Home | History | Annotate | Download | only in clang-include
      1 /*===---- arm_acle.h - ARM Non-Neon intrinsics -----------------------------===
      2  *
      3  * Permission is hereby granted, free of charge, to any person obtaining a copy
      4  * of this software and associated documentation files (the "Software"), to deal
      5  * in the Software without restriction, including without limitation the rights
      6  * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
      7  * copies of the Software, and to permit persons to whom the Software is
      8  * furnished to do so, subject to the following conditions:
      9  *
     10  * The above copyright notice and this permission notice shall be included in
     11  * all copies or substantial portions of the Software.
     12  *
     13  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
     14  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
     15  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
     16  * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
     17  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
     18  * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
     19  * THE SOFTWARE.
     20  *
     21  *===-----------------------------------------------------------------------===
     22  */
     23 
     24 #ifndef __ARM_ACLE_H
     25 #define __ARM_ACLE_H
     26 
     27 #ifndef __ARM_ACLE
     28 #error "ACLE intrinsics support not enabled."
     29 #endif
     30 
     31 #include <stdint.h>
     32 
     33 #if defined(__cplusplus)
     34 extern "C" {
     35 #endif
     36 
     37 /* 8 SYNCHRONIZATION, BARRIER AND HINT INTRINSICS */
     38 /* 8.3 Memory barriers */
     39 #if !defined(_MSC_VER)
     40 #define __dmb(i) __builtin_arm_dmb(i)
     41 #define __dsb(i) __builtin_arm_dsb(i)
     42 #define __isb(i) __builtin_arm_isb(i)
     43 #endif
     44 
     45 /* 8.4 Hints */
     46 
     47 #if !defined(_MSC_VER)
     48 static __inline__ void __attribute__((__always_inline__, __nodebug__)) __wfi(void) {
     49   __builtin_arm_wfi();
     50 }
     51 
     52 static __inline__ void __attribute__((__always_inline__, __nodebug__)) __wfe(void) {
     53   __builtin_arm_wfe();
     54 }
     55 
     56 static __inline__ void __attribute__((__always_inline__, __nodebug__)) __sev(void) {
     57   __builtin_arm_sev();
     58 }
     59 
     60 static __inline__ void __attribute__((__always_inline__, __nodebug__)) __sevl(void) {
     61   __builtin_arm_sevl();
     62 }
     63 
     64 static __inline__ void __attribute__((__always_inline__, __nodebug__)) __yield(void) {
     65   __builtin_arm_yield();
     66 }
     67 #endif
     68 
     69 #if __ARM_32BIT_STATE
     70 #define __dbg(t) __builtin_arm_dbg(t)
     71 #endif
     72 
     73 /* 8.5 Swap */
     74 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__))
     75   __swp(uint32_t x, volatile uint32_t *p) {
     76   uint32_t v;
     77   do v = __builtin_arm_ldrex(p); while (__builtin_arm_strex(x, p));
     78   return v;
     79 }
     80 
     81 /* 8.6 Memory prefetch intrinsics */
     82 /* 8.6.1 Data prefetch */
     83 #define __pld(addr) __pldx(0, 0, 0, addr)
     84 
     85 #if __ARM_32BIT_STATE
     86 #define __pldx(access_kind, cache_level, retention_policy, addr) \
     87   __builtin_arm_prefetch(addr, access_kind, 1)
     88 #else
     89 #define __pldx(access_kind, cache_level, retention_policy, addr) \
     90   __builtin_arm_prefetch(addr, access_kind, cache_level, retention_policy, 1)
     91 #endif
     92 
     93 /* 8.6.2 Instruction prefetch */
     94 #define __pli(addr) __plix(0, 0, addr)
     95 
     96 #if __ARM_32BIT_STATE
     97 #define __plix(cache_level, retention_policy, addr) \
     98   __builtin_arm_prefetch(addr, 0, 0)
     99 #else
    100 #define __plix(cache_level, retention_policy, addr) \
    101   __builtin_arm_prefetch(addr, 0, cache_level, retention_policy, 0)
    102 #endif
    103 
    104 /* 8.7 NOP */
    105 static __inline__ void __attribute__((__always_inline__, __nodebug__)) __nop(void) {
    106   __builtin_arm_nop();
    107 }
    108 
    109 /* 9 DATA-PROCESSING INTRINSICS */
    110 /* 9.2 Miscellaneous data-processing intrinsics */
    111 /* ROR */
    112 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__))
    113   __ror(uint32_t x, uint32_t y) {
    114   y %= 32;
    115   if (y == 0)  return x;
    116   return (x >> y) | (x << (32 - y));
    117 }
    118 
    119 static __inline__ uint64_t __attribute__((__always_inline__, __nodebug__))
    120   __rorll(uint64_t x, uint32_t y) {
    121   y %= 64;
    122   if (y == 0)  return x;
    123   return (x >> y) | (x << (64 - y));
    124 }
    125 
    126 static __inline__ unsigned long __attribute__((__always_inline__, __nodebug__))
    127   __rorl(unsigned long x, uint32_t y) {
    128 #if __SIZEOF_LONG__ == 4
    129   return __ror(x, y);
    130 #else
    131   return __rorll(x, y);
    132 #endif
    133 }
    134 
    135 
    136 /* CLZ */
    137 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__))
    138   __clz(uint32_t t) {
    139   return __builtin_clz(t);
    140 }
    141 
    142 static __inline__ unsigned long __attribute__((__always_inline__, __nodebug__))
    143   __clzl(unsigned long t) {
    144   return __builtin_clzl(t);
    145 }
    146 
    147 static __inline__ uint64_t __attribute__((__always_inline__, __nodebug__))
    148   __clzll(uint64_t t) {
    149   return __builtin_clzll(t);
    150 }
    151 
    152 /* REV */
    153 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__))
    154   __rev(uint32_t t) {
    155   return __builtin_bswap32(t);
    156 }
    157 
    158 static __inline__ unsigned long __attribute__((__always_inline__, __nodebug__))
    159   __revl(unsigned long t) {
    160 #if __SIZEOF_LONG__ == 4
    161   return __builtin_bswap32(t);
    162 #else
    163   return __builtin_bswap64(t);
    164 #endif
    165 }
    166 
    167 static __inline__ uint64_t __attribute__((__always_inline__, __nodebug__))
    168   __revll(uint64_t t) {
    169   return __builtin_bswap64(t);
    170 }
    171 
    172 /* REV16 */
    173 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__))
    174   __rev16(uint32_t t) {
    175   return __ror(__rev(t), 16);
    176 }
    177 
    178 static __inline__ uint64_t __attribute__((__always_inline__, __nodebug__))
    179   __rev16ll(uint64_t t) {
    180   return (((uint64_t)__rev16(t >> 32)) << 32) | __rev16(t);
    181 }
    182 
    183 static __inline__ unsigned long __attribute__((__always_inline__, __nodebug__))
    184   __rev16l(unsigned long t) {
    185 #if __SIZEOF_LONG__ == 4
    186     return __rev16(t);
    187 #else
    188     return __rev16ll(t);
    189 #endif
    190 }
    191 
    192 /* REVSH */
    193 static __inline__ int16_t __attribute__((__always_inline__, __nodebug__))
    194   __revsh(int16_t t) {
    195   return __builtin_bswap16(t);
    196 }
    197 
    198 /* RBIT */
    199 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__))
    200   __rbit(uint32_t t) {
    201   return __builtin_arm_rbit(t);
    202 }
    203 
    204 static __inline__ uint64_t __attribute__((__always_inline__, __nodebug__))
    205   __rbitll(uint64_t t) {
    206 #if __ARM_32BIT_STATE
    207   return (((uint64_t) __builtin_arm_rbit(t)) << 32) |
    208     __builtin_arm_rbit(t >> 32);
    209 #else
    210   return __builtin_arm_rbit64(t);
    211 #endif
    212 }
    213 
    214 static __inline__ unsigned long __attribute__((__always_inline__, __nodebug__))
    215   __rbitl(unsigned long t) {
    216 #if __SIZEOF_LONG__ == 4
    217   return __rbit(t);
    218 #else
    219   return __rbitll(t);
    220 #endif
    221 }
    222 
    223 /*
    224  * 9.4 Saturating intrinsics
    225  *
    226  * FIXME: Change guard to their corrosponding __ARM_FEATURE flag when Q flag
    227  * intrinsics are implemented and the flag is enabled.
    228  */
    229 /* 9.4.1 Width-specified saturation intrinsics */
    230 #if __ARM_32BIT_STATE
    231 #define __ssat(x, y) __builtin_arm_ssat(x, y)
    232 #define __usat(x, y) __builtin_arm_usat(x, y)
    233 #endif
    234 
    235 /* 9.4.2 Saturating addition and subtraction intrinsics */
    236 #if __ARM_32BIT_STATE
    237 static __inline__ int32_t __attribute__((__always_inline__, __nodebug__))
    238   __qadd(int32_t t, int32_t v) {
    239   return __builtin_arm_qadd(t, v);
    240 }
    241 
    242 static __inline__ int32_t __attribute__((__always_inline__, __nodebug__))
    243   __qsub(int32_t t, int32_t v) {
    244   return __builtin_arm_qsub(t, v);
    245 }
    246 
    247 static __inline__ int32_t __attribute__((__always_inline__, __nodebug__))
    248 __qdbl(int32_t t) {
    249   return __builtin_arm_qadd(t, t);
    250 }
    251 #endif
    252 
    253 /* 9.7 CRC32 intrinsics */
    254 #if __ARM_FEATURE_CRC32
    255 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__))
    256   __crc32b(uint32_t a, uint8_t b) {
    257   return __builtin_arm_crc32b(a, b);
    258 }
    259 
    260 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__))
    261   __crc32h(uint32_t a, uint16_t b) {
    262   return __builtin_arm_crc32h(a, b);
    263 }
    264 
    265 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__))
    266   __crc32w(uint32_t a, uint32_t b) {
    267   return __builtin_arm_crc32w(a, b);
    268 }
    269 
    270 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__))
    271   __crc32d(uint32_t a, uint64_t b) {
    272   return __builtin_arm_crc32d(a, b);
    273 }
    274 
    275 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__))
    276   __crc32cb(uint32_t a, uint8_t b) {
    277   return __builtin_arm_crc32cb(a, b);
    278 }
    279 
    280 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__))
    281   __crc32ch(uint32_t a, uint16_t b) {
    282   return __builtin_arm_crc32ch(a, b);
    283 }
    284 
    285 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__))
    286   __crc32cw(uint32_t a, uint32_t b) {
    287   return __builtin_arm_crc32cw(a, b);
    288 }
    289 
    290 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__))
    291   __crc32cd(uint32_t a, uint64_t b) {
    292   return __builtin_arm_crc32cd(a, b);
    293 }
    294 #endif
    295 
    296 /* 10.1 Special register intrinsics */
    297 #define __arm_rsr(sysreg) __builtin_arm_rsr(sysreg)
    298 #define __arm_rsr64(sysreg) __builtin_arm_rsr64(sysreg)
    299 #define __arm_rsrp(sysreg) __builtin_arm_rsrp(sysreg)
    300 #define __arm_wsr(sysreg, v) __builtin_arm_wsr(sysreg, v)
    301 #define __arm_wsr64(sysreg, v) __builtin_arm_wsr64(sysreg, v)
    302 #define __arm_wsrp(sysreg, v) __builtin_arm_wsrp(sysreg, v)
    303 
    304 #if defined(__cplusplus)
    305 }
    306 #endif
    307 
    308 #endif /* __ARM_ACLE_H */
    309