1 /*===---- arm_acle.h - ARM Non-Neon intrinsics -----------------------------=== 2 * 3 * Permission is hereby granted, free of charge, to any person obtaining a copy 4 * of this software and associated documentation files (the "Software"), to deal 5 * in the Software without restriction, including without limitation the rights 6 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell 7 * copies of the Software, and to permit persons to whom the Software is 8 * furnished to do so, subject to the following conditions: 9 * 10 * The above copyright notice and this permission notice shall be included in 11 * all copies or substantial portions of the Software. 12 * 13 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 14 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 15 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE 16 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 17 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, 18 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN 19 * THE SOFTWARE. 20 * 21 *===-----------------------------------------------------------------------=== 22 */ 23 24 #ifndef __ARM_ACLE_H 25 #define __ARM_ACLE_H 26 27 #ifndef __ARM_ACLE 28 #error "ACLE intrinsics support not enabled." 29 #endif 30 31 #include <stdint.h> 32 33 #if defined(__cplusplus) 34 extern "C" { 35 #endif 36 37 /* 8 SYNCHRONIZATION, BARRIER AND HINT INTRINSICS */ 38 /* 8.3 Memory barriers */ 39 #if !defined(_MSC_VER) 40 #define __dmb(i) __builtin_arm_dmb(i) 41 #define __dsb(i) __builtin_arm_dsb(i) 42 #define __isb(i) __builtin_arm_isb(i) 43 #endif 44 45 /* 8.4 Hints */ 46 47 #if !defined(_MSC_VER) 48 static __inline__ void __attribute__((__always_inline__, __nodebug__)) __wfi(void) { 49 __builtin_arm_wfi(); 50 } 51 52 static __inline__ void __attribute__((__always_inline__, __nodebug__)) __wfe(void) { 53 __builtin_arm_wfe(); 54 } 55 56 static __inline__ void __attribute__((__always_inline__, __nodebug__)) __sev(void) { 57 __builtin_arm_sev(); 58 } 59 60 static __inline__ void __attribute__((__always_inline__, __nodebug__)) __sevl(void) { 61 __builtin_arm_sevl(); 62 } 63 64 static __inline__ void __attribute__((__always_inline__, __nodebug__)) __yield(void) { 65 __builtin_arm_yield(); 66 } 67 #endif 68 69 #if __ARM_32BIT_STATE 70 #define __dbg(t) __builtin_arm_dbg(t) 71 #endif 72 73 /* 8.5 Swap */ 74 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__)) 75 __swp(uint32_t x, volatile uint32_t *p) { 76 uint32_t v; 77 do v = __builtin_arm_ldrex(p); while (__builtin_arm_strex(x, p)); 78 return v; 79 } 80 81 /* 8.6 Memory prefetch intrinsics */ 82 /* 8.6.1 Data prefetch */ 83 #define __pld(addr) __pldx(0, 0, 0, addr) 84 85 #if __ARM_32BIT_STATE 86 #define __pldx(access_kind, cache_level, retention_policy, addr) \ 87 __builtin_arm_prefetch(addr, access_kind, 1) 88 #else 89 #define __pldx(access_kind, cache_level, retention_policy, addr) \ 90 __builtin_arm_prefetch(addr, access_kind, cache_level, retention_policy, 1) 91 #endif 92 93 /* 8.6.2 Instruction prefetch */ 94 #define __pli(addr) __plix(0, 0, addr) 95 96 #if __ARM_32BIT_STATE 97 #define __plix(cache_level, retention_policy, addr) \ 98 __builtin_arm_prefetch(addr, 0, 0) 99 #else 100 #define __plix(cache_level, retention_policy, addr) \ 101 __builtin_arm_prefetch(addr, 0, cache_level, retention_policy, 0) 102 #endif 103 104 /* 8.7 NOP */ 105 static __inline__ void __attribute__((__always_inline__, __nodebug__)) __nop(void) { 106 __builtin_arm_nop(); 107 } 108 109 /* 9 DATA-PROCESSING INTRINSICS */ 110 /* 9.2 Miscellaneous data-processing intrinsics */ 111 /* ROR */ 112 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__)) 113 __ror(uint32_t x, uint32_t y) { 114 y %= 32; 115 if (y == 0) return x; 116 return (x >> y) | (x << (32 - y)); 117 } 118 119 static __inline__ uint64_t __attribute__((__always_inline__, __nodebug__)) 120 __rorll(uint64_t x, uint32_t y) { 121 y %= 64; 122 if (y == 0) return x; 123 return (x >> y) | (x << (64 - y)); 124 } 125 126 static __inline__ unsigned long __attribute__((__always_inline__, __nodebug__)) 127 __rorl(unsigned long x, uint32_t y) { 128 #if __SIZEOF_LONG__ == 4 129 return __ror(x, y); 130 #else 131 return __rorll(x, y); 132 #endif 133 } 134 135 136 /* CLZ */ 137 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__)) 138 __clz(uint32_t t) { 139 return __builtin_clz(t); 140 } 141 142 static __inline__ unsigned long __attribute__((__always_inline__, __nodebug__)) 143 __clzl(unsigned long t) { 144 return __builtin_clzl(t); 145 } 146 147 static __inline__ uint64_t __attribute__((__always_inline__, __nodebug__)) 148 __clzll(uint64_t t) { 149 return __builtin_clzll(t); 150 } 151 152 /* REV */ 153 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__)) 154 __rev(uint32_t t) { 155 return __builtin_bswap32(t); 156 } 157 158 static __inline__ unsigned long __attribute__((__always_inline__, __nodebug__)) 159 __revl(unsigned long t) { 160 #if __SIZEOF_LONG__ == 4 161 return __builtin_bswap32(t); 162 #else 163 return __builtin_bswap64(t); 164 #endif 165 } 166 167 static __inline__ uint64_t __attribute__((__always_inline__, __nodebug__)) 168 __revll(uint64_t t) { 169 return __builtin_bswap64(t); 170 } 171 172 /* REV16 */ 173 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__)) 174 __rev16(uint32_t t) { 175 return __ror(__rev(t), 16); 176 } 177 178 static __inline__ uint64_t __attribute__((__always_inline__, __nodebug__)) 179 __rev16ll(uint64_t t) { 180 return (((uint64_t)__rev16(t >> 32)) << 32) | __rev16(t); 181 } 182 183 static __inline__ unsigned long __attribute__((__always_inline__, __nodebug__)) 184 __rev16l(unsigned long t) { 185 #if __SIZEOF_LONG__ == 4 186 return __rev16(t); 187 #else 188 return __rev16ll(t); 189 #endif 190 } 191 192 /* REVSH */ 193 static __inline__ int16_t __attribute__((__always_inline__, __nodebug__)) 194 __revsh(int16_t t) { 195 return __builtin_bswap16(t); 196 } 197 198 /* RBIT */ 199 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__)) 200 __rbit(uint32_t t) { 201 return __builtin_arm_rbit(t); 202 } 203 204 static __inline__ uint64_t __attribute__((__always_inline__, __nodebug__)) 205 __rbitll(uint64_t t) { 206 #if __ARM_32BIT_STATE 207 return (((uint64_t) __builtin_arm_rbit(t)) << 32) | 208 __builtin_arm_rbit(t >> 32); 209 #else 210 return __builtin_arm_rbit64(t); 211 #endif 212 } 213 214 static __inline__ unsigned long __attribute__((__always_inline__, __nodebug__)) 215 __rbitl(unsigned long t) { 216 #if __SIZEOF_LONG__ == 4 217 return __rbit(t); 218 #else 219 return __rbitll(t); 220 #endif 221 } 222 223 /* 224 * 9.4 Saturating intrinsics 225 * 226 * FIXME: Change guard to their corrosponding __ARM_FEATURE flag when Q flag 227 * intrinsics are implemented and the flag is enabled. 228 */ 229 /* 9.4.1 Width-specified saturation intrinsics */ 230 #if __ARM_32BIT_STATE 231 #define __ssat(x, y) __builtin_arm_ssat(x, y) 232 #define __usat(x, y) __builtin_arm_usat(x, y) 233 #endif 234 235 /* 9.4.2 Saturating addition and subtraction intrinsics */ 236 #if __ARM_32BIT_STATE 237 static __inline__ int32_t __attribute__((__always_inline__, __nodebug__)) 238 __qadd(int32_t t, int32_t v) { 239 return __builtin_arm_qadd(t, v); 240 } 241 242 static __inline__ int32_t __attribute__((__always_inline__, __nodebug__)) 243 __qsub(int32_t t, int32_t v) { 244 return __builtin_arm_qsub(t, v); 245 } 246 247 static __inline__ int32_t __attribute__((__always_inline__, __nodebug__)) 248 __qdbl(int32_t t) { 249 return __builtin_arm_qadd(t, t); 250 } 251 #endif 252 253 /* 9.7 CRC32 intrinsics */ 254 #if __ARM_FEATURE_CRC32 255 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__)) 256 __crc32b(uint32_t a, uint8_t b) { 257 return __builtin_arm_crc32b(a, b); 258 } 259 260 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__)) 261 __crc32h(uint32_t a, uint16_t b) { 262 return __builtin_arm_crc32h(a, b); 263 } 264 265 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__)) 266 __crc32w(uint32_t a, uint32_t b) { 267 return __builtin_arm_crc32w(a, b); 268 } 269 270 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__)) 271 __crc32d(uint32_t a, uint64_t b) { 272 return __builtin_arm_crc32d(a, b); 273 } 274 275 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__)) 276 __crc32cb(uint32_t a, uint8_t b) { 277 return __builtin_arm_crc32cb(a, b); 278 } 279 280 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__)) 281 __crc32ch(uint32_t a, uint16_t b) { 282 return __builtin_arm_crc32ch(a, b); 283 } 284 285 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__)) 286 __crc32cw(uint32_t a, uint32_t b) { 287 return __builtin_arm_crc32cw(a, b); 288 } 289 290 static __inline__ uint32_t __attribute__((__always_inline__, __nodebug__)) 291 __crc32cd(uint32_t a, uint64_t b) { 292 return __builtin_arm_crc32cd(a, b); 293 } 294 #endif 295 296 /* 10.1 Special register intrinsics */ 297 #define __arm_rsr(sysreg) __builtin_arm_rsr(sysreg) 298 #define __arm_rsr64(sysreg) __builtin_arm_rsr64(sysreg) 299 #define __arm_rsrp(sysreg) __builtin_arm_rsrp(sysreg) 300 #define __arm_wsr(sysreg, v) __builtin_arm_wsr(sysreg, v) 301 #define __arm_wsr64(sysreg, v) __builtin_arm_wsr64(sysreg, v) 302 #define __arm_wsrp(sysreg, v) __builtin_arm_wsrp(sysreg, v) 303 304 #if defined(__cplusplus) 305 } 306 #endif 307 308 #endif /* __ARM_ACLE_H */ 309