Home | History | Annotate | Download | only in cmsis
      1 /**************************************************************************//**
      2  * @file     core_cmInstr.h
      3  * @brief    CMSIS Cortex-M Core Instruction Access Header File
      4  * @version  V4.00
      5  * @date     28. August 2014
      6  *
      7  * @note
      8  *
      9  ******************************************************************************/
     10 /* Copyright (c) 2009 - 2014 ARM LIMITED
     11 
     12    All rights reserved.
     13    Redistribution and use in source and binary forms, with or without
     14    modification, are permitted provided that the following conditions are met:
     15    - Redistributions of source code must retain the above copyright
     16      notice, this list of conditions and the following disclaimer.
     17    - Redistributions in binary form must reproduce the above copyright
     18      notice, this list of conditions and the following disclaimer in the
     19      documentation and/or other materials provided with the distribution.
     20    - Neither the name of ARM nor the names of its contributors may be used
     21      to endorse or promote products derived from this software without
     22      specific prior written permission.
     23    *
     24    THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
     25    AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
     26    IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
     27    ARE DISCLAIMED. IN NO EVENT SHALL COPYRIGHT HOLDERS AND CONTRIBUTORS BE
     28    LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
     29    CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
     30    SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
     31    INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
     32    CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
     33    ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
     34    POSSIBILITY OF SUCH DAMAGE.
     35    ---------------------------------------------------------------------------*/
     36 
     37 
     38 #ifndef __CORE_CMINSTR_H
     39 #define __CORE_CMINSTR_H
     40 
     41 
     42 /* ##########################  Core Instruction Access  ######################### */
     43 /** \defgroup CMSIS_Core_InstructionInterface CMSIS Core Instruction Interface
     44   Access to dedicated instructions
     45   @{
     46 */
     47 
     48 #if   defined ( __CC_ARM ) /*------------------RealView Compiler -----------------*/
     49 /* ARM armcc specific functions */
     50 
     51 #if (__ARMCC_VERSION < 400677)
     52   #error "Please use ARM Compiler Toolchain V4.0.677 or later!"
     53 #endif
     54 
     55 
     56 /** \brief  No Operation
     57 
     58     No Operation does nothing. This instruction can be used for code alignment purposes.
     59  */
     60 #define __NOP                             __nop
     61 
     62 
     63 /** \brief  Wait For Interrupt
     64 
     65     Wait For Interrupt is a hint instruction that suspends execution
     66     until one of a number of events occurs.
     67  */
     68 #define __WFI                             __wfi
     69 
     70 
     71 /** \brief  Wait For Event
     72 
     73     Wait For Event is a hint instruction that permits the processor to enter
     74     a low-power state until one of a number of events occurs.
     75  */
     76 #define __WFE                             __wfe
     77 
     78 
     79 /** \brief  Send Event
     80 
     81     Send Event is a hint instruction. It causes an event to be signaled to the CPU.
     82  */
     83 #define __SEV                             __sev
     84 
     85 
     86 /** \brief  Instruction Synchronization Barrier
     87 
     88     Instruction Synchronization Barrier flushes the pipeline in the processor,
     89     so that all instructions following the ISB are fetched from cache or
     90     memory, after the instruction has been completed.
     91  */
     92 #define __ISB()                           __isb(0xF)
     93 
     94 
     95 /** \brief  Data Synchronization Barrier
     96 
     97     This function acts as a special kind of Data Memory Barrier.
     98     It completes when all explicit memory accesses before this instruction complete.
     99  */
    100 #define __DSB()                           __dsb(0xF)
    101 
    102 
    103 /** \brief  Data Memory Barrier
    104 
    105     This function ensures the apparent order of the explicit memory operations before
    106     and after the instruction, without ensuring their completion.
    107  */
    108 #define __DMB()                           __dmb(0xF)
    109 
    110 
    111 /** \brief  Reverse byte order (32 bit)
    112 
    113     This function reverses the byte order in integer value.
    114 
    115     \param [in]    value  Value to reverse
    116     \return               Reversed value
    117  */
    118 #define __REV                             __rev
    119 
    120 
    121 /** \brief  Reverse byte order (16 bit)
    122 
    123     This function reverses the byte order in two unsigned short values.
    124 
    125     \param [in]    value  Value to reverse
    126     \return               Reversed value
    127  */
    128 #ifndef __NO_EMBEDDED_ASM
    129 __attribute__((section(".rev16_text"))) __STATIC_INLINE __ASM uint32_t __REV16(uint32_t value)
    130 {
    131   rev16 r0, r0
    132   bx lr
    133 }
    134 #endif
    135 
    136 /** \brief  Reverse byte order in signed short value
    137 
    138     This function reverses the byte order in a signed short value with sign extension to integer.
    139 
    140     \param [in]    value  Value to reverse
    141     \return               Reversed value
    142  */
    143 #ifndef __NO_EMBEDDED_ASM
    144 __attribute__((section(".revsh_text"))) __STATIC_INLINE __ASM int32_t __REVSH(int32_t value)
    145 {
    146   revsh r0, r0
    147   bx lr
    148 }
    149 #endif
    150 
    151 
    152 /** \brief  Rotate Right in unsigned value (32 bit)
    153 
    154     This function Rotate Right (immediate) provides the value of the contents of a register rotated by a variable number of bits.
    155 
    156     \param [in]    value  Value to rotate
    157     \param [in]    value  Number of Bits to rotate
    158     \return               Rotated value
    159  */
    160 #define __ROR                             __ror
    161 
    162 
    163 /** \brief  Breakpoint
    164 
    165     This function causes the processor to enter Debug state.
    166     Debug tools can use this to investigate system state when the instruction at a particular address is reached.
    167 
    168     \param [in]    value  is ignored by the processor.
    169                    If required, a debugger can use it to store additional information about the breakpoint.
    170  */
    171 #define __BKPT(value)                       __breakpoint(value)
    172 
    173 
    174 #if       (__CORTEX_M >= 0x03) || (__CORTEX_SC >= 300)
    175 
    176 /** \brief  Reverse bit order of value
    177 
    178     This function reverses the bit order of the given value.
    179 
    180     \param [in]    value  Value to reverse
    181     \return               Reversed value
    182  */
    183 #define __RBIT                            __rbit
    184 
    185 
    186 /** \brief  LDR Exclusive (8 bit)
    187 
    188     This function executes a exclusive LDR instruction for 8 bit value.
    189 
    190     \param [in]    ptr  Pointer to data
    191     \return             value of type uint8_t at (*ptr)
    192  */
    193 #define __LDREXB(ptr)                     ((uint8_t ) __ldrex(ptr))
    194 
    195 
    196 /** \brief  LDR Exclusive (16 bit)
    197 
    198     This function executes a exclusive LDR instruction for 16 bit values.
    199 
    200     \param [in]    ptr  Pointer to data
    201     \return        value of type uint16_t at (*ptr)
    202  */
    203 #define __LDREXH(ptr)                     ((uint16_t) __ldrex(ptr))
    204 
    205 
    206 /** \brief  LDR Exclusive (32 bit)
    207 
    208     This function executes a exclusive LDR instruction for 32 bit values.
    209 
    210     \param [in]    ptr  Pointer to data
    211     \return        value of type uint32_t at (*ptr)
    212  */
    213 #define __LDREXW(ptr)                     ((uint32_t ) __ldrex(ptr))
    214 
    215 
    216 /** \brief  STR Exclusive (8 bit)
    217 
    218     This function executes a exclusive STR instruction for 8 bit values.
    219 
    220     \param [in]  value  Value to store
    221     \param [in]    ptr  Pointer to location
    222     \return          0  Function succeeded
    223     \return          1  Function failed
    224  */
    225 #define __STREXB(value, ptr)              __strex(value, ptr)
    226 
    227 
    228 /** \brief  STR Exclusive (16 bit)
    229 
    230     This function executes a exclusive STR instruction for 16 bit values.
    231 
    232     \param [in]  value  Value to store
    233     \param [in]    ptr  Pointer to location
    234     \return          0  Function succeeded
    235     \return          1  Function failed
    236  */
    237 #define __STREXH(value, ptr)              __strex(value, ptr)
    238 
    239 
    240 /** \brief  STR Exclusive (32 bit)
    241 
    242     This function executes a exclusive STR instruction for 32 bit values.
    243 
    244     \param [in]  value  Value to store
    245     \param [in]    ptr  Pointer to location
    246     \return          0  Function succeeded
    247     \return          1  Function failed
    248  */
    249 #define __STREXW(value, ptr)              __strex(value, ptr)
    250 
    251 
    252 /** \brief  Remove the exclusive lock
    253 
    254     This function removes the exclusive lock which is created by LDREX.
    255 
    256  */
    257 #define __CLREX                           __clrex
    258 
    259 
    260 /** \brief  Signed Saturate
    261 
    262     This function saturates a signed value.
    263 
    264     \param [in]  value  Value to be saturated
    265     \param [in]    sat  Bit position to saturate to (1..32)
    266     \return             Saturated value
    267  */
    268 #define __SSAT                            __ssat
    269 
    270 
    271 /** \brief  Unsigned Saturate
    272 
    273     This function saturates an unsigned value.
    274 
    275     \param [in]  value  Value to be saturated
    276     \param [in]    sat  Bit position to saturate to (0..31)
    277     \return             Saturated value
    278  */
    279 #define __USAT                            __usat
    280 
    281 
    282 /** \brief  Count leading zeros
    283 
    284     This function counts the number of leading zeros of a data value.
    285 
    286     \param [in]  value  Value to count the leading zeros
    287     \return             number of leading zeros in value
    288  */
    289 #define __CLZ                             __clz
    290 
    291 
    292 /** \brief  Rotate Right with Extend (32 bit)
    293 
    294     This function moves each bit of a bitstring right by one bit. The carry input is shifted in at the left end of the bitstring.
    295 
    296     \param [in]    value  Value to rotate
    297     \return               Rotated value
    298  */
    299 #ifndef __NO_EMBEDDED_ASM
    300 __attribute__((section(".rrx_text"))) __STATIC_INLINE __ASM uint32_t __RRX(uint32_t value)
    301 {
    302   rrx r0, r0
    303   bx lr
    304 }
    305 #endif
    306 
    307 
    308 /** \brief  LDRT Unprivileged (8 bit)
    309 
    310     This function executes a Unprivileged LDRT instruction for 8 bit value.
    311 
    312     \param [in]    ptr  Pointer to data
    313     \return             value of type uint8_t at (*ptr)
    314  */
    315 #define __LDRBT(ptr)                      ((uint8_t )  __ldrt(ptr))
    316 
    317 
    318 /** \brief  LDRT Unprivileged (16 bit)
    319 
    320     This function executes a Unprivileged LDRT instruction for 16 bit values.
    321 
    322     \param [in]    ptr  Pointer to data
    323     \return        value of type uint16_t at (*ptr)
    324  */
    325 #define __LDRHT(ptr)                      ((uint16_t)  __ldrt(ptr))
    326 
    327 
    328 /** \brief  LDRT Unprivileged (32 bit)
    329 
    330     This function executes a Unprivileged LDRT instruction for 32 bit values.
    331 
    332     \param [in]    ptr  Pointer to data
    333     \return        value of type uint32_t at (*ptr)
    334  */
    335 #define __LDRT(ptr)                       ((uint32_t ) __ldrt(ptr))
    336 
    337 
    338 /** \brief  STRT Unprivileged (8 bit)
    339 
    340     This function executes a Unprivileged STRT instruction for 8 bit values.
    341 
    342     \param [in]  value  Value to store
    343     \param [in]    ptr  Pointer to location
    344  */
    345 #define __STRBT(value, ptr)               __strt(value, ptr)
    346 
    347 
    348 /** \brief  STRT Unprivileged (16 bit)
    349 
    350     This function executes a Unprivileged STRT instruction for 16 bit values.
    351 
    352     \param [in]  value  Value to store
    353     \param [in]    ptr  Pointer to location
    354  */
    355 #define __STRHT(value, ptr)               __strt(value, ptr)
    356 
    357 
    358 /** \brief  STRT Unprivileged (32 bit)
    359 
    360     This function executes a Unprivileged STRT instruction for 32 bit values.
    361 
    362     \param [in]  value  Value to store
    363     \param [in]    ptr  Pointer to location
    364  */
    365 #define __STRT(value, ptr)                __strt(value, ptr)
    366 
    367 #endif /* (__CORTEX_M >= 0x03) || (__CORTEX_SC >= 300) */
    368 
    369 
    370 #elif defined ( __GNUC__ ) /*------------------ GNU Compiler ---------------------*/
    371 /* GNU gcc specific functions */
    372 
    373 /* Define macros for porting to both thumb1 and thumb2.
    374  * For thumb1, use low register (r0-r7), specified by constrant "l"
    375  * Otherwise, use general registers, specified by constrant "r" */
    376 #if defined (__thumb__) && !defined (__thumb2__)
    377 #define __CMSIS_GCC_OUT_REG(r) "=l" (r)
    378 #define __CMSIS_GCC_USE_REG(r) "l" (r)
    379 #else
    380 #define __CMSIS_GCC_OUT_REG(r) "=r" (r)
    381 #define __CMSIS_GCC_USE_REG(r) "r" (r)
    382 #endif
    383 
    384 /** \brief  No Operation
    385 
    386     No Operation does nothing. This instruction can be used for code alignment purposes.
    387  */
    388 __attribute__( ( always_inline ) ) __STATIC_INLINE void __NOP(void)
    389 {
    390   __ASM volatile ("nop");
    391 }
    392 
    393 
    394 /** \brief  Wait For Interrupt
    395 
    396     Wait For Interrupt is a hint instruction that suspends execution
    397     until one of a number of events occurs.
    398  */
    399 __attribute__( ( always_inline ) ) __STATIC_INLINE void __WFI(void)
    400 {
    401   __ASM volatile ("wfi");
    402 }
    403 
    404 
    405 /** \brief  Wait For Event
    406 
    407     Wait For Event is a hint instruction that permits the processor to enter
    408     a low-power state until one of a number of events occurs.
    409  */
    410 __attribute__( ( always_inline ) ) __STATIC_INLINE void __WFE(void)
    411 {
    412   __ASM volatile ("wfe");
    413 }
    414 
    415 
    416 /** \brief  Send Event
    417 
    418     Send Event is a hint instruction. It causes an event to be signaled to the CPU.
    419  */
    420 __attribute__( ( always_inline ) ) __STATIC_INLINE void __SEV(void)
    421 {
    422   __ASM volatile ("sev");
    423 }
    424 
    425 
    426 /** \brief  Instruction Synchronization Barrier
    427 
    428     Instruction Synchronization Barrier flushes the pipeline in the processor,
    429     so that all instructions following the ISB are fetched from cache or
    430     memory, after the instruction has been completed.
    431  */
    432 __attribute__( ( always_inline ) ) __STATIC_INLINE void __ISB(void)
    433 {
    434   __ASM volatile ("isb");
    435 }
    436 
    437 
    438 /** \brief  Data Synchronization Barrier
    439 
    440     This function acts as a special kind of Data Memory Barrier.
    441     It completes when all explicit memory accesses before this instruction complete.
    442  */
    443 __attribute__( ( always_inline ) ) __STATIC_INLINE void __DSB(void)
    444 {
    445   __ASM volatile ("dsb");
    446 }
    447 
    448 
    449 /** \brief  Data Memory Barrier
    450 
    451     This function ensures the apparent order of the explicit memory operations before
    452     and after the instruction, without ensuring their completion.
    453  */
    454 __attribute__( ( always_inline ) ) __STATIC_INLINE void __DMB(void)
    455 {
    456   __ASM volatile ("dmb");
    457 }
    458 
    459 
    460 /** \brief  Reverse byte order (32 bit)
    461 
    462     This function reverses the byte order in integer value.
    463 
    464     \param [in]    value  Value to reverse
    465     \return               Reversed value
    466  */
    467 __attribute__( ( always_inline ) ) __STATIC_INLINE uint32_t __REV(uint32_t value)
    468 {
    469 #if (__GNUC__ > 4) || (__GNUC__ == 4 && __GNUC_MINOR__ >= 5)
    470   return __builtin_bswap32(value);
    471 #else
    472   uint32_t result;
    473 
    474   __ASM volatile ("rev %0, %1" : __CMSIS_GCC_OUT_REG (result) : __CMSIS_GCC_USE_REG (value) );
    475   return(result);
    476 #endif
    477 }
    478 
    479 
    480 /** \brief  Reverse byte order (16 bit)
    481 
    482     This function reverses the byte order in two unsigned short values.
    483 
    484     \param [in]    value  Value to reverse
    485     \return               Reversed value
    486  */
    487 __attribute__( ( always_inline ) ) __STATIC_INLINE uint32_t __REV16(uint32_t value)
    488 {
    489   uint32_t result;
    490 
    491   __ASM volatile ("rev16 %0, %1" : __CMSIS_GCC_OUT_REG (result) : __CMSIS_GCC_USE_REG (value) );
    492   return(result);
    493 }
    494 
    495 
    496 /** \brief  Reverse byte order in signed short value
    497 
    498     This function reverses the byte order in a signed short value with sign extension to integer.
    499 
    500     \param [in]    value  Value to reverse
    501     \return               Reversed value
    502  */
    503 __attribute__( ( always_inline ) ) __STATIC_INLINE int32_t __REVSH(int32_t value)
    504 {
    505 #if (__GNUC__ > 4) || (__GNUC__ == 4 && __GNUC_MINOR__ >= 8)
    506   return (short)__builtin_bswap16(value);
    507 #else
    508   uint32_t result;
    509 
    510   __ASM volatile ("revsh %0, %1" : __CMSIS_GCC_OUT_REG (result) : __CMSIS_GCC_USE_REG (value) );
    511   return(result);
    512 #endif
    513 }
    514 
    515 
    516 /** \brief  Rotate Right in unsigned value (32 bit)
    517 
    518     This function Rotate Right (immediate) provides the value of the contents of a register rotated by a variable number of bits.
    519 
    520     \param [in]    value  Value to rotate
    521     \param [in]    value  Number of Bits to rotate
    522     \return               Rotated value
    523  */
    524 __attribute__( ( always_inline ) ) __STATIC_INLINE uint32_t __ROR(uint32_t op1, uint32_t op2)
    525 {
    526   return (op1 >> op2) | (op1 << (32 - op2));
    527 }
    528 
    529 
    530 /** \brief  Breakpoint
    531 
    532     This function causes the processor to enter Debug state.
    533     Debug tools can use this to investigate system state when the instruction at a particular address is reached.
    534 
    535     \param [in]    value  is ignored by the processor.
    536                    If required, a debugger can use it to store additional information about the breakpoint.
    537  */
    538 #define __BKPT(value)                       __ASM volatile ("bkpt "#value)
    539 
    540 
    541 #if       (__CORTEX_M >= 0x03) || (__CORTEX_SC >= 300)
    542 
    543 /** \brief  Reverse bit order of value
    544 
    545     This function reverses the bit order of the given value.
    546 
    547     \param [in]    value  Value to reverse
    548     \return               Reversed value
    549  */
    550 __attribute__( ( always_inline ) ) __STATIC_INLINE uint32_t __RBIT(uint32_t value)
    551 {
    552   uint32_t result;
    553 
    554    __ASM volatile ("rbit %0, %1" : "=r" (result) : "r" (value) );
    555    return(result);
    556 }
    557 
    558 
    559 /** \brief  LDR Exclusive (8 bit)
    560 
    561     This function executes a exclusive LDR instruction for 8 bit value.
    562 
    563     \param [in]    ptr  Pointer to data
    564     \return             value of type uint8_t at (*ptr)
    565  */
    566 __attribute__( ( always_inline ) ) __STATIC_INLINE uint8_t __LDREXB(volatile uint8_t *addr)
    567 {
    568     uint32_t result;
    569 
    570 #if (__GNUC__ > 4) || (__GNUC__ == 4 && __GNUC_MINOR__ >= 8)
    571    __ASM volatile ("ldrexb %0, %1" : "=r" (result) : "Q" (*addr) );
    572 #else
    573     /* Prior to GCC 4.8, "Q" will be expanded to [rx, #0] which is not
    574        accepted by assembler. So has to use following less efficient pattern.
    575     */
    576    __ASM volatile ("ldrexb %0, [%1]" : "=r" (result) : "r" (addr) : "memory" );
    577 #endif
    578    return ((uint8_t) result);    /* Add explicit type cast here */
    579 }
    580 
    581 
    582 /** \brief  LDR Exclusive (16 bit)
    583 
    584     This function executes a exclusive LDR instruction for 16 bit values.
    585 
    586     \param [in]    ptr  Pointer to data
    587     \return        value of type uint16_t at (*ptr)
    588  */
    589 __attribute__( ( always_inline ) ) __STATIC_INLINE uint16_t __LDREXH(volatile uint16_t *addr)
    590 {
    591     uint32_t result;
    592 
    593 #if (__GNUC__ > 4) || (__GNUC__ == 4 && __GNUC_MINOR__ >= 8)
    594    __ASM volatile ("ldrexh %0, %1" : "=r" (result) : "Q" (*addr) );
    595 #else
    596     /* Prior to GCC 4.8, "Q" will be expanded to [rx, #0] which is not
    597        accepted by assembler. So has to use following less efficient pattern.
    598     */
    599    __ASM volatile ("ldrexh %0, [%1]" : "=r" (result) : "r" (addr) : "memory" );
    600 #endif
    601    return ((uint16_t) result);    /* Add explicit type cast here */
    602 }
    603 
    604 
    605 /** \brief  LDR Exclusive (32 bit)
    606 
    607     This function executes a exclusive LDR instruction for 32 bit values.
    608 
    609     \param [in]    ptr  Pointer to data
    610     \return        value of type uint32_t at (*ptr)
    611  */
    612 __attribute__( ( always_inline ) ) __STATIC_INLINE uint32_t __LDREXW(volatile uint32_t *addr)
    613 {
    614     uint32_t result;
    615 
    616    __ASM volatile ("ldrex %0, %1" : "=r" (result) : "Q" (*addr) );
    617    return(result);
    618 }
    619 
    620 
    621 /** \brief  STR Exclusive (8 bit)
    622 
    623     This function executes a exclusive STR instruction for 8 bit values.
    624 
    625     \param [in]  value  Value to store
    626     \param [in]    ptr  Pointer to location
    627     \return          0  Function succeeded
    628     \return          1  Function failed
    629  */
    630 __attribute__( ( always_inline ) ) __STATIC_INLINE uint32_t __STREXB(uint8_t value, volatile uint8_t *addr)
    631 {
    632    uint32_t result;
    633 
    634    __ASM volatile ("strexb %0, %2, %1" : "=&r" (result), "=Q" (*addr) : "r" ((uint32_t)value) );
    635    return(result);
    636 }
    637 
    638 
    639 /** \brief  STR Exclusive (16 bit)
    640 
    641     This function executes a exclusive STR instruction for 16 bit values.
    642 
    643     \param [in]  value  Value to store
    644     \param [in]    ptr  Pointer to location
    645     \return          0  Function succeeded
    646     \return          1  Function failed
    647  */
    648 __attribute__( ( always_inline ) ) __STATIC_INLINE uint32_t __STREXH(uint16_t value, volatile uint16_t *addr)
    649 {
    650    uint32_t result;
    651 
    652    __ASM volatile ("strexh %0, %2, %1" : "=&r" (result), "=Q" (*addr) : "r" ((uint32_t)value) );
    653    return(result);
    654 }
    655 
    656 
    657 /** \brief  STR Exclusive (32 bit)
    658 
    659     This function executes a exclusive STR instruction for 32 bit values.
    660 
    661     \param [in]  value  Value to store
    662     \param [in]    ptr  Pointer to location
    663     \return          0  Function succeeded
    664     \return          1  Function failed
    665  */
    666 __attribute__( ( always_inline ) ) __STATIC_INLINE uint32_t __STREXW(uint32_t value, volatile uint32_t *addr)
    667 {
    668    uint32_t result;
    669 
    670    __ASM volatile ("strex %0, %2, %1" : "=&r" (result), "=Q" (*addr) : "r" (value) );
    671    return(result);
    672 }
    673 
    674 
    675 /** \brief  Remove the exclusive lock
    676 
    677     This function removes the exclusive lock which is created by LDREX.
    678 
    679  */
    680 __attribute__( ( always_inline ) ) __STATIC_INLINE void __CLREX(void)
    681 {
    682   __ASM volatile ("clrex" ::: "memory");
    683 }
    684 
    685 
    686 /** \brief  Signed Saturate
    687 
    688     This function saturates a signed value.
    689 
    690     \param [in]  value  Value to be saturated
    691     \param [in]    sat  Bit position to saturate to (1..32)
    692     \return             Saturated value
    693  */
    694 #define __SSAT(ARG1,ARG2) \
    695 ({                          \
    696   uint32_t __RES, __ARG1 = (ARG1); \
    697   __ASM ("ssat %0, %1, %2" : "=r" (__RES) :  "I" (ARG2), "r" (__ARG1) ); \
    698   __RES; \
    699  })
    700 
    701 
    702 /** \brief  Unsigned Saturate
    703 
    704     This function saturates an unsigned value.
    705 
    706     \param [in]  value  Value to be saturated
    707     \param [in]    sat  Bit position to saturate to (0..31)
    708     \return             Saturated value
    709  */
    710 #define __USAT(ARG1,ARG2) \
    711 ({                          \
    712   uint32_t __RES, __ARG1 = (ARG1); \
    713   __ASM ("usat %0, %1, %2" : "=r" (__RES) :  "I" (ARG2), "r" (__ARG1) ); \
    714   __RES; \
    715  })
    716 
    717 
    718 /** \brief  Count leading zeros
    719 
    720     This function counts the number of leading zeros of a data value.
    721 
    722     \param [in]  value  Value to count the leading zeros
    723     \return             number of leading zeros in value
    724  */
    725 __attribute__( ( always_inline ) ) __STATIC_INLINE uint8_t __CLZ(uint32_t value)
    726 {
    727   uint32_t result;
    728 
    729   __ASM volatile ("clz %0, %1" : "=r" (result) : "r" (value) );
    730    return ((uint8_t) result);    /* Add explicit type cast here */
    731 }
    732 
    733 
    734 /** \brief  Rotate Right with Extend (32 bit)
    735 
    736     This function moves each bit of a bitstring right by one bit. The carry input is shifted in at the left end of the bitstring.
    737 
    738     \param [in]    value  Value to rotate
    739     \return               Rotated value
    740  */
    741 __attribute__( ( always_inline ) ) __STATIC_INLINE uint32_t __RRX(uint32_t value)
    742 {
    743   uint32_t result;
    744 
    745   __ASM volatile ("rrx %0, %1" : __CMSIS_GCC_OUT_REG (result) : __CMSIS_GCC_USE_REG (value) );
    746   return(result);
    747 }
    748 
    749 
    750 /** \brief  LDRT Unprivileged (8 bit)
    751 
    752     This function executes a Unprivileged LDRT instruction for 8 bit value.
    753 
    754     \param [in]    ptr  Pointer to data
    755     \return             value of type uint8_t at (*ptr)
    756  */
    757 __attribute__( ( always_inline ) ) __STATIC_INLINE uint8_t __LDRBT(volatile uint8_t *addr)
    758 {
    759     uint32_t result;
    760 
    761 #if (__GNUC__ > 4) || (__GNUC__ == 4 && __GNUC_MINOR__ >= 8)
    762    __ASM volatile ("ldrbt %0, %1" : "=r" (result) : "Q" (*addr) );
    763 #else
    764     /* Prior to GCC 4.8, "Q" will be expanded to [rx, #0] which is not
    765        accepted by assembler. So has to use following less efficient pattern.
    766     */
    767    __ASM volatile ("ldrbt %0, [%1]" : "=r" (result) : "r" (addr) : "memory" );
    768 #endif
    769    return ((uint8_t) result);    /* Add explicit type cast here */
    770 }
    771 
    772 
    773 /** \brief  LDRT Unprivileged (16 bit)
    774 
    775     This function executes a Unprivileged LDRT instruction for 16 bit values.
    776 
    777     \param [in]    ptr  Pointer to data
    778     \return        value of type uint16_t at (*ptr)
    779  */
    780 __attribute__( ( always_inline ) ) __STATIC_INLINE uint16_t __LDRHT(volatile uint16_t *addr)
    781 {
    782     uint32_t result;
    783 
    784 #if (__GNUC__ > 4) || (__GNUC__ == 4 && __GNUC_MINOR__ >= 8)
    785    __ASM volatile ("ldrht %0, %1" : "=r" (result) : "Q" (*addr) );
    786 #else
    787     /* Prior to GCC 4.8, "Q" will be expanded to [rx, #0] which is not
    788        accepted by assembler. So has to use following less efficient pattern.
    789     */
    790    __ASM volatile ("ldrht %0, [%1]" : "=r" (result) : "r" (addr) : "memory" );
    791 #endif
    792    return ((uint16_t) result);    /* Add explicit type cast here */
    793 }
    794 
    795 
    796 /** \brief  LDRT Unprivileged (32 bit)
    797 
    798     This function executes a Unprivileged LDRT instruction for 32 bit values.
    799 
    800     \param [in]    ptr  Pointer to data
    801     \return        value of type uint32_t at (*ptr)
    802  */
    803 __attribute__( ( always_inline ) ) __STATIC_INLINE uint32_t __LDRT(volatile uint32_t *addr)
    804 {
    805     uint32_t result;
    806 
    807    __ASM volatile ("ldrt %0, %1" : "=r" (result) : "Q" (*addr) );
    808    return(result);
    809 }
    810 
    811 
    812 /** \brief  STRT Unprivileged (8 bit)
    813 
    814     This function executes a Unprivileged STRT instruction for 8 bit values.
    815 
    816     \param [in]  value  Value to store
    817     \param [in]    ptr  Pointer to location
    818  */
    819 __attribute__( ( always_inline ) ) __STATIC_INLINE void __STRBT(uint8_t value, volatile uint8_t *addr)
    820 {
    821    __ASM volatile ("strbt %1, %0" : "=Q" (*addr) : "r" ((uint32_t)value) );
    822 }
    823 
    824 
    825 /** \brief  STRT Unprivileged (16 bit)
    826 
    827     This function executes a Unprivileged STRT instruction for 16 bit values.
    828 
    829     \param [in]  value  Value to store
    830     \param [in]    ptr  Pointer to location
    831  */
    832 __attribute__( ( always_inline ) ) __STATIC_INLINE void __STRHT(uint16_t value, volatile uint16_t *addr)
    833 {
    834    __ASM volatile ("strht %1, %0" : "=Q" (*addr) : "r" ((uint32_t)value) );
    835 }
    836 
    837 
    838 /** \brief  STRT Unprivileged (32 bit)
    839 
    840     This function executes a Unprivileged STRT instruction for 32 bit values.
    841 
    842     \param [in]  value  Value to store
    843     \param [in]    ptr  Pointer to location
    844  */
    845 __attribute__( ( always_inline ) ) __STATIC_INLINE void __STRT(uint32_t value, volatile uint32_t *addr)
    846 {
    847    __ASM volatile ("strt %1, %0" : "=Q" (*addr) : "r" (value) );
    848 }
    849 
    850 #endif /* (__CORTEX_M >= 0x03) || (__CORTEX_SC >= 300) */
    851 
    852 
    853 #elif defined ( __ICCARM__ ) /*------------------ ICC Compiler -------------------*/
    854 /* IAR iccarm specific functions */
    855 #include <cmsis_iar.h>
    856 
    857 
    858 #elif defined ( __TMS470__ ) /*---------------- TI CCS Compiler ------------------*/
    859 /* TI CCS specific functions */
    860 #include <cmsis_ccs.h>
    861 
    862 
    863 #elif defined ( __TASKING__ ) /*------------------ TASKING Compiler --------------*/
    864 /* TASKING carm specific functions */
    865 /*
    866  * The CMSIS functions have been implemented as intrinsics in the compiler.
    867  * Please use "carm -?i" to get an up to date list of all intrinsics,
    868  * Including the CMSIS ones.
    869  */
    870 
    871 
    872 #elif defined ( __CSMC__ ) /*------------------ COSMIC Compiler -------------------*/
    873 /* Cosmic specific functions */
    874 #include <cmsis_csm.h>
    875 
    876 #endif
    877 
    878 /*@}*/ /* end of group CMSIS_Core_InstructionInterface */
    879 
    880 #endif /* __CORE_CMINSTR_H */
    881