29 #pragma GCC diagnostic push 30 #pragma GCC diagnostic ignored "-Wsign-conversion" 31 #pragma GCC diagnostic ignored "-Wconversion" 32 #pragma GCC diagnostic ignored "-Wunused-parameter" 36 #define __has_builtin(x) (0) 44 #define __INLINE inline 46 #ifndef __STATIC_INLINE 47 #define __STATIC_INLINE static inline 49 #ifndef __STATIC_FORCEINLINE 50 #define __STATIC_FORCEINLINE __attribute__((always_inline)) static inline 53 #define __NO_RETURN __attribute__((__noreturn__)) 56 #define __USED __attribute__((used)) 59 #define __WEAK __attribute__((weak)) 62 #define __PACKED __attribute__((packed, aligned(1))) 64 #ifndef __PACKED_STRUCT 65 #define __PACKED_STRUCT struct __attribute__((packed, aligned(1))) 67 #ifndef __PACKED_UNION 68 #define __PACKED_UNION union __attribute__((packed, aligned(1))) 70 #ifndef __UNALIGNED_UINT32 71 #pragma GCC diagnostic push 72 #pragma GCC diagnostic ignored "-Wpacked" 73 #pragma GCC diagnostic ignored "-Wattributes" 74 struct __attribute__((packed))
T_UINT32 { uint32_t v; };
75 #pragma GCC diagnostic pop 76 #define __UNALIGNED_UINT32(x) (((struct T_UINT32 *)(x))->v) 78 #ifndef __UNALIGNED_UINT16_WRITE 79 #pragma GCC diagnostic push 80 #pragma GCC diagnostic ignored "-Wpacked" 81 #pragma GCC diagnostic ignored "-Wattributes" 83 #pragma GCC diagnostic pop 84 #define __UNALIGNED_UINT16_WRITE(addr, val) (void)((((struct T_UINT16_WRITE *)(void *)(addr))->v) = (val)) 86 #ifndef __UNALIGNED_UINT16_READ 87 #pragma GCC diagnostic push 88 #pragma GCC diagnostic ignored "-Wpacked" 89 #pragma GCC diagnostic ignored "-Wattributes" 91 #pragma GCC diagnostic pop 92 #define __UNALIGNED_UINT16_READ(addr) (((const struct T_UINT16_READ *)(const void *)(addr))->v) 94 #ifndef __UNALIGNED_UINT32_WRITE 95 #pragma GCC diagnostic push 96 #pragma GCC diagnostic ignored "-Wpacked" 97 #pragma GCC diagnostic ignored "-Wattributes" 99 #pragma GCC diagnostic pop 100 #define __UNALIGNED_UINT32_WRITE(addr, val) (void)((((struct T_UINT32_WRITE *)(void *)(addr))->v) = (val)) 102 #ifndef __UNALIGNED_UINT32_READ 103 #pragma GCC diagnostic push 104 #pragma GCC diagnostic ignored "-Wpacked" 105 #pragma GCC diagnostic ignored "-Wattributes" 107 #pragma GCC diagnostic pop 108 #define __UNALIGNED_UINT32_READ(addr) (((const struct T_UINT32_READ *)(const void *)(addr))->v) 111 #define __ALIGNED(x) __attribute__((aligned(x))) 114 #define __RESTRICT __restrict 131 __ASM volatile (
"cpsie i" : : :
"memory");
142 __ASM volatile (
"cpsid i" : : :
"memory");
155 __ASM volatile (
"MRS %0, control" :
"=r" (result) );
160 #if (defined (__ARM_FEATURE_CMSE ) && (__ARM_FEATURE_CMSE == 3)) 170 __ASM volatile (
"MRS %0, control_ns" :
"=r" (result) );
183 __ASM volatile (
"MSR control, %0" : :
"r" (control) :
"memory");
187 #if (defined (__ARM_FEATURE_CMSE ) && (__ARM_FEATURE_CMSE == 3)) 195 __ASM volatile (
"MSR control_ns, %0" : :
"r" (control) :
"memory");
209 __ASM volatile (
"MRS %0, ipsr" :
"=r" (result) );
223 __ASM volatile (
"MRS %0, apsr" :
"=r" (result) );
237 __ASM volatile (
"MRS %0, xpsr" :
"=r" (result) );
249 register uint32_t result;
251 __ASM volatile (
"MRS %0, psp" :
"=r" (result) );
256 #if (defined (__ARM_FEATURE_CMSE ) && (__ARM_FEATURE_CMSE == 3)) 264 register uint32_t result;
266 __ASM volatile (
"MRS %0, psp_ns" :
"=r" (result) );
279 __ASM volatile (
"MSR psp, %0" : :
"r" (topOfProcStack) : );
283 #if (defined (__ARM_FEATURE_CMSE ) && (__ARM_FEATURE_CMSE == 3)) 291 __ASM volatile (
"MSR psp_ns, %0" : :
"r" (topOfProcStack) : );
303 register uint32_t result;
305 __ASM volatile (
"MRS %0, msp" :
"=r" (result) );
310 #if (defined (__ARM_FEATURE_CMSE ) && (__ARM_FEATURE_CMSE == 3)) 318 register uint32_t result;
320 __ASM volatile (
"MRS %0, msp_ns" :
"=r" (result) );
333 __ASM volatile (
"MSR msp, %0" : :
"r" (topOfMainStack) : );
337 #if (defined (__ARM_FEATURE_CMSE ) && (__ARM_FEATURE_CMSE == 3)) 345 __ASM volatile (
"MSR msp_ns, %0" : :
"r" (topOfMainStack) : );
350 #if (defined (__ARM_FEATURE_CMSE ) && (__ARM_FEATURE_CMSE == 3)) 358 register uint32_t result;
360 __ASM volatile (
"MRS %0, sp_ns" :
"=r" (result) );
372 __ASM volatile (
"MSR sp_ns, %0" : :
"r" (topOfStack) : );
386 __ASM volatile (
"MRS %0, primask" :
"=r" (result) ::
"memory");
391 #if (defined (__ARM_FEATURE_CMSE ) && (__ARM_FEATURE_CMSE == 3)) 401 __ASM volatile (
"MRS %0, primask_ns" :
"=r" (result) ::
"memory");
414 __ASM volatile (
"MSR primask, %0" : :
"r" (priMask) :
"memory");
418 #if (defined (__ARM_FEATURE_CMSE ) && (__ARM_FEATURE_CMSE == 3)) 426 __ASM volatile (
"MSR primask_ns, %0" : :
"r" (priMask) :
"memory");
431 #if ((defined (__ARM_ARCH_7M__ ) && (__ARM_ARCH_7M__ == 1)) || \ 432 (defined (__ARM_ARCH_7EM__ ) && (__ARM_ARCH_7EM__ == 1)) || \ 433 (defined (__ARM_ARCH_8M_MAIN__ ) && (__ARM_ARCH_8M_MAIN__ == 1)) ) 441 __ASM volatile (
"cpsie f" : : :
"memory");
452 __ASM volatile (
"cpsid f" : : :
"memory");
465 __ASM volatile (
"MRS %0, basepri" :
"=r" (result) );
470 #if (defined (__ARM_FEATURE_CMSE ) && (__ARM_FEATURE_CMSE == 3)) 480 __ASM volatile (
"MRS %0, basepri_ns" :
"=r" (result) );
493 __ASM volatile (
"MSR basepri, %0" : :
"r" (basePri) :
"memory");
497 #if (defined (__ARM_FEATURE_CMSE ) && (__ARM_FEATURE_CMSE == 3)) 505 __ASM volatile (
"MSR basepri_ns, %0" : :
"r" (basePri) :
"memory");
518 __ASM volatile (
"MSR basepri_max, %0" : :
"r" (basePri) :
"memory");
531 __ASM volatile (
"MRS %0, faultmask" :
"=r" (result) );
536 #if (defined (__ARM_FEATURE_CMSE ) && (__ARM_FEATURE_CMSE == 3)) 546 __ASM volatile (
"MRS %0, faultmask_ns" :
"=r" (result) );
559 __ASM volatile (
"MSR faultmask, %0" : :
"r" (faultMask) :
"memory");
563 #if (defined (__ARM_FEATURE_CMSE ) && (__ARM_FEATURE_CMSE == 3)) 571 __ASM volatile (
"MSR faultmask_ns, %0" : :
"r" (faultMask) :
"memory");
580 #if ((defined (__ARM_ARCH_8M_MAIN__ ) && (__ARM_ARCH_8M_MAIN__ == 1)) || \ 581 (defined (__ARM_ARCH_8M_BASE__ ) && (__ARM_ARCH_8M_BASE__ == 1)) ) 594 #if (!(defined (__ARM_ARCH_8M_MAIN__ ) && (__ARM_ARCH_8M_MAIN__ == 1)) && \ 595 (!defined (__ARM_FEATURE_CMSE) || (__ARM_FEATURE_CMSE < 3))) 599 register uint32_t result;
600 __ASM volatile (
"MRS %0, psplim" :
"=r" (result) );
605 #if (defined (__ARM_FEATURE_CMSE) && (__ARM_FEATURE_CMSE == 3)) 616 #if (!(defined (__ARM_ARCH_8M_MAIN__ ) && (__ARM_ARCH_8M_MAIN__ == 1))) 620 register uint32_t result;
621 __ASM volatile (
"MRS %0, psplim_ns" :
"=r" (result) );
639 #if (!(defined (__ARM_ARCH_8M_MAIN__ ) && (__ARM_ARCH_8M_MAIN__ == 1)) && \ 640 (!defined (__ARM_FEATURE_CMSE) || (__ARM_FEATURE_CMSE < 3))) 642 (void)ProcStackPtrLimit;
644 __ASM volatile (
"MSR psplim, %0" : :
"r" (ProcStackPtrLimit));
649 #if (defined (__ARM_FEATURE_CMSE ) && (__ARM_FEATURE_CMSE == 3)) 660 #if (!(defined (__ARM_ARCH_8M_MAIN__ ) && (__ARM_ARCH_8M_MAIN__ == 1))) 662 (void)ProcStackPtrLimit;
664 __ASM volatile (
"MSR psplim_ns, %0\n" : :
"r" (ProcStackPtrLimit));
681 #if (!(defined (__ARM_ARCH_8M_MAIN__ ) && (__ARM_ARCH_8M_MAIN__ == 1)) && \ 682 (!defined (__ARM_FEATURE_CMSE) || (__ARM_FEATURE_CMSE < 3))) 686 register uint32_t result;
687 __ASM volatile (
"MRS %0, msplim" :
"=r" (result) );
693 #if (defined (__ARM_FEATURE_CMSE ) && (__ARM_FEATURE_CMSE == 3)) 704 #if (!(defined (__ARM_ARCH_8M_MAIN__ ) && (__ARM_ARCH_8M_MAIN__ == 1))) 708 register uint32_t result;
709 __ASM volatile (
"MRS %0, msplim_ns" :
"=r" (result) );
727 #if (!(defined (__ARM_ARCH_8M_MAIN__ ) && (__ARM_ARCH_8M_MAIN__ == 1)) && \ 728 (!defined (__ARM_FEATURE_CMSE) || (__ARM_FEATURE_CMSE < 3))) 730 (void)MainStackPtrLimit;
732 __ASM volatile (
"MSR msplim, %0" : :
"r" (MainStackPtrLimit));
737 #if (defined (__ARM_FEATURE_CMSE ) && (__ARM_FEATURE_CMSE == 3)) 748 #if (!(defined (__ARM_ARCH_8M_MAIN__ ) && (__ARM_ARCH_8M_MAIN__ == 1))) 750 (void)MainStackPtrLimit;
752 __ASM volatile (
"MSR msplim_ns, %0" : :
"r" (MainStackPtrLimit));
761 #if ((defined (__ARM_ARCH_7EM__ ) && (__ARM_ARCH_7EM__ == 1)) || \ 762 (defined (__ARM_ARCH_8M_MAIN__ ) && (__ARM_ARCH_8M_MAIN__ == 1)) ) 771 #if ((defined (__FPU_PRESENT) && (__FPU_PRESENT == 1U)) && \ 772 (defined (__FPU_USED ) && (__FPU_USED == 1U)) ) 773 #if __has_builtin(__builtin_arm_get_fpscr) || (__GNUC__ > 7) || (__GNUC__ == 7 && __GNUC_MINOR__ >= 2) 775 return __builtin_arm_get_fpscr();
779 __ASM volatile (
"VMRS %0, fpscr" :
"=r" (result) );
795 #if ((defined (__FPU_PRESENT) && (__FPU_PRESENT == 1U)) && \ 796 (defined (__FPU_USED ) && (__FPU_USED == 1U)) ) 797 #if __has_builtin(__builtin_arm_set_fpscr) || (__GNUC__ > 7) || (__GNUC__ == 7 && __GNUC_MINOR__ >= 2) 799 __builtin_arm_set_fpscr(fpscr);
801 __ASM volatile (
"VMSR fpscr, %0" : :
"r" (fpscr) :
"vfpcc",
"memory");
825 #if defined (__thumb__) && !defined (__thumb2__) 826 #define __CMSIS_GCC_OUT_REG(r) "=l" (r) 827 #define __CMSIS_GCC_RW_REG(r) "+l" (r) 828 #define __CMSIS_GCC_USE_REG(r) "l" (r) 830 #define __CMSIS_GCC_OUT_REG(r) "=r" (r) 831 #define __CMSIS_GCC_RW_REG(r) "+r" (r) 832 #define __CMSIS_GCC_USE_REG(r) "r" (r) 839 #define __NOP() __ASM volatile ("nop") 845 #define __WFI() __ASM volatile ("wfi") 853 #define __WFE() __ASM volatile ("wfe") 860 #define __SEV() __ASM volatile ("sev") 871 __ASM volatile (
"isb 0xF":::
"memory");
882 __ASM volatile (
"dsb 0xF":::
"memory");
893 __ASM volatile (
"dmb 0xF":::
"memory");
905 #if (__GNUC__ > 4) || (__GNUC__ == 4 && __GNUC_MINOR__ >= 5) 906 return __builtin_bswap32(value);
939 #if (__GNUC__ > 4) || (__GNUC__ == 4 && __GNUC_MINOR__ >= 8) 940 return (int16_t)__builtin_bswap16(value);
964 return (op1 >> op2) | (op1 << (32U - op2));
975 #define __BKPT(value) __ASM volatile ("bkpt "#value) 988 #if ((defined (__ARM_ARCH_7M__ ) && (__ARM_ARCH_7M__ == 1)) || \ 989 (defined (__ARM_ARCH_7EM__ ) && (__ARM_ARCH_7EM__ == 1)) || \ 990 (defined (__ARM_ARCH_8M_MAIN__ ) && (__ARM_ARCH_8M_MAIN__ == 1)) ) 991 __ASM volatile (
"rbit %0, %1" :
"=r" (result) :
"r" (value) );
993 uint32_t s = (4U * 8U) - 1U;
996 for (value >>= 1U; value != 0U; value >>= 1U)
999 result |= value & 1U;
1014 #define __CLZ (uint8_t)__builtin_clz 1017 #if ((defined (__ARM_ARCH_7M__ ) && (__ARM_ARCH_7M__ == 1)) || \ 1018 (defined (__ARM_ARCH_7EM__ ) && (__ARM_ARCH_7EM__ == 1)) || \ 1019 (defined (__ARM_ARCH_8M_MAIN__ ) && (__ARM_ARCH_8M_MAIN__ == 1)) || \ 1020 (defined (__ARM_ARCH_8M_BASE__ ) && (__ARM_ARCH_8M_BASE__ == 1)) ) 1031 #if (__GNUC__ > 4) || (__GNUC__ == 4 && __GNUC_MINOR__ >= 8) 1032 __ASM volatile (
"ldrexb %0, %1" :
"=r" (result) :
"Q" (*addr) );
1037 __ASM volatile (
"ldrexb %0, [%1]" :
"=r" (result) :
"r" (addr) :
"memory" );
1039 return ((uint8_t) result);
1053 #if (__GNUC__ > 4) || (__GNUC__ == 4 && __GNUC_MINOR__ >= 8) 1054 __ASM volatile (
"ldrexh %0, %1" :
"=r" (result) :
"Q" (*addr) );
1059 __ASM volatile (
"ldrexh %0, [%1]" :
"=r" (result) :
"r" (addr) :
"memory" );
1061 return ((uint16_t) result);
1075 __ASM volatile (
"ldrex %0, %1" :
"=r" (result) :
"Q" (*addr) );
1092 __ASM volatile (
"strexb %0, %2, %1" :
"=&r" (result),
"=Q" (*addr) :
"r" ((uint32_t)value) );
1109 __ASM volatile (
"strexh %0, %2, %1" :
"=&r" (result),
"=Q" (*addr) :
"r" ((uint32_t)value) );
1126 __ASM volatile (
"strex %0, %2, %1" :
"=&r" (result),
"=Q" (*addr) :
"r" (value) );
1137 __ASM volatile (
"clrex" :::
"memory");
1146 #if ((defined (__ARM_ARCH_7M__ ) && (__ARM_ARCH_7M__ == 1)) || \ 1147 (defined (__ARM_ARCH_7EM__ ) && (__ARM_ARCH_7EM__ == 1)) || \ 1148 (defined (__ARM_ARCH_8M_MAIN__ ) && (__ARM_ARCH_8M_MAIN__ == 1)) ) 1156 #define __SSAT(ARG1,ARG2) \ 1159 int32_t __RES, __ARG1 = (ARG1); \ 1160 __ASM ("ssat %0, %1, %2" : "=r" (__RES) : "I" (ARG2), "r" (__ARG1) ); \ 1172 #define __USAT(ARG1,ARG2) \ 1175 uint32_t __RES, __ARG1 = (ARG1); \ 1176 __ASM ("usat %0, %1, %2" : "=r" (__RES) : "I" (ARG2), "r" (__ARG1) ); \ 1207 #if (__GNUC__ > 4) || (__GNUC__ == 4 && __GNUC_MINOR__ >= 8) 1208 __ASM volatile (
"ldrbt %0, %1" :
"=r" (result) :
"Q" (*ptr) );
1213 __ASM volatile (
"ldrbt %0, [%1]" :
"=r" (result) :
"r" (ptr) :
"memory" );
1215 return ((uint8_t) result);
1229 #if (__GNUC__ > 4) || (__GNUC__ == 4 && __GNUC_MINOR__ >= 8) 1230 __ASM volatile (
"ldrht %0, %1" :
"=r" (result) :
"Q" (*ptr) );
1235 __ASM volatile (
"ldrht %0, [%1]" :
"=r" (result) :
"r" (ptr) :
"memory" );
1237 return ((uint16_t) result);
1251 __ASM volatile (
"ldrt %0, %1" :
"=r" (result) :
"Q" (*ptr) );
1264 __ASM volatile (
"strbt %1, %0" :
"=Q" (*ptr) :
"r" ((uint32_t)value) );
1276 __ASM volatile (
"strht %1, %0" :
"=Q" (*ptr) :
"r" ((uint32_t)value) );
1288 __ASM volatile (
"strt %1, %0" :
"=Q" (*ptr) :
"r" (value) );
1304 if ((sat >= 1U) && (sat <= 32U))
1306 const int32_t max = (int32_t)((1U << (sat - 1U)) - 1U);
1307 const int32_t min = -1 - max ;
1331 const uint32_t max = ((1U << sat) - 1U);
1332 if (val > (int32_t)max)
1341 return (uint32_t)val;
1349 #if ((defined (__ARM_ARCH_8M_MAIN__ ) && (__ARM_ARCH_8M_MAIN__ == 1)) || \ 1350 (defined (__ARM_ARCH_8M_BASE__ ) && (__ARM_ARCH_8M_BASE__ == 1)) ) 1361 __ASM volatile (
"ldab %0, %1" :
"=r" (result) :
"Q" (*ptr) );
1362 return ((uint8_t) result);
1376 __ASM volatile (
"ldah %0, %1" :
"=r" (result) :
"Q" (*ptr) );
1377 return ((uint16_t) result);
1391 __ASM volatile (
"lda %0, %1" :
"=r" (result) :
"Q" (*ptr) );
1404 __ASM volatile (
"stlb %1, %0" :
"=Q" (*ptr) :
"r" ((uint32_t)value) );
1416 __ASM volatile (
"stlh %1, %0" :
"=Q" (*ptr) :
"r" ((uint32_t)value) );
1428 __ASM volatile (
"stl %1, %0" :
"=Q" (*ptr) :
"r" ((uint32_t)value) );
1442 __ASM volatile (
"ldaexb %0, %1" :
"=r" (result) :
"Q" (*ptr) );
1443 return ((uint8_t) result);
1457 __ASM volatile (
"ldaexh %0, %1" :
"=r" (result) :
"Q" (*ptr) );
1458 return ((uint16_t) result);
1472 __ASM volatile (
"ldaex %0, %1" :
"=r" (result) :
"Q" (*ptr) );
1489 __ASM volatile (
"stlexb %0, %2, %1" :
"=&r" (result),
"=Q" (*ptr) :
"r" ((uint32_t)value) );
1506 __ASM volatile (
"stlexh %0, %2, %1" :
"=&r" (result),
"=Q" (*ptr) :
"r" ((uint32_t)value) );
1523 __ASM volatile (
"stlex %0, %2, %1" :
"=&r" (result),
"=Q" (*ptr) :
"r" ((uint32_t)value) );
1539 #if (defined (__ARM_FEATURE_DSP) && (__ARM_FEATURE_DSP == 1)) 1545 __ASM volatile (
"sadd8 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1553 __ASM volatile (
"qadd8 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1561 __ASM volatile (
"shadd8 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1569 __ASM volatile (
"uadd8 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1577 __ASM volatile (
"uqadd8 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1585 __ASM volatile (
"uhadd8 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1594 __ASM volatile (
"ssub8 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1602 __ASM volatile (
"qsub8 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1610 __ASM volatile (
"shsub8 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1618 __ASM volatile (
"usub8 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1626 __ASM volatile (
"uqsub8 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1634 __ASM volatile (
"uhsub8 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1643 __ASM volatile (
"sadd16 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1651 __ASM volatile (
"qadd16 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1659 __ASM volatile (
"shadd16 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1667 __ASM volatile (
"uadd16 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1675 __ASM volatile (
"uqadd16 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1683 __ASM volatile (
"uhadd16 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1691 __ASM volatile (
"ssub16 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1699 __ASM volatile (
"qsub16 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1707 __ASM volatile (
"shsub16 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1715 __ASM volatile (
"usub16 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1723 __ASM volatile (
"uqsub16 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1731 __ASM volatile (
"uhsub16 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1739 __ASM volatile (
"sasx %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1747 __ASM volatile (
"qasx %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1755 __ASM volatile (
"shasx %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1763 __ASM volatile (
"uasx %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1771 __ASM volatile (
"uqasx %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1779 __ASM volatile (
"uhasx %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1787 __ASM volatile (
"ssax %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1795 __ASM volatile (
"qsax %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1803 __ASM volatile (
"shsax %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1811 __ASM volatile (
"usax %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1819 __ASM volatile (
"uqsax %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1827 __ASM volatile (
"uhsax %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1835 __ASM volatile (
"usad8 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1843 __ASM volatile (
"usada8 %0, %1, %2, %3" :
"=r" (result) :
"r" (op1),
"r" (op2),
"r" (op3) );
1847 #define __SSAT16(ARG1,ARG2) \ 1849 int32_t __RES, __ARG1 = (ARG1); \ 1850 __ASM ("ssat16 %0, %1, %2" : "=r" (__RES) : "I" (ARG2), "r" (__ARG1) ); \ 1854 #define __USAT16(ARG1,ARG2) \ 1856 uint32_t __RES, __ARG1 = (ARG1); \ 1857 __ASM ("usat16 %0, %1, %2" : "=r" (__RES) : "I" (ARG2), "r" (__ARG1) ); \ 1865 __ASM volatile (
"uxtb16 %0, %1" :
"=r" (result) :
"r" (op1));
1873 __ASM volatile (
"uxtab16 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1881 __ASM volatile (
"sxtb16 %0, %1" :
"=r" (result) :
"r" (op1));
1889 __ASM volatile (
"sxtab16 %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1897 __ASM volatile (
"smuad %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1905 __ASM volatile (
"smuadx %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1913 __ASM volatile (
"smlad %0, %1, %2, %3" :
"=r" (result) :
"r" (op1),
"r" (op2),
"r" (op3) );
1921 __ASM volatile (
"smladx %0, %1, %2, %3" :
"=r" (result) :
"r" (op1),
"r" (op2),
"r" (op3) );
1934 __ASM volatile (
"smlald %0, %1, %2, %3" :
"=r" (llr.w32[0]),
"=r" (llr.w32[1]):
"r" (op1),
"r" (op2) ,
"0" (llr.w32[0]),
"1" (llr.w32[1]) );
1936 __ASM volatile (
"smlald %0, %1, %2, %3" :
"=r" (llr.w32[1]),
"=r" (llr.w32[0]):
"r" (op1),
"r" (op2) ,
"0" (llr.w32[1]),
"1" (llr.w32[0]) );
1951 __ASM volatile (
"smlaldx %0, %1, %2, %3" :
"=r" (llr.w32[0]),
"=r" (llr.w32[1]):
"r" (op1),
"r" (op2) ,
"0" (llr.w32[0]),
"1" (llr.w32[1]) );
1953 __ASM volatile (
"smlaldx %0, %1, %2, %3" :
"=r" (llr.w32[1]),
"=r" (llr.w32[0]):
"r" (op1),
"r" (op2) ,
"0" (llr.w32[1]),
"1" (llr.w32[0]) );
1963 __ASM volatile (
"smusd %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1971 __ASM volatile (
"smusdx %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
1979 __ASM volatile (
"smlsd %0, %1, %2, %3" :
"=r" (result) :
"r" (op1),
"r" (op2),
"r" (op3) );
1987 __ASM volatile (
"smlsdx %0, %1, %2, %3" :
"=r" (result) :
"r" (op1),
"r" (op2),
"r" (op3) );
2000 __ASM volatile (
"smlsld %0, %1, %2, %3" :
"=r" (llr.w32[0]),
"=r" (llr.w32[1]):
"r" (op1),
"r" (op2) ,
"0" (llr.w32[0]),
"1" (llr.w32[1]) );
2002 __ASM volatile (
"smlsld %0, %1, %2, %3" :
"=r" (llr.w32[1]),
"=r" (llr.w32[0]):
"r" (op1),
"r" (op2) ,
"0" (llr.w32[1]),
"1" (llr.w32[0]) );
2017 __ASM volatile (
"smlsldx %0, %1, %2, %3" :
"=r" (llr.w32[0]),
"=r" (llr.w32[1]):
"r" (op1),
"r" (op2) ,
"0" (llr.w32[0]),
"1" (llr.w32[1]) );
2019 __ASM volatile (
"smlsldx %0, %1, %2, %3" :
"=r" (llr.w32[1]),
"=r" (llr.w32[0]):
"r" (op1),
"r" (op2) ,
"0" (llr.w32[1]),
"1" (llr.w32[0]) );
2029 __ASM volatile (
"sel %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
2037 __ASM volatile (
"qadd %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
2045 __ASM volatile (
"qsub %0, %1, %2" :
"=r" (result) :
"r" (op1),
"r" (op2) );
2050 #define __PKHBT(ARG1,ARG2,ARG3) \ 2052 uint32_t __RES, __ARG1 = (ARG1), __ARG2 = (ARG2); \ 2053 __ASM ("pkhbt %0, %1, %2, lsl %3" : "=r" (__RES) : "r" (__ARG1), "r" (__ARG2), "I" (ARG3) ); \ 2057 #define __PKHTB(ARG1,ARG2,ARG3) \ 2059 uint32_t __RES, __ARG1 = (ARG1), __ARG2 = (ARG2); \ 2061 __ASM ("pkhtb %0, %1, %2" : "=r" (__RES) : "r" (__ARG1), "r" (__ARG2) ); \ 2063 __ASM ("pkhtb %0, %1, %2, asr %3" : "=r" (__RES) : "r" (__ARG1), "r" (__ARG2), "I" (ARG3) ); \ 2068 #define __PKHBT(ARG1,ARG2,ARG3) ( ((((uint32_t)(ARG1)) ) & 0x0000FFFFUL) | \ 2069 ((((uint32_t)(ARG2)) << (ARG3)) & 0xFFFF0000UL) ) 2071 #define __PKHTB(ARG1,ARG2,ARG3) ( ((((uint32_t)(ARG1)) ) & 0xFFFF0000UL) | \ 2072 ((((uint32_t)(ARG2)) >> (ARG3)) & 0x0000FFFFUL) ) 2078 __ASM volatile (
"smmla %0, %1, %2, %3" :
"=r" (result):
"r" (op1),
"r" (op2),
"r" (op3) );
2086 #pragma GCC diagnostic pop __STATIC_INLINE uint32_t __get_CONTROL(void)
Enable IRQ Interrupts.
#define __CMSIS_GCC_USE_REG(r)
__STATIC_INLINE uint32_t __get_IPSR(void)
Get IPSR Register.
__STATIC_INLINE uint32_t __USAT(int32_t val, uint32_t sat)
Unsigned Saturate.
__STATIC_INLINE uint32_t __get_PSP(void)
Get Process Stack Pointer.
__STATIC_INLINE void __set_PSP(uint32_t topOfProcStack)
Set Process Stack Pointer.
__STATIC_FORCEINLINE void __DMB(void)
Data Memory Barrier.
__PACKED_STRUCT T_UINT16_WRITE
__STATIC_INLINE uint32_t __get_xPSR(void)
Get xPSR Register.
__STATIC_INLINE __ASM int16_t __REVSH(int16_t value)
Reverse byte order (16 bit)
__STATIC_INLINE void __set_MSP(uint32_t topOfMainStack)
Set Main Stack Pointer.
__PACKED_STRUCT T_UINT32_WRITE
__STATIC_FORCEINLINE void __DSB(void)
Data Synchronization Barrier.
__STATIC_INLINE void __set_PRIMASK(uint32_t priMask)
Set Priority Mask.
__STATIC_INLINE __ASM uint32_t __REV16(uint32_t value)
Reverse byte order (16 bit)
__STATIC_FORCEINLINE void __disable_irq(void)
Disable IRQ Interrupts.
#define __STATIC_FORCEINLINE
__STATIC_INLINE uint32_t __RBIT(uint32_t value)
Reverse bit order of value.
__STATIC_INLINE uint32_t __get_APSR(void)
Get APSR Register.
__STATIC_INLINE uint32_t __get_PRIMASK(void)
Get Priority Mask.
__STATIC_INLINE uint32_t __get_MSP(void)
Get Main Stack Pointer.
#define __ROR
Rotate Right in unsigned value (32 bit)
__PACKED_STRUCT T_UINT32_READ
#define __CMSIS_GCC_OUT_REG(r)
__STATIC_INLINE void __set_CONTROL(uint32_t control)
Set Control Register.
__STATIC_FORCEINLINE void __enable_irq(void)
Enable IRQ Interrupts.
__IAR_FT uint32_t __LDREXW(uint32_t volatile *ptr)
__STATIC_INLINE int32_t __SSAT(int32_t val, uint32_t sat)
Signed Saturate.
__PACKED_STRUCT T_UINT16_READ
__STATIC_FORCEINLINE void __ISB(void)
Instruction Synchronization Barrier.
__IAR_FT uint32_t __STREXW(uint32_t value, uint32_t volatile *ptr)
__STATIC_FORCEINLINE uint32_t __REV(uint32_t value)
Reverse byte order (32 bit)