| /* Shared RTM header. Emulate TSX intrinsics for compilers and assemblers |
| that do not support the intrinsics and instructions yet. */ |
| #ifndef _HLE_H |
| #define _HLE_H 1 |
| |
| #ifdef __ASSEMBLER__ |
| |
| .macro XBEGIN target |
| .byte 0xc7,0xf8 |
| .long \target-1f |
| 1: |
| .endm |
| |
| .macro XEND |
| .byte 0x0f,0x01,0xd5 |
| .endm |
| |
| .macro XABORT code |
| .byte 0xc6,0xf8,\code |
| .endm |
| |
| .macro XTEST |
| .byte 0x0f,0x01,0xd6 |
| .endm |
| |
| #endif |
| |
| /* Official RTM intrinsics interface matching gcc/icc, but works |
| on older gcc compatible compilers and binutils. |
| We should somehow detect if the compiler supports it, because |
| it may be able to generate slightly better code. */ |
| |
| #define _XBEGIN_STARTED (~0u) |
| #define _XABORT_EXPLICIT (1 << 0) |
| #define _XABORT_RETRY (1 << 1) |
| #define _XABORT_CONFLICT (1 << 2) |
| #define _XABORT_CAPACITY (1 << 3) |
| #define _XABORT_DEBUG (1 << 4) |
| #define _XABORT_NESTED (1 << 5) |
| #define _XABORT_CODE(x) (((x) >> 24) & 0xff) |
| |
| #define _ABORT_LOCK_BUSY 0xff |
| #define _ABORT_LOCK_IS_LOCKED 0xfe |
| #define _ABORT_NESTED_TRYLOCK 0xfd |
| |
| #ifndef __ASSEMBLER__ |
| |
| #define __force_inline __attribute__((__always_inline__)) inline |
| |
| static __force_inline int _xbegin(void) |
| { |
| int ret = _XBEGIN_STARTED; |
| asm volatile (".byte 0xc7,0xf8 ; .long 0" : "+a" (ret) :: "memory"); |
| return ret; |
| } |
| |
| static __force_inline void _xend(void) |
| { |
| asm volatile (".byte 0x0f,0x01,0xd5" ::: "memory"); |
| } |
| |
| static __force_inline void _xabort(const unsigned int status) |
| { |
| asm volatile (".byte 0xc6,0xf8,%P0" :: "i" (status) : "memory"); |
| } |
| |
| static __force_inline int _xtest(void) |
| { |
| unsigned char out; |
| asm volatile (".byte 0x0f,0x01,0xd6 ; setnz %0" : "=r" (out) :: "memory"); |
| return out; |
| } |
| |
| #endif |
| #endif |