mirror of
https://github.com/FEX-Emu/linux.git
synced 2024-12-22 17:33:01 +00:00
a3ebdb6c42
__clear_bit_unlock does not need to perform atomic operations on the variable. Avoid a cmpxchg and simply do a store with release semantics. Add a barrier to be safe that the compiler does not do funky things. Tony: Use intrinsic rather than inline assembler Signed-off-by: Christoph Lameter <clameter@sgi.com> Acked-by: Nick Piggin <nickpiggin@yahoo.com.au> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Tony Luck <tony.luck@intel.com>
161 lines
4.7 KiB
C
161 lines
4.7 KiB
C
#ifndef _ASM_IA64_INTEL_INTRIN_H
|
|
#define _ASM_IA64_INTEL_INTRIN_H
|
|
/*
|
|
* Intel Compiler Intrinsics
|
|
*
|
|
* Copyright (C) 2002,2003 Jun Nakajima <jun.nakajima@intel.com>
|
|
* Copyright (C) 2002,2003 Suresh Siddha <suresh.b.siddha@intel.com>
|
|
* Copyright (C) 2005,2006 Hongjiu Lu <hongjiu.lu@intel.com>
|
|
*
|
|
*/
|
|
#include <ia64intrin.h>
|
|
|
|
#define ia64_barrier() __memory_barrier()
|
|
|
|
#define ia64_stop() /* Nothing: As of now stop bit is generated for each
|
|
* intrinsic
|
|
*/
|
|
|
|
#define ia64_getreg __getReg
|
|
#define ia64_setreg __setReg
|
|
|
|
#define ia64_hint __hint
|
|
#define ia64_hint_pause __hint_pause
|
|
|
|
#define ia64_mux1_brcst _m64_mux1_brcst
|
|
#define ia64_mux1_mix _m64_mux1_mix
|
|
#define ia64_mux1_shuf _m64_mux1_shuf
|
|
#define ia64_mux1_alt _m64_mux1_alt
|
|
#define ia64_mux1_rev _m64_mux1_rev
|
|
|
|
#define ia64_mux1(x,v) _m_to_int64(_m64_mux1(_m_from_int64(x), (v)))
|
|
#define ia64_popcnt _m64_popcnt
|
|
#define ia64_getf_exp __getf_exp
|
|
#define ia64_shrp _m64_shrp
|
|
|
|
#define ia64_tpa __tpa
|
|
#define ia64_invala __invala
|
|
#define ia64_invala_gr __invala_gr
|
|
#define ia64_invala_fr __invala_fr
|
|
#define ia64_nop __nop
|
|
#define ia64_sum __sum
|
|
#define ia64_ssm __ssm
|
|
#define ia64_rum __rum
|
|
#define ia64_rsm __rsm
|
|
#define ia64_fc __fc
|
|
|
|
#define ia64_ldfs __ldfs
|
|
#define ia64_ldfd __ldfd
|
|
#define ia64_ldfe __ldfe
|
|
#define ia64_ldf8 __ldf8
|
|
#define ia64_ldf_fill __ldf_fill
|
|
|
|
#define ia64_stfs __stfs
|
|
#define ia64_stfd __stfd
|
|
#define ia64_stfe __stfe
|
|
#define ia64_stf8 __stf8
|
|
#define ia64_stf_spill __stf_spill
|
|
|
|
#define ia64_mf __mf
|
|
#define ia64_mfa __mfa
|
|
|
|
#define ia64_fetchadd4_acq __fetchadd4_acq
|
|
#define ia64_fetchadd4_rel __fetchadd4_rel
|
|
#define ia64_fetchadd8_acq __fetchadd8_acq
|
|
#define ia64_fetchadd8_rel __fetchadd8_rel
|
|
|
|
#define ia64_xchg1 _InterlockedExchange8
|
|
#define ia64_xchg2 _InterlockedExchange16
|
|
#define ia64_xchg4 _InterlockedExchange
|
|
#define ia64_xchg8 _InterlockedExchange64
|
|
|
|
#define ia64_cmpxchg1_rel _InterlockedCompareExchange8_rel
|
|
#define ia64_cmpxchg1_acq _InterlockedCompareExchange8_acq
|
|
#define ia64_cmpxchg2_rel _InterlockedCompareExchange16_rel
|
|
#define ia64_cmpxchg2_acq _InterlockedCompareExchange16_acq
|
|
#define ia64_cmpxchg4_rel _InterlockedCompareExchange_rel
|
|
#define ia64_cmpxchg4_acq _InterlockedCompareExchange_acq
|
|
#define ia64_cmpxchg8_rel _InterlockedCompareExchange64_rel
|
|
#define ia64_cmpxchg8_acq _InterlockedCompareExchange64_acq
|
|
|
|
#define __ia64_set_dbr(index, val) \
|
|
__setIndReg(_IA64_REG_INDR_DBR, index, val)
|
|
#define ia64_set_ibr(index, val) \
|
|
__setIndReg(_IA64_REG_INDR_IBR, index, val)
|
|
#define ia64_set_pkr(index, val) \
|
|
__setIndReg(_IA64_REG_INDR_PKR, index, val)
|
|
#define ia64_set_pmc(index, val) \
|
|
__setIndReg(_IA64_REG_INDR_PMC, index, val)
|
|
#define ia64_set_pmd(index, val) \
|
|
__setIndReg(_IA64_REG_INDR_PMD, index, val)
|
|
#define ia64_set_rr(index, val) \
|
|
__setIndReg(_IA64_REG_INDR_RR, index, val)
|
|
|
|
#define ia64_get_cpuid(index) __getIndReg(_IA64_REG_INDR_CPUID, index)
|
|
#define __ia64_get_dbr(index) __getIndReg(_IA64_REG_INDR_DBR, index)
|
|
#define ia64_get_ibr(index) __getIndReg(_IA64_REG_INDR_IBR, index)
|
|
#define ia64_get_pkr(index) __getIndReg(_IA64_REG_INDR_PKR, index)
|
|
#define ia64_get_pmc(index) __getIndReg(_IA64_REG_INDR_PMC, index)
|
|
#define ia64_get_pmd(index) __getIndReg(_IA64_REG_INDR_PMD, index)
|
|
#define ia64_get_rr(index) __getIndReg(_IA64_REG_INDR_RR, index)
|
|
|
|
#define ia64_srlz_d __dsrlz
|
|
#define ia64_srlz_i __isrlz
|
|
|
|
#define ia64_dv_serialize_data()
|
|
#define ia64_dv_serialize_instruction()
|
|
|
|
#define ia64_st1_rel __st1_rel
|
|
#define ia64_st2_rel __st2_rel
|
|
#define ia64_st4_rel __st4_rel
|
|
#define ia64_st8_rel __st8_rel
|
|
|
|
/* FIXME: need st4.rel.nta intrinsic */
|
|
#define ia64_st4_rel_nta __st4_rel
|
|
|
|
#define ia64_ld1_acq __ld1_acq
|
|
#define ia64_ld2_acq __ld2_acq
|
|
#define ia64_ld4_acq __ld4_acq
|
|
#define ia64_ld8_acq __ld8_acq
|
|
|
|
#define ia64_sync_i __synci
|
|
#define ia64_thash __thash
|
|
#define ia64_ttag __ttag
|
|
#define ia64_itcd __itcd
|
|
#define ia64_itci __itci
|
|
#define ia64_itrd __itrd
|
|
#define ia64_itri __itri
|
|
#define ia64_ptce __ptce
|
|
#define ia64_ptcl __ptcl
|
|
#define ia64_ptcg __ptcg
|
|
#define ia64_ptcga __ptcga
|
|
#define ia64_ptri __ptri
|
|
#define ia64_ptrd __ptrd
|
|
#define ia64_dep_mi _m64_dep_mi
|
|
|
|
/* Values for lfhint in __lfetch and __lfetch_fault */
|
|
|
|
#define ia64_lfhint_none __lfhint_none
|
|
#define ia64_lfhint_nt1 __lfhint_nt1
|
|
#define ia64_lfhint_nt2 __lfhint_nt2
|
|
#define ia64_lfhint_nta __lfhint_nta
|
|
|
|
#define ia64_lfetch __lfetch
|
|
#define ia64_lfetch_excl __lfetch_excl
|
|
#define ia64_lfetch_fault __lfetch_fault
|
|
#define ia64_lfetch_fault_excl __lfetch_fault_excl
|
|
|
|
#define ia64_intrin_local_irq_restore(x) \
|
|
do { \
|
|
if ((x) != 0) { \
|
|
ia64_ssm(IA64_PSR_I); \
|
|
ia64_srlz_d(); \
|
|
} else { \
|
|
ia64_rsm(IA64_PSR_I); \
|
|
} \
|
|
} while (0)
|
|
|
|
#define __builtin_trap() __break(0);
|
|
|
|
#endif /* _ASM_IA64_INTEL_INTRIN_H */
|