mirror of
https://github.com/xemu-project/xemu.git
synced 2024-11-30 15:00:34 +00:00
cbbab9226d
This is just code movement, and moving the fpu/ include path from target-dependent to target-independent Make variables. Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> Signed-off-by: Anthony Liguori <aliguori@us.ibm.com>
715 lines
14 KiB
C
715 lines
14 KiB
C
#ifndef BSWAP_H
|
|
#define BSWAP_H
|
|
|
|
#include "config-host.h"
|
|
|
|
#include <inttypes.h>
|
|
|
|
#ifdef CONFIG_MACHINE_BSWAP_H
|
|
#include <sys/endian.h>
|
|
#include <sys/types.h>
|
|
#include <machine/bswap.h>
|
|
#else
|
|
|
|
#include "softfloat.h"
|
|
|
|
#ifdef CONFIG_BYTESWAP_H
|
|
#include <byteswap.h>
|
|
#else
|
|
|
|
#define bswap_16(x) \
|
|
({ \
|
|
uint16_t __x = (x); \
|
|
((uint16_t)( \
|
|
(((uint16_t)(__x) & (uint16_t)0x00ffU) << 8) | \
|
|
(((uint16_t)(__x) & (uint16_t)0xff00U) >> 8) )); \
|
|
})
|
|
|
|
#define bswap_32(x) \
|
|
({ \
|
|
uint32_t __x = (x); \
|
|
((uint32_t)( \
|
|
(((uint32_t)(__x) & (uint32_t)0x000000ffUL) << 24) | \
|
|
(((uint32_t)(__x) & (uint32_t)0x0000ff00UL) << 8) | \
|
|
(((uint32_t)(__x) & (uint32_t)0x00ff0000UL) >> 8) | \
|
|
(((uint32_t)(__x) & (uint32_t)0xff000000UL) >> 24) )); \
|
|
})
|
|
|
|
#define bswap_64(x) \
|
|
({ \
|
|
uint64_t __x = (x); \
|
|
((uint64_t)( \
|
|
(uint64_t)(((uint64_t)(__x) & (uint64_t)0x00000000000000ffULL) << 56) | \
|
|
(uint64_t)(((uint64_t)(__x) & (uint64_t)0x000000000000ff00ULL) << 40) | \
|
|
(uint64_t)(((uint64_t)(__x) & (uint64_t)0x0000000000ff0000ULL) << 24) | \
|
|
(uint64_t)(((uint64_t)(__x) & (uint64_t)0x00000000ff000000ULL) << 8) | \
|
|
(uint64_t)(((uint64_t)(__x) & (uint64_t)0x000000ff00000000ULL) >> 8) | \
|
|
(uint64_t)(((uint64_t)(__x) & (uint64_t)0x0000ff0000000000ULL) >> 24) | \
|
|
(uint64_t)(((uint64_t)(__x) & (uint64_t)0x00ff000000000000ULL) >> 40) | \
|
|
(uint64_t)(((uint64_t)(__x) & (uint64_t)0xff00000000000000ULL) >> 56) )); \
|
|
})
|
|
|
|
#endif /* !CONFIG_BYTESWAP_H */
|
|
|
|
static inline uint16_t bswap16(uint16_t x)
|
|
{
|
|
return bswap_16(x);
|
|
}
|
|
|
|
static inline uint32_t bswap32(uint32_t x)
|
|
{
|
|
return bswap_32(x);
|
|
}
|
|
|
|
static inline uint64_t bswap64(uint64_t x)
|
|
{
|
|
return bswap_64(x);
|
|
}
|
|
|
|
#endif /* ! CONFIG_MACHINE_BSWAP_H */
|
|
|
|
static inline void bswap16s(uint16_t *s)
|
|
{
|
|
*s = bswap16(*s);
|
|
}
|
|
|
|
static inline void bswap32s(uint32_t *s)
|
|
{
|
|
*s = bswap32(*s);
|
|
}
|
|
|
|
static inline void bswap64s(uint64_t *s)
|
|
{
|
|
*s = bswap64(*s);
|
|
}
|
|
|
|
#if defined(HOST_WORDS_BIGENDIAN)
|
|
#define be_bswap(v, size) (v)
|
|
#define le_bswap(v, size) bswap ## size(v)
|
|
#define be_bswaps(v, size)
|
|
#define le_bswaps(p, size) *p = bswap ## size(*p);
|
|
#else
|
|
#define le_bswap(v, size) (v)
|
|
#define be_bswap(v, size) bswap ## size(v)
|
|
#define le_bswaps(v, size)
|
|
#define be_bswaps(p, size) *p = bswap ## size(*p);
|
|
#endif
|
|
|
|
#define CPU_CONVERT(endian, size, type)\
|
|
static inline type endian ## size ## _to_cpu(type v)\
|
|
{\
|
|
return endian ## _bswap(v, size);\
|
|
}\
|
|
\
|
|
static inline type cpu_to_ ## endian ## size(type v)\
|
|
{\
|
|
return endian ## _bswap(v, size);\
|
|
}\
|
|
\
|
|
static inline void endian ## size ## _to_cpus(type *p)\
|
|
{\
|
|
endian ## _bswaps(p, size)\
|
|
}\
|
|
\
|
|
static inline void cpu_to_ ## endian ## size ## s(type *p)\
|
|
{\
|
|
endian ## _bswaps(p, size)\
|
|
}\
|
|
\
|
|
static inline type endian ## size ## _to_cpup(const type *p)\
|
|
{\
|
|
return endian ## size ## _to_cpu(*p);\
|
|
}\
|
|
\
|
|
static inline void cpu_to_ ## endian ## size ## w(type *p, type v)\
|
|
{\
|
|
*p = cpu_to_ ## endian ## size(v);\
|
|
}
|
|
|
|
CPU_CONVERT(be, 16, uint16_t)
|
|
CPU_CONVERT(be, 32, uint32_t)
|
|
CPU_CONVERT(be, 64, uint64_t)
|
|
|
|
CPU_CONVERT(le, 16, uint16_t)
|
|
CPU_CONVERT(le, 32, uint32_t)
|
|
CPU_CONVERT(le, 64, uint64_t)
|
|
|
|
/* unaligned versions (optimized for frequent unaligned accesses)*/
|
|
|
|
#if defined(__i386__) || defined(_ARCH_PPC)
|
|
|
|
#define cpu_to_le16wu(p, v) cpu_to_le16w(p, v)
|
|
#define cpu_to_le32wu(p, v) cpu_to_le32w(p, v)
|
|
#define le16_to_cpupu(p) le16_to_cpup(p)
|
|
#define le32_to_cpupu(p) le32_to_cpup(p)
|
|
#define be32_to_cpupu(p) be32_to_cpup(p)
|
|
|
|
#define cpu_to_be16wu(p, v) cpu_to_be16w(p, v)
|
|
#define cpu_to_be32wu(p, v) cpu_to_be32w(p, v)
|
|
#define cpu_to_be64wu(p, v) cpu_to_be64w(p, v)
|
|
|
|
#else
|
|
|
|
static inline void cpu_to_le16wu(uint16_t *p, uint16_t v)
|
|
{
|
|
uint8_t *p1 = (uint8_t *)p;
|
|
|
|
p1[0] = v & 0xff;
|
|
p1[1] = v >> 8;
|
|
}
|
|
|
|
static inline void cpu_to_le32wu(uint32_t *p, uint32_t v)
|
|
{
|
|
uint8_t *p1 = (uint8_t *)p;
|
|
|
|
p1[0] = v & 0xff;
|
|
p1[1] = v >> 8;
|
|
p1[2] = v >> 16;
|
|
p1[3] = v >> 24;
|
|
}
|
|
|
|
static inline uint16_t le16_to_cpupu(const uint16_t *p)
|
|
{
|
|
const uint8_t *p1 = (const uint8_t *)p;
|
|
return p1[0] | (p1[1] << 8);
|
|
}
|
|
|
|
static inline uint32_t le32_to_cpupu(const uint32_t *p)
|
|
{
|
|
const uint8_t *p1 = (const uint8_t *)p;
|
|
return p1[0] | (p1[1] << 8) | (p1[2] << 16) | (p1[3] << 24);
|
|
}
|
|
|
|
static inline uint32_t be32_to_cpupu(const uint32_t *p)
|
|
{
|
|
const uint8_t *p1 = (const uint8_t *)p;
|
|
return p1[3] | (p1[2] << 8) | (p1[1] << 16) | (p1[0] << 24);
|
|
}
|
|
|
|
static inline void cpu_to_be16wu(uint16_t *p, uint16_t v)
|
|
{
|
|
uint8_t *p1 = (uint8_t *)p;
|
|
|
|
p1[0] = v >> 8;
|
|
p1[1] = v & 0xff;
|
|
}
|
|
|
|
static inline void cpu_to_be32wu(uint32_t *p, uint32_t v)
|
|
{
|
|
uint8_t *p1 = (uint8_t *)p;
|
|
|
|
p1[0] = v >> 24;
|
|
p1[1] = v >> 16;
|
|
p1[2] = v >> 8;
|
|
p1[3] = v & 0xff;
|
|
}
|
|
|
|
static inline void cpu_to_be64wu(uint64_t *p, uint64_t v)
|
|
{
|
|
uint8_t *p1 = (uint8_t *)p;
|
|
|
|
p1[0] = v >> 56;
|
|
p1[1] = v >> 48;
|
|
p1[2] = v >> 40;
|
|
p1[3] = v >> 32;
|
|
p1[4] = v >> 24;
|
|
p1[5] = v >> 16;
|
|
p1[6] = v >> 8;
|
|
p1[7] = v & 0xff;
|
|
}
|
|
|
|
#endif
|
|
|
|
#ifdef HOST_WORDS_BIGENDIAN
|
|
#define cpu_to_32wu cpu_to_be32wu
|
|
#define leul_to_cpu(v) glue(glue(le,HOST_LONG_BITS),_to_cpu)(v)
|
|
#else
|
|
#define cpu_to_32wu cpu_to_le32wu
|
|
#define leul_to_cpu(v) (v)
|
|
#endif
|
|
|
|
#undef le_bswap
|
|
#undef be_bswap
|
|
#undef le_bswaps
|
|
#undef be_bswaps
|
|
|
|
/* len must be one of 1, 2, 4 */
|
|
static inline uint32_t qemu_bswap_len(uint32_t value, int len)
|
|
{
|
|
return bswap32(value) >> (32 - 8 * len);
|
|
}
|
|
|
|
typedef union {
|
|
float32 f;
|
|
uint32_t l;
|
|
} CPU_FloatU;
|
|
|
|
typedef union {
|
|
float64 d;
|
|
#if defined(HOST_WORDS_BIGENDIAN)
|
|
struct {
|
|
uint32_t upper;
|
|
uint32_t lower;
|
|
} l;
|
|
#else
|
|
struct {
|
|
uint32_t lower;
|
|
uint32_t upper;
|
|
} l;
|
|
#endif
|
|
uint64_t ll;
|
|
} CPU_DoubleU;
|
|
|
|
typedef union {
|
|
floatx80 d;
|
|
struct {
|
|
uint64_t lower;
|
|
uint16_t upper;
|
|
} l;
|
|
} CPU_LDoubleU;
|
|
|
|
typedef union {
|
|
float128 q;
|
|
#if defined(HOST_WORDS_BIGENDIAN)
|
|
struct {
|
|
uint32_t upmost;
|
|
uint32_t upper;
|
|
uint32_t lower;
|
|
uint32_t lowest;
|
|
} l;
|
|
struct {
|
|
uint64_t upper;
|
|
uint64_t lower;
|
|
} ll;
|
|
#else
|
|
struct {
|
|
uint32_t lowest;
|
|
uint32_t lower;
|
|
uint32_t upper;
|
|
uint32_t upmost;
|
|
} l;
|
|
struct {
|
|
uint64_t lower;
|
|
uint64_t upper;
|
|
} ll;
|
|
#endif
|
|
} CPU_QuadU;
|
|
|
|
/* unaligned/endian-independent pointer access */
|
|
|
|
/*
|
|
* the generic syntax is:
|
|
*
|
|
* load: ld{type}{sign}{size}{endian}_p(ptr)
|
|
*
|
|
* store: st{type}{size}{endian}_p(ptr, val)
|
|
*
|
|
* Note there are small differences with the softmmu access API!
|
|
*
|
|
* type is:
|
|
* (empty): integer access
|
|
* f : float access
|
|
*
|
|
* sign is:
|
|
* (empty): for floats or 32 bit size
|
|
* u : unsigned
|
|
* s : signed
|
|
*
|
|
* size is:
|
|
* b: 8 bits
|
|
* w: 16 bits
|
|
* l: 32 bits
|
|
* q: 64 bits
|
|
*
|
|
* endian is:
|
|
* (empty): 8 bit access
|
|
* be : big endian
|
|
* le : little endian
|
|
*/
|
|
static inline int ldub_p(const void *ptr)
|
|
{
|
|
return *(uint8_t *)ptr;
|
|
}
|
|
|
|
static inline int ldsb_p(const void *ptr)
|
|
{
|
|
return *(int8_t *)ptr;
|
|
}
|
|
|
|
static inline void stb_p(void *ptr, int v)
|
|
{
|
|
*(uint8_t *)ptr = v;
|
|
}
|
|
|
|
/* NOTE: on arm, putting 2 in /proc/sys/debug/alignment so that the
|
|
kernel handles unaligned load/stores may give better results, but
|
|
it is a system wide setting : bad */
|
|
#if defined(HOST_WORDS_BIGENDIAN) || defined(WORDS_ALIGNED)
|
|
|
|
/* conservative code for little endian unaligned accesses */
|
|
static inline int lduw_le_p(const void *ptr)
|
|
{
|
|
#ifdef _ARCH_PPC
|
|
int val;
|
|
__asm__ __volatile__ ("lhbrx %0,0,%1" : "=r" (val) : "r" (ptr));
|
|
return val;
|
|
#else
|
|
const uint8_t *p = ptr;
|
|
return p[0] | (p[1] << 8);
|
|
#endif
|
|
}
|
|
|
|
static inline int ldsw_le_p(const void *ptr)
|
|
{
|
|
#ifdef _ARCH_PPC
|
|
int val;
|
|
__asm__ __volatile__ ("lhbrx %0,0,%1" : "=r" (val) : "r" (ptr));
|
|
return (int16_t)val;
|
|
#else
|
|
const uint8_t *p = ptr;
|
|
return (int16_t)(p[0] | (p[1] << 8));
|
|
#endif
|
|
}
|
|
|
|
static inline int ldl_le_p(const void *ptr)
|
|
{
|
|
#ifdef _ARCH_PPC
|
|
int val;
|
|
__asm__ __volatile__ ("lwbrx %0,0,%1" : "=r" (val) : "r" (ptr));
|
|
return val;
|
|
#else
|
|
const uint8_t *p = ptr;
|
|
return p[0] | (p[1] << 8) | (p[2] << 16) | (p[3] << 24);
|
|
#endif
|
|
}
|
|
|
|
static inline uint64_t ldq_le_p(const void *ptr)
|
|
{
|
|
const uint8_t *p = ptr;
|
|
uint32_t v1, v2;
|
|
v1 = ldl_le_p(p);
|
|
v2 = ldl_le_p(p + 4);
|
|
return v1 | ((uint64_t)v2 << 32);
|
|
}
|
|
|
|
static inline void stw_le_p(void *ptr, int v)
|
|
{
|
|
#ifdef _ARCH_PPC
|
|
__asm__ __volatile__ ("sthbrx %1,0,%2" : "=m" (*(uint16_t *)ptr) : "r" (v), "r" (ptr));
|
|
#else
|
|
uint8_t *p = ptr;
|
|
p[0] = v;
|
|
p[1] = v >> 8;
|
|
#endif
|
|
}
|
|
|
|
static inline void stl_le_p(void *ptr, int v)
|
|
{
|
|
#ifdef _ARCH_PPC
|
|
__asm__ __volatile__ ("stwbrx %1,0,%2" : "=m" (*(uint32_t *)ptr) : "r" (v), "r" (ptr));
|
|
#else
|
|
uint8_t *p = ptr;
|
|
p[0] = v;
|
|
p[1] = v >> 8;
|
|
p[2] = v >> 16;
|
|
p[3] = v >> 24;
|
|
#endif
|
|
}
|
|
|
|
static inline void stq_le_p(void *ptr, uint64_t v)
|
|
{
|
|
uint8_t *p = ptr;
|
|
stl_le_p(p, (uint32_t)v);
|
|
stl_le_p(p + 4, v >> 32);
|
|
}
|
|
|
|
/* float access */
|
|
|
|
static inline float32 ldfl_le_p(const void *ptr)
|
|
{
|
|
union {
|
|
float32 f;
|
|
uint32_t i;
|
|
} u;
|
|
u.i = ldl_le_p(ptr);
|
|
return u.f;
|
|
}
|
|
|
|
static inline void stfl_le_p(void *ptr, float32 v)
|
|
{
|
|
union {
|
|
float32 f;
|
|
uint32_t i;
|
|
} u;
|
|
u.f = v;
|
|
stl_le_p(ptr, u.i);
|
|
}
|
|
|
|
static inline float64 ldfq_le_p(const void *ptr)
|
|
{
|
|
CPU_DoubleU u;
|
|
u.l.lower = ldl_le_p(ptr);
|
|
u.l.upper = ldl_le_p(ptr + 4);
|
|
return u.d;
|
|
}
|
|
|
|
static inline void stfq_le_p(void *ptr, float64 v)
|
|
{
|
|
CPU_DoubleU u;
|
|
u.d = v;
|
|
stl_le_p(ptr, u.l.lower);
|
|
stl_le_p(ptr + 4, u.l.upper);
|
|
}
|
|
|
|
#else
|
|
|
|
static inline int lduw_le_p(const void *ptr)
|
|
{
|
|
return *(uint16_t *)ptr;
|
|
}
|
|
|
|
static inline int ldsw_le_p(const void *ptr)
|
|
{
|
|
return *(int16_t *)ptr;
|
|
}
|
|
|
|
static inline int ldl_le_p(const void *ptr)
|
|
{
|
|
return *(uint32_t *)ptr;
|
|
}
|
|
|
|
static inline uint64_t ldq_le_p(const void *ptr)
|
|
{
|
|
return *(uint64_t *)ptr;
|
|
}
|
|
|
|
static inline void stw_le_p(void *ptr, int v)
|
|
{
|
|
*(uint16_t *)ptr = v;
|
|
}
|
|
|
|
static inline void stl_le_p(void *ptr, int v)
|
|
{
|
|
*(uint32_t *)ptr = v;
|
|
}
|
|
|
|
static inline void stq_le_p(void *ptr, uint64_t v)
|
|
{
|
|
*(uint64_t *)ptr = v;
|
|
}
|
|
|
|
/* float access */
|
|
|
|
static inline float32 ldfl_le_p(const void *ptr)
|
|
{
|
|
return *(float32 *)ptr;
|
|
}
|
|
|
|
static inline float64 ldfq_le_p(const void *ptr)
|
|
{
|
|
return *(float64 *)ptr;
|
|
}
|
|
|
|
static inline void stfl_le_p(void *ptr, float32 v)
|
|
{
|
|
*(float32 *)ptr = v;
|
|
}
|
|
|
|
static inline void stfq_le_p(void *ptr, float64 v)
|
|
{
|
|
*(float64 *)ptr = v;
|
|
}
|
|
#endif
|
|
|
|
#if !defined(HOST_WORDS_BIGENDIAN) || defined(WORDS_ALIGNED)
|
|
|
|
static inline int lduw_be_p(const void *ptr)
|
|
{
|
|
#if defined(__i386__)
|
|
int val;
|
|
asm volatile ("movzwl %1, %0\n"
|
|
"xchgb %b0, %h0\n"
|
|
: "=q" (val)
|
|
: "m" (*(uint16_t *)ptr));
|
|
return val;
|
|
#else
|
|
const uint8_t *b = ptr;
|
|
return ((b[0] << 8) | b[1]);
|
|
#endif
|
|
}
|
|
|
|
static inline int ldsw_be_p(const void *ptr)
|
|
{
|
|
#if defined(__i386__)
|
|
int val;
|
|
asm volatile ("movzwl %1, %0\n"
|
|
"xchgb %b0, %h0\n"
|
|
: "=q" (val)
|
|
: "m" (*(uint16_t *)ptr));
|
|
return (int16_t)val;
|
|
#else
|
|
const uint8_t *b = ptr;
|
|
return (int16_t)((b[0] << 8) | b[1]);
|
|
#endif
|
|
}
|
|
|
|
static inline int ldl_be_p(const void *ptr)
|
|
{
|
|
#if defined(__i386__) || defined(__x86_64__)
|
|
int val;
|
|
asm volatile ("movl %1, %0\n"
|
|
"bswap %0\n"
|
|
: "=r" (val)
|
|
: "m" (*(uint32_t *)ptr));
|
|
return val;
|
|
#else
|
|
const uint8_t *b = ptr;
|
|
return (b[0] << 24) | (b[1] << 16) | (b[2] << 8) | b[3];
|
|
#endif
|
|
}
|
|
|
|
static inline uint64_t ldq_be_p(const void *ptr)
|
|
{
|
|
uint32_t a,b;
|
|
a = ldl_be_p(ptr);
|
|
b = ldl_be_p((uint8_t *)ptr + 4);
|
|
return (((uint64_t)a<<32)|b);
|
|
}
|
|
|
|
static inline void stw_be_p(void *ptr, int v)
|
|
{
|
|
#if defined(__i386__)
|
|
asm volatile ("xchgb %b0, %h0\n"
|
|
"movw %w0, %1\n"
|
|
: "=q" (v)
|
|
: "m" (*(uint16_t *)ptr), "0" (v));
|
|
#else
|
|
uint8_t *d = (uint8_t *) ptr;
|
|
d[0] = v >> 8;
|
|
d[1] = v;
|
|
#endif
|
|
}
|
|
|
|
static inline void stl_be_p(void *ptr, int v)
|
|
{
|
|
#if defined(__i386__) || defined(__x86_64__)
|
|
asm volatile ("bswap %0\n"
|
|
"movl %0, %1\n"
|
|
: "=r" (v)
|
|
: "m" (*(uint32_t *)ptr), "0" (v));
|
|
#else
|
|
uint8_t *d = (uint8_t *) ptr;
|
|
d[0] = v >> 24;
|
|
d[1] = v >> 16;
|
|
d[2] = v >> 8;
|
|
d[3] = v;
|
|
#endif
|
|
}
|
|
|
|
static inline void stq_be_p(void *ptr, uint64_t v)
|
|
{
|
|
stl_be_p(ptr, v >> 32);
|
|
stl_be_p((uint8_t *)ptr + 4, v);
|
|
}
|
|
|
|
/* float access */
|
|
|
|
static inline float32 ldfl_be_p(const void *ptr)
|
|
{
|
|
union {
|
|
float32 f;
|
|
uint32_t i;
|
|
} u;
|
|
u.i = ldl_be_p(ptr);
|
|
return u.f;
|
|
}
|
|
|
|
static inline void stfl_be_p(void *ptr, float32 v)
|
|
{
|
|
union {
|
|
float32 f;
|
|
uint32_t i;
|
|
} u;
|
|
u.f = v;
|
|
stl_be_p(ptr, u.i);
|
|
}
|
|
|
|
static inline float64 ldfq_be_p(const void *ptr)
|
|
{
|
|
CPU_DoubleU u;
|
|
u.l.upper = ldl_be_p(ptr);
|
|
u.l.lower = ldl_be_p((uint8_t *)ptr + 4);
|
|
return u.d;
|
|
}
|
|
|
|
static inline void stfq_be_p(void *ptr, float64 v)
|
|
{
|
|
CPU_DoubleU u;
|
|
u.d = v;
|
|
stl_be_p(ptr, u.l.upper);
|
|
stl_be_p((uint8_t *)ptr + 4, u.l.lower);
|
|
}
|
|
|
|
#else
|
|
|
|
static inline int lduw_be_p(const void *ptr)
|
|
{
|
|
return *(uint16_t *)ptr;
|
|
}
|
|
|
|
static inline int ldsw_be_p(const void *ptr)
|
|
{
|
|
return *(int16_t *)ptr;
|
|
}
|
|
|
|
static inline int ldl_be_p(const void *ptr)
|
|
{
|
|
return *(uint32_t *)ptr;
|
|
}
|
|
|
|
static inline uint64_t ldq_be_p(const void *ptr)
|
|
{
|
|
return *(uint64_t *)ptr;
|
|
}
|
|
|
|
static inline void stw_be_p(void *ptr, int v)
|
|
{
|
|
*(uint16_t *)ptr = v;
|
|
}
|
|
|
|
static inline void stl_be_p(void *ptr, int v)
|
|
{
|
|
*(uint32_t *)ptr = v;
|
|
}
|
|
|
|
static inline void stq_be_p(void *ptr, uint64_t v)
|
|
{
|
|
*(uint64_t *)ptr = v;
|
|
}
|
|
|
|
/* float access */
|
|
|
|
static inline float32 ldfl_be_p(const void *ptr)
|
|
{
|
|
return *(float32 *)ptr;
|
|
}
|
|
|
|
static inline float64 ldfq_be_p(const void *ptr)
|
|
{
|
|
return *(float64 *)ptr;
|
|
}
|
|
|
|
static inline void stfl_be_p(void *ptr, float32 v)
|
|
{
|
|
*(float32 *)ptr = v;
|
|
}
|
|
|
|
static inline void stfq_be_p(void *ptr, float64 v)
|
|
{
|
|
*(float64 *)ptr = v;
|
|
}
|
|
|
|
#endif
|
|
|
|
#endif /* BSWAP_H */
|