18281b2578
The const_le64() macro introduced in commit845d80a8c7
turns out to have a bug which means that on big-endian systems the compiler complains if the argument isn't already a 64-bit type. This hasn't caused a problem yet, because there are no in-tree uses, but it means it's not possible for anybody to add one without it failing CI. This example is from an attempted use of it with the argument '0', from the s390 CI runner's gcc: ../block/blklogwrites.c: In function ‘blk_log_writes_co_do_log’: ../include/qemu/bswap.h:148:36: error: left shift count >= width of type [-Werror=shift-count-overflow] 148 | ((((_x) & 0x00000000000000ffU) << 56) | \ | ^~ ../block/blklogwrites.c:409:27: note: in expansion of macro ‘const_le64’ 409 | .nr_entries = const_le64(0), | ^~~~~~~~~~ ../include/qemu/bswap.h:149:36: error: left shift count >= width of type [-Werror=shift-count-overflow] 149 | (((_x) & 0x000000000000ff00U) << 40) | \ | ^~ ../block/blklogwrites.c:409:27: note: in expansion of macro ‘const_le64’ 409 | .nr_entries = const_le64(0), | ^~~~~~~~~~ cc1: all warnings being treated as errors Fix this by making all the constants in the macro have the ULL suffix. This will cause them all to be 64-bit integers, which means the result of the logical & will also be an unsigned 64-bit type, even if the input to the macro is a smaller type, and so the shifts will be in range. Fixes:845d80a8c7
("qemu/bswap: Add const_le64()") Signed-off-by: Peter Maydell <peter.maydell@linaro.org> Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> Reviewed-by: Thomas Huth <thuth@redhat.com> Reviewed-by: Kevin Wolf <kwolf@redhat.com> Reviewed-by: Ira Weiny <ira.weiny@intel.com> Message-id: 20240122173735.472951-1-peter.maydell@linaro.org
430 lines
12 KiB
C
430 lines
12 KiB
C
#ifndef BSWAP_H
|
|
#define BSWAP_H
|
|
|
|
#undef bswap16
|
|
#define bswap16(_x) __builtin_bswap16(_x)
|
|
#undef bswap32
|
|
#define bswap32(_x) __builtin_bswap32(_x)
|
|
#undef bswap64
|
|
#define bswap64(_x) __builtin_bswap64(_x)
|
|
|
|
static inline uint32_t bswap24(uint32_t x)
|
|
{
|
|
return (((x & 0x000000ffU) << 16) |
|
|
((x & 0x0000ff00U) << 0) |
|
|
((x & 0x00ff0000U) >> 16));
|
|
}
|
|
|
|
static inline void bswap16s(uint16_t *s)
|
|
{
|
|
*s = __builtin_bswap16(*s);
|
|
}
|
|
|
|
static inline void bswap24s(uint32_t *s)
|
|
{
|
|
*s = bswap24(*s & 0x00ffffffU);
|
|
}
|
|
|
|
static inline void bswap32s(uint32_t *s)
|
|
{
|
|
*s = __builtin_bswap32(*s);
|
|
}
|
|
|
|
static inline void bswap64s(uint64_t *s)
|
|
{
|
|
*s = __builtin_bswap64(*s);
|
|
}
|
|
|
|
#if HOST_BIG_ENDIAN
|
|
#define be_bswap(v, size) (v)
|
|
#define le_bswap(v, size) glue(__builtin_bswap, size)(v)
|
|
#define le_bswap24(v) bswap24(v)
|
|
#define be_bswaps(v, size)
|
|
#define le_bswaps(p, size) \
|
|
do { *p = glue(__builtin_bswap, size)(*p); } while (0)
|
|
#else
|
|
#define le_bswap(v, size) (v)
|
|
#define le_bswap24(v) (v)
|
|
#define be_bswap(v, size) glue(__builtin_bswap, size)(v)
|
|
#define le_bswaps(v, size)
|
|
#define be_bswaps(p, size) \
|
|
do { *p = glue(__builtin_bswap, size)(*p); } while (0)
|
|
#endif
|
|
|
|
/**
|
|
* Endianness conversion functions between host cpu and specified endianness.
|
|
* (We list the complete set of prototypes produced by the macros below
|
|
* to assist people who search the headers to find their definitions.)
|
|
*
|
|
* uint16_t le16_to_cpu(uint16_t v);
|
|
* uint32_t le32_to_cpu(uint32_t v);
|
|
* uint64_t le64_to_cpu(uint64_t v);
|
|
* uint16_t be16_to_cpu(uint16_t v);
|
|
* uint32_t be32_to_cpu(uint32_t v);
|
|
* uint64_t be64_to_cpu(uint64_t v);
|
|
*
|
|
* Convert the value @v from the specified format to the native
|
|
* endianness of the host CPU by byteswapping if necessary, and
|
|
* return the converted value.
|
|
*
|
|
* uint16_t cpu_to_le16(uint16_t v);
|
|
* uint32_t cpu_to_le32(uint32_t v);
|
|
* uint64_t cpu_to_le64(uint64_t v);
|
|
* uint16_t cpu_to_be16(uint16_t v);
|
|
* uint32_t cpu_to_be32(uint32_t v);
|
|
* uint64_t cpu_to_be64(uint64_t v);
|
|
*
|
|
* Convert the value @v from the native endianness of the host CPU to
|
|
* the specified format by byteswapping if necessary, and return
|
|
* the converted value.
|
|
*
|
|
* void le16_to_cpus(uint16_t *v);
|
|
* void le32_to_cpus(uint32_t *v);
|
|
* void le64_to_cpus(uint64_t *v);
|
|
* void be16_to_cpus(uint16_t *v);
|
|
* void be32_to_cpus(uint32_t *v);
|
|
* void be64_to_cpus(uint64_t *v);
|
|
*
|
|
* Do an in-place conversion of the value pointed to by @v from the
|
|
* specified format to the native endianness of the host CPU.
|
|
*
|
|
* void cpu_to_le16s(uint16_t *v);
|
|
* void cpu_to_le32s(uint32_t *v);
|
|
* void cpu_to_le64s(uint64_t *v);
|
|
* void cpu_to_be16s(uint16_t *v);
|
|
* void cpu_to_be32s(uint32_t *v);
|
|
* void cpu_to_be64s(uint64_t *v);
|
|
*
|
|
* Do an in-place conversion of the value pointed to by @v from the
|
|
* native endianness of the host CPU to the specified format.
|
|
*
|
|
* Both X_to_cpu() and cpu_to_X() perform the same operation; you
|
|
* should use whichever one is better documenting of the function your
|
|
* code is performing.
|
|
*
|
|
* Do not use these functions for conversion of values which are in guest
|
|
* memory, since the data may not be sufficiently aligned for the host CPU's
|
|
* load and store instructions. Instead you should use the ld*_p() and
|
|
* st*_p() functions, which perform loads and stores of data of any
|
|
* required size and endianness and handle possible misalignment.
|
|
*/
|
|
|
|
#define CPU_CONVERT(endian, size, type)\
|
|
static inline type endian ## size ## _to_cpu(type v)\
|
|
{\
|
|
return glue(endian, _bswap)(v, size);\
|
|
}\
|
|
\
|
|
static inline type cpu_to_ ## endian ## size(type v)\
|
|
{\
|
|
return glue(endian, _bswap)(v, size);\
|
|
}\
|
|
\
|
|
static inline void endian ## size ## _to_cpus(type *p)\
|
|
{\
|
|
glue(endian, _bswaps)(p, size);\
|
|
}\
|
|
\
|
|
static inline void cpu_to_ ## endian ## size ## s(type *p)\
|
|
{\
|
|
glue(endian, _bswaps)(p, size);\
|
|
}
|
|
|
|
CPU_CONVERT(be, 16, uint16_t)
|
|
CPU_CONVERT(be, 32, uint32_t)
|
|
CPU_CONVERT(be, 64, uint64_t)
|
|
|
|
CPU_CONVERT(le, 16, uint16_t)
|
|
CPU_CONVERT(le, 32, uint32_t)
|
|
CPU_CONVERT(le, 64, uint64_t)
|
|
|
|
/*
|
|
* Same as cpu_to_le{16,32,64}, except that gcc will figure the result is
|
|
* a compile-time constant if you pass in a constant. So this can be
|
|
* used to initialize static variables.
|
|
*/
|
|
#if HOST_BIG_ENDIAN
|
|
# define const_le64(_x) \
|
|
((((_x) & 0x00000000000000ffULL) << 56) | \
|
|
(((_x) & 0x000000000000ff00ULL) << 40) | \
|
|
(((_x) & 0x0000000000ff0000ULL) << 24) | \
|
|
(((_x) & 0x00000000ff000000ULL) << 8) | \
|
|
(((_x) & 0x000000ff00000000ULL) >> 8) | \
|
|
(((_x) & 0x0000ff0000000000ULL) >> 24) | \
|
|
(((_x) & 0x00ff000000000000ULL) >> 40) | \
|
|
(((_x) & 0xff00000000000000ULL) >> 56))
|
|
# define const_le32(_x) \
|
|
((((_x) & 0x000000ffU) << 24) | \
|
|
(((_x) & 0x0000ff00U) << 8) | \
|
|
(((_x) & 0x00ff0000U) >> 8) | \
|
|
(((_x) & 0xff000000U) >> 24))
|
|
# define const_le16(_x) \
|
|
((((_x) & 0x00ff) << 8) | \
|
|
(((_x) & 0xff00) >> 8))
|
|
#else
|
|
# define const_le64(_x) (_x)
|
|
# define const_le32(_x) (_x)
|
|
# define const_le16(_x) (_x)
|
|
#endif
|
|
|
|
/* unaligned/endian-independent pointer access */
|
|
|
|
/*
|
|
* the generic syntax is:
|
|
*
|
|
* load: ld{type}{sign}{size}_{endian}_p(ptr)
|
|
*
|
|
* store: st{type}{size}_{endian}_p(ptr, val)
|
|
*
|
|
* Note there are small differences with the softmmu access API!
|
|
*
|
|
* type is:
|
|
* (empty): integer access
|
|
* f : float access
|
|
*
|
|
* sign is:
|
|
* (empty): for 32 or 64 bit sizes (including floats and doubles)
|
|
* u : unsigned
|
|
* s : signed
|
|
*
|
|
* size is:
|
|
* b: 8 bits
|
|
* w: 16 bits
|
|
* 24: 24 bits
|
|
* l: 32 bits
|
|
* q: 64 bits
|
|
*
|
|
* endian is:
|
|
* he : host endian
|
|
* be : big endian
|
|
* le : little endian
|
|
* te : target endian
|
|
* (except for byte accesses, which have no endian infix).
|
|
*
|
|
* The target endian accessors are obviously only available to source
|
|
* files which are built per-target; they are defined in cpu-all.h.
|
|
*
|
|
* In all cases these functions take a host pointer.
|
|
* For accessors that take a guest address rather than a
|
|
* host address, see the cpu_{ld,st}_* accessors defined in
|
|
* cpu_ldst.h.
|
|
*
|
|
* For cases where the size to be used is not fixed at compile time,
|
|
* there are
|
|
* stn_{endian}_p(ptr, sz, val)
|
|
* which stores @val to @ptr as an @endian-order number @sz bytes in size
|
|
* and
|
|
* ldn_{endian}_p(ptr, sz)
|
|
* which loads @sz bytes from @ptr as an unsigned @endian-order number
|
|
* and returns it in a uint64_t.
|
|
*/
|
|
|
|
static inline int ldub_p(const void *ptr)
|
|
{
|
|
return *(uint8_t *)ptr;
|
|
}
|
|
|
|
static inline int ldsb_p(const void *ptr)
|
|
{
|
|
return *(int8_t *)ptr;
|
|
}
|
|
|
|
static inline void stb_p(void *ptr, uint8_t v)
|
|
{
|
|
*(uint8_t *)ptr = v;
|
|
}
|
|
|
|
/*
|
|
* Any compiler worth its salt will turn these memcpy into native unaligned
|
|
* operations. Thus we don't need to play games with packed attributes, or
|
|
* inline byte-by-byte stores.
|
|
* Some compilation environments (eg some fortify-source implementations)
|
|
* may intercept memcpy() in a way that defeats the compiler optimization,
|
|
* though, so we use __builtin_memcpy() to give ourselves the best chance
|
|
* of good performance.
|
|
*/
|
|
|
|
static inline int lduw_he_p(const void *ptr)
|
|
{
|
|
uint16_t r;
|
|
__builtin_memcpy(&r, ptr, sizeof(r));
|
|
return r;
|
|
}
|
|
|
|
static inline int ldsw_he_p(const void *ptr)
|
|
{
|
|
int16_t r;
|
|
__builtin_memcpy(&r, ptr, sizeof(r));
|
|
return r;
|
|
}
|
|
|
|
static inline void stw_he_p(void *ptr, uint16_t v)
|
|
{
|
|
__builtin_memcpy(ptr, &v, sizeof(v));
|
|
}
|
|
|
|
static inline void st24_he_p(void *ptr, uint32_t v)
|
|
{
|
|
__builtin_memcpy(ptr, &v, 3);
|
|
}
|
|
|
|
static inline int ldl_he_p(const void *ptr)
|
|
{
|
|
int32_t r;
|
|
__builtin_memcpy(&r, ptr, sizeof(r));
|
|
return r;
|
|
}
|
|
|
|
static inline void stl_he_p(void *ptr, uint32_t v)
|
|
{
|
|
__builtin_memcpy(ptr, &v, sizeof(v));
|
|
}
|
|
|
|
static inline uint64_t ldq_he_p(const void *ptr)
|
|
{
|
|
uint64_t r;
|
|
__builtin_memcpy(&r, ptr, sizeof(r));
|
|
return r;
|
|
}
|
|
|
|
static inline void stq_he_p(void *ptr, uint64_t v)
|
|
{
|
|
__builtin_memcpy(ptr, &v, sizeof(v));
|
|
}
|
|
|
|
static inline int lduw_le_p(const void *ptr)
|
|
{
|
|
return (uint16_t)le_bswap(lduw_he_p(ptr), 16);
|
|
}
|
|
|
|
static inline int ldsw_le_p(const void *ptr)
|
|
{
|
|
return (int16_t)le_bswap(lduw_he_p(ptr), 16);
|
|
}
|
|
|
|
static inline int ldl_le_p(const void *ptr)
|
|
{
|
|
return le_bswap(ldl_he_p(ptr), 32);
|
|
}
|
|
|
|
static inline uint64_t ldq_le_p(const void *ptr)
|
|
{
|
|
return le_bswap(ldq_he_p(ptr), 64);
|
|
}
|
|
|
|
static inline void stw_le_p(void *ptr, uint16_t v)
|
|
{
|
|
stw_he_p(ptr, le_bswap(v, 16));
|
|
}
|
|
|
|
static inline void st24_le_p(void *ptr, uint32_t v)
|
|
{
|
|
st24_he_p(ptr, le_bswap24(v));
|
|
}
|
|
|
|
static inline void stl_le_p(void *ptr, uint32_t v)
|
|
{
|
|
stl_he_p(ptr, le_bswap(v, 32));
|
|
}
|
|
|
|
static inline void stq_le_p(void *ptr, uint64_t v)
|
|
{
|
|
stq_he_p(ptr, le_bswap(v, 64));
|
|
}
|
|
|
|
static inline int lduw_be_p(const void *ptr)
|
|
{
|
|
return (uint16_t)be_bswap(lduw_he_p(ptr), 16);
|
|
}
|
|
|
|
static inline int ldsw_be_p(const void *ptr)
|
|
{
|
|
return (int16_t)be_bswap(lduw_he_p(ptr), 16);
|
|
}
|
|
|
|
static inline int ldl_be_p(const void *ptr)
|
|
{
|
|
return be_bswap(ldl_he_p(ptr), 32);
|
|
}
|
|
|
|
static inline uint64_t ldq_be_p(const void *ptr)
|
|
{
|
|
return be_bswap(ldq_he_p(ptr), 64);
|
|
}
|
|
|
|
static inline void stw_be_p(void *ptr, uint16_t v)
|
|
{
|
|
stw_he_p(ptr, be_bswap(v, 16));
|
|
}
|
|
|
|
static inline void stl_be_p(void *ptr, uint32_t v)
|
|
{
|
|
stl_he_p(ptr, be_bswap(v, 32));
|
|
}
|
|
|
|
static inline void stq_be_p(void *ptr, uint64_t v)
|
|
{
|
|
stq_he_p(ptr, be_bswap(v, 64));
|
|
}
|
|
|
|
static inline unsigned long leul_to_cpu(unsigned long v)
|
|
{
|
|
#if HOST_LONG_BITS == 32
|
|
return le_bswap(v, 32);
|
|
#elif HOST_LONG_BITS == 64
|
|
return le_bswap(v, 64);
|
|
#else
|
|
# error Unknown sizeof long
|
|
#endif
|
|
}
|
|
|
|
/* Store v to p as a sz byte value in host order */
|
|
#define DO_STN_LDN_P(END) \
|
|
static inline void stn_## END ## _p(void *ptr, int sz, uint64_t v) \
|
|
{ \
|
|
switch (sz) { \
|
|
case 1: \
|
|
stb_p(ptr, v); \
|
|
break; \
|
|
case 2: \
|
|
stw_ ## END ## _p(ptr, v); \
|
|
break; \
|
|
case 4: \
|
|
stl_ ## END ## _p(ptr, v); \
|
|
break; \
|
|
case 8: \
|
|
stq_ ## END ## _p(ptr, v); \
|
|
break; \
|
|
default: \
|
|
g_assert_not_reached(); \
|
|
} \
|
|
} \
|
|
static inline uint64_t ldn_## END ## _p(const void *ptr, int sz) \
|
|
{ \
|
|
switch (sz) { \
|
|
case 1: \
|
|
return ldub_p(ptr); \
|
|
case 2: \
|
|
return lduw_ ## END ## _p(ptr); \
|
|
case 4: \
|
|
return (uint32_t)ldl_ ## END ## _p(ptr); \
|
|
case 8: \
|
|
return ldq_ ## END ## _p(ptr); \
|
|
default: \
|
|
g_assert_not_reached(); \
|
|
} \
|
|
}
|
|
|
|
DO_STN_LDN_P(he)
|
|
DO_STN_LDN_P(le)
|
|
DO_STN_LDN_P(be)
|
|
|
|
#undef DO_STN_LDN_P
|
|
|
|
#undef le_bswap
|
|
#undef be_bswap
|
|
#undef le_bswaps
|
|
#undef be_bswaps
|
|
|
|
#endif /* BSWAP_H */
|