[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [PATCH] xen: Consistently use alignof()
We have a mix of all 3 spellings in Xen, as well as having compatibility in compiler.h for older C standards. Remove the use of __alignof() and __alignof__(), which reduced code volume a little. No functional change. Signed-off-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx> --- CC: Jan Beulich <JBeulich@xxxxxxxx> CC: Roger Pau Monné <roger.pau@xxxxxxxxxx> CC: Stefano Stabellini <sstabellini@xxxxxxxxxx> CC: Julien Grall <julien@xxxxxxx> CC: Volodymyr Babchuk <Volodymyr_Babchuk@xxxxxxxx> CC: Bertrand Marquis <bertrand.marquis@xxxxxxx> CC: Michal Orzel <michal.orzel@xxxxxxx> x86_emulate.c and compat-build-header.py still use __alignof() because they have code used outside of Xen. The compiler.h compatibilty checks for < C11, whereas it was C23 when they became keywords. Nothing seems to mind, so I've left it alone. --- xen/arch/x86/xstate.c | 4 ++-- xen/common/coverage/gcc_3_4.c | 4 ++-- xen/common/device-tree/device-tree.c | 8 ++++---- xen/include/xen/config.h | 2 +- xen/include/xen/percpu.h | 2 +- xen/include/xen/xmalloc.h | 16 ++++++++-------- xen/include/xen/xvmalloc.h | 16 ++++++++-------- 7 files changed, 26 insertions(+), 26 deletions(-) diff --git a/xen/arch/x86/xstate.c b/xen/arch/x86/xstate.c index 3d249518a1b7..d423bf7978da 100644 --- a/xen/arch/x86/xstate.c +++ b/xen/arch/x86/xstate.c @@ -542,8 +542,8 @@ int xstate_alloc_save_area(struct vcpu *v) } /* XSAVE/XRSTOR requires the save area be 64-byte-boundary aligned. */ - BUILD_BUG_ON(__alignof(*save_area) < 64); - save_area = _xvzalloc(size, __alignof(*save_area)); + BUILD_BUG_ON(alignof(*save_area) < 64); + save_area = _xvzalloc(size, alignof(*save_area)); if ( save_area == NULL ) return -ENOMEM; diff --git a/xen/common/coverage/gcc_3_4.c b/xen/common/coverage/gcc_3_4.c index 3631f4bc2535..393e3c866d74 100644 --- a/xen/common/coverage/gcc_3_4.c +++ b/xen/common/coverage/gcc_3_4.c @@ -199,8 +199,8 @@ static size_t get_fn_size(const struct gcov_info *info) size = sizeof(struct gcov_fn_info) + num_counter_active(info) * sizeof(unsigned int); - if ( __alignof__(struct gcov_fn_info) > sizeof(unsigned int) ) - size = ROUNDUP(size, __alignof__(struct gcov_fn_info)); + if ( alignof(struct gcov_fn_info) > sizeof(unsigned int) ) + size = ROUNDUP(size, alignof(struct gcov_fn_info)); return size; } diff --git a/xen/common/device-tree/device-tree.c b/xen/common/device-tree/device-tree.c index 90fee2ba0315..e4dea4b42a1d 100644 --- a/xen/common/device-tree/device-tree.c +++ b/xen/common/device-tree/device-tree.c @@ -1885,7 +1885,7 @@ static unsigned long unflatten_dt_node(const void *fdt, } np = unflatten_dt_alloc(&mem, sizeof(struct dt_device_node) + allocl, - __alignof__(struct dt_device_node)); + alignof(struct dt_device_node)); if ( allnextpp ) { memset(np, 0, sizeof(*np)); @@ -1963,7 +1963,7 @@ static unsigned long unflatten_dt_node(const void *fdt, has_name = 1; l = strlen(pname) + 1; pp = unflatten_dt_alloc(&mem, sizeof(struct dt_property), - __alignof__(struct dt_property)); + alignof(struct dt_property)); if ( allnextpp ) { /* We accept flattened tree phandles either in @@ -2010,7 +2010,7 @@ static unsigned long unflatten_dt_node(const void *fdt, pa = p1; sz = (pa - ps) + 1; pp = unflatten_dt_alloc(&mem, sizeof(struct dt_property) + sz, - __alignof__(struct dt_property)); + alignof(struct dt_property)); if ( allnextpp ) { pp->name = "name"; @@ -2088,7 +2088,7 @@ int unflatten_device_tree(const void *fdt, struct dt_device_node **mynodes) dt_dprintk(" size is %#lx allocating...\n", size); /* Allocate memory for the expanded device tree */ - mem = (unsigned long)_xmalloc (size + 4, __alignof__(struct dt_device_node)); + mem = (unsigned long)_xmalloc(size + 4, alignof(struct dt_device_node)); if ( !mem ) return -ENOMEM; diff --git a/xen/include/xen/config.h b/xen/include/xen/config.h index 1d7195066c08..479c3bac8925 100644 --- a/xen/include/xen/config.h +++ b/xen/include/xen/config.h @@ -105,7 +105,7 @@ #define BITS_PER_LONG (BITS_PER_BYTE * BYTES_PER_LONG) #define BITS_PER_LLONG (BITS_PER_BYTE * __SIZEOF_LONG_LONG__) -/* It is assumed that sizeof(void *) == __alignof(void *) */ +/* It is assumed that sizeof(void *) == alignof(void *) */ #define POINTER_ALIGN __SIZEOF_POINTER__ #if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__ diff --git a/xen/include/xen/percpu.h b/xen/include/xen/percpu.h index e7f585c7ed69..62a6259b2bc1 100644 --- a/xen/include/xen/percpu.h +++ b/xen/include/xen/percpu.h @@ -18,7 +18,7 @@ #define DEFINE_PER_CPU_PAGE_ALIGNED(type, name) \ typedef char name ## _chk_t \ - [BUILD_BUG_ON_ZERO(__alignof(type) & (PAGE_SIZE - 1))]; \ + [BUILD_BUG_ON_ZERO(alignof(type) & (PAGE_SIZE - 1))]; \ __DEFINE_PER_CPU(__section(".bss.percpu.page_aligned"), \ type, _ ## name) diff --git a/xen/include/xen/xmalloc.h b/xen/include/xen/xmalloc.h index f0412fb4e021..b7616c3b9038 100644 --- a/xen/include/xen/xmalloc.h +++ b/xen/include/xen/xmalloc.h @@ -13,8 +13,8 @@ */ /* Allocate space for typed object. */ -#define xmalloc(_type) ((_type *)_xmalloc(sizeof(_type), __alignof__(_type))) -#define xzalloc(_type) ((_type *)_xzalloc(sizeof(_type), __alignof__(_type))) +#define xmalloc(_type) ((_type *)_xmalloc(sizeof(_type), alignof(_type))) +#define xzalloc(_type) ((_type *)_xzalloc(sizeof(_type), alignof(_type))) /* * Allocate space for a typed object and copy an existing instance. @@ -34,24 +34,24 @@ /* Allocate space for array of typed objects. */ #define xmalloc_array(_type, _num) \ - ((_type *)_xmalloc_array(sizeof(_type), __alignof__(_type), _num)) + ((_type *)_xmalloc_array(sizeof(_type), alignof(_type), _num)) #define xzalloc_array(_type, _num) \ - ((_type *)_xzalloc_array(sizeof(_type), __alignof__(_type), _num)) + ((_type *)_xzalloc_array(sizeof(_type), alignof(_type), _num)) #define xrealloc_array(_ptr, _num) \ ((typeof(_ptr))_xrealloc_array(_ptr, sizeof(typeof(*(_ptr))), \ - __alignof__(typeof(*(_ptr))), _num)) + alignof(typeof(*(_ptr))), _num)) /* Allocate space for a structure with a flexible array of typed objects. */ #define xzalloc_flex_struct(type, field, nr) \ - ((type *)_xzalloc(offsetof(type, field[nr]), __alignof__(type))) + ((type *)_xzalloc(offsetof(type, field[nr]), alignof(type))) #define xmalloc_flex_struct(type, field, nr) \ - ((type *)_xmalloc(offsetof(type, field[nr]), __alignof__(type))) + ((type *)_xmalloc(offsetof(type, field[nr]), alignof(type))) /* Re-allocate space for a structure with a flexible array of typed objects. */ #define xrealloc_flex_struct(ptr, field, nr) \ ((typeof(ptr))_xrealloc(ptr, offsetof(typeof(*(ptr)), field[nr]), \ - __alignof__(typeof(*(ptr))))) + alignof(typeof(*(ptr))))) /* Allocate untyped storage. */ #define xmalloc_bytes(_bytes) _xmalloc(_bytes, SMP_CACHE_BYTES) diff --git a/xen/include/xen/xvmalloc.h b/xen/include/xen/xvmalloc.h index 7686d49f8154..d2288c175814 100644 --- a/xen/include/xen/xvmalloc.h +++ b/xen/include/xen/xvmalloc.h @@ -10,13 +10,13 @@ */ /* Allocate space for typed object. */ -#define xvmalloc(_type) ((_type *)_xvmalloc(sizeof(_type), __alignof__(_type))) -#define xvzalloc(_type) ((_type *)_xvzalloc(sizeof(_type), __alignof__(_type))) +#define xvmalloc(_type) ((_type *)_xvmalloc(sizeof(_type), alignof(_type))) +#define xvzalloc(_type) ((_type *)_xvzalloc(sizeof(_type), alignof(_type))) /* Allocate space for a typed object and copy an existing instance. */ #define xvmemdup(ptr) \ ({ \ - void *p_ = _xvmalloc(sizeof(*(ptr)), __alignof__(*(ptr))); \ + void *p_ = _xvmalloc(sizeof(*(ptr)), alignof(*(ptr))); \ if ( p_ ) \ memcpy(p_, ptr, sizeof(*(ptr))); \ (typeof(*(ptr)) *)p_; \ @@ -24,21 +24,21 @@ /* Allocate space for array of typed objects. */ #define xvmalloc_array(_type, _num) \ - ((_type *)_xvmalloc_array(sizeof(_type), __alignof__(_type), _num)) + ((_type *)_xvmalloc_array(sizeof(_type), alignof(_type), _num)) #define xvzalloc_array(_type, _num) \ - ((_type *)_xvzalloc_array(sizeof(_type), __alignof__(_type), _num)) + ((_type *)_xvzalloc_array(sizeof(_type), alignof(_type), _num)) /* Allocate space for a structure with a flexible array of typed objects. */ #define xvzalloc_flex_struct(type, field, nr) \ - ((type *)_xvzalloc(offsetof(type, field[nr]), __alignof__(type))) + ((type *)_xvzalloc(offsetof(type, field[nr]), alignof(type))) #define xvmalloc_flex_struct(type, field, nr) \ - ((type *)_xvmalloc(offsetof(type, field[nr]), __alignof__(type))) + ((type *)_xvmalloc(offsetof(type, field[nr]), alignof(type))) /* Re-allocate space for a structure with a flexible array of typed objects. */ #define xvrealloc_flex_struct(ptr, field, nr) \ ((typeof(ptr))_xvrealloc(ptr, offsetof(typeof(*(ptr)), field[nr]), \ - __alignof__(typeof(*(ptr))))) + alignof(typeof(*(ptr))))) #ifdef CONFIG_HAS_VMAP -- 2.39.5
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |