[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [PATCH v2 2/2] types: replace remaining uses of s64
... and move the type itself to linux-compat.h. While doing so - correct the type of union uu's uq field in lib/divmod.c, - switch a few adjacent types as well, for (a little bit of) consistency. Signed-off-by: Jan Beulich <jbeulich@xxxxxxxx> --- v2: Split off ubsan.[ch] adjustments. Re-base. --- a/xen/arch/arm/arm64/cpufeature.c +++ b/xen/arch/arm/arm64/cpufeature.c @@ -484,8 +484,8 @@ static const struct arm64_ftr_bits ftr_r ARM64_FTR_END, }; -static u64 arm64_ftr_set_value(const struct arm64_ftr_bits *ftrp, s64 reg, - s64 ftr_val) +static uint64_t arm64_ftr_set_value(const struct arm64_ftr_bits *ftrp, + int64_t reg, int64_t ftr_val) { u64 mask = arm64_ftr_mask(ftrp); @@ -494,10 +494,10 @@ static u64 arm64_ftr_set_value(const str return reg; } -static s64 arm64_ftr_safe_value(const struct arm64_ftr_bits *ftrp, s64 new, - s64 cur) +static int64_t arm64_ftr_safe_value(const struct arm64_ftr_bits *ftrp, + int64_t new, int64_t cur) { - s64 ret = 0; + int64_t ret = 0; switch (ftrp->type) { case FTR_EXACT: @@ -532,8 +532,8 @@ static void sanitize_reg(u64 *cur_reg, u for (;ftrp->width != 0;ftrp++) { - s64 cur_field = arm64_ftr_value(ftrp, *cur_reg); - s64 new_field = arm64_ftr_value(ftrp, new_reg); + int64_t cur_field = arm64_ftr_value(ftrp, *cur_reg); + int64_t new_field = arm64_ftr_value(ftrp, new_reg); if (cur_field == new_field) continue; --- a/xen/arch/arm/arm64/livepatch.c +++ b/xen/arch/arm/arm64/livepatch.c @@ -119,7 +119,7 @@ static u64 do_reloc(enum aarch64_reloc_o static int reloc_data(enum aarch64_reloc_op op, void *place, u64 val, int len) { - s64 sval = do_reloc(op, place, val); + int64_t sval = do_reloc(op, place, val); switch ( len ) { @@ -136,7 +136,7 @@ static int reloc_data(enum aarch64_reloc break; case 64: - *(s64 *)place = sval; + *(int64_t *)place = sval; break; default: @@ -155,9 +155,9 @@ enum aarch64_insn_movw_imm_type { static int reloc_insn_movw(enum aarch64_reloc_op op, void *dest, u64 val, int lsb, enum aarch64_insn_movw_imm_type imm_type) { - u64 imm; - s64 sval; - u32 insn = *(u32 *)dest; + uint64_t imm; + int64_t sval; + uint32_t insn = *(uint32_t *)dest; sval = do_reloc(op, dest, val); imm = sval >> lsb; @@ -200,9 +200,9 @@ static int reloc_insn_movw(enum aarch64_ static int reloc_insn_imm(enum aarch64_reloc_op op, void *dest, u64 val, int lsb, int len, enum aarch64_insn_imm_type imm_type) { - u64 imm, imm_mask; - s64 sval; - u32 insn = *(u32 *)dest; + uint64_t imm, imm_mask; + int64_t sval; + uint32_t insn = *(uint32_t *)dest; /* Calculate the relocation value. */ sval = do_reloc(op, dest, val); @@ -220,7 +220,7 @@ static int reloc_insn_imm(enum aarch64_r * Extract the upper value bits (including the sign bit) and * shift them to bit 0. */ - sval = (s64)(sval & ~(imm_mask >> 1)) >> (len - 1); + sval = (int64_t)(sval & ~(imm_mask >> 1)) >> (len - 1); /* * Overflow has occurred if the upper bits are not all equal to --- a/xen/arch/arm/include/asm/arm64/cpufeature.h +++ b/xen/arch/arm/include/asm/arm64/cpufeature.h @@ -39,15 +39,15 @@ struct arm64_ftr_bits { bool visible; bool strict; /* CPU Sanity check: strict matching required ? */ enum ftr_type type; - u8 shift; - u8 width; - s64 safe_val; /* safe value for FTR_EXACT features */ + uint8_t shift; + uint8_t width; + int64_t safe_val; /* safe value for FTR_EXACT features */ }; static inline int attr_const cpuid_feature_extract_signed_field_width(u64 features, int field, int width) { - return (s64)(features << (64 - width - field)) >> (64 - width); + return (int64_t)(features << (64 - width - field)) >> (64 - width); } static inline int attr_const @@ -87,9 +87,12 @@ cpuid_feature_extract_field(u64 features return cpuid_feature_extract_field_width(features, field, 4, sign); } -static inline s64 arm64_ftr_value(const struct arm64_ftr_bits *ftrp, u64 val) +static inline int64_t arm64_ftr_value(const struct arm64_ftr_bits *ftrp, + uint64_t val) { - return (s64)cpuid_feature_extract_field_width(val, ftrp->shift, ftrp->width, ftrp->sign); + return (int64_t)cpuid_feature_extract_field_width(val, ftrp->shift, + ftrp->width, + ftrp->sign); } #endif /* _ASM_ARM_ARM64_CPUFEATURES_H */ --- a/xen/arch/x86/hvm/vlapic.c +++ b/xen/arch/x86/hvm/vlapic.c @@ -1208,7 +1208,7 @@ void vlapic_tdt_msr_set(struct vlapic *v if ( value > guest_tsc ) { uint64_t delta = gtsc_to_gtime(v->domain, value - guest_tsc); - delta = max_t(s64, delta, 0); + delta = max_t(int64_t, delta, 0); HVM_DBG_LOG(DBG_LEVEL_VLAPIC_TIMER, "delta[0x%016"PRIx64"]", delta); --- a/xen/arch/x86/include/asm/hvm/vcpu.h +++ b/xen/arch/x86/include/asm/hvm/vcpu.h @@ -134,8 +134,8 @@ struct hvm_vcpu { unsigned long hw_cr[5]; struct vlapic vlapic; - s64 cache_tsc_offset; - u64 guest_time; + int64_t cache_tsc_offset; + uint64_t guest_time; /* Lock and list for virtual platform timers. */ spinlock_t tm_lock; --- a/xen/arch/x86/time.c +++ b/xen/arch/x86/time.c @@ -66,10 +66,10 @@ struct cpu_time { struct platform_timesource { const char *id; const char *name; - u64 frequency; + uint64_t frequency; /* Post-init this hook may only be invoked via the read_counter() wrapper! */ - u64 (*read_counter)(void); - s64 (*init)(struct platform_timesource *); + uint64_t (*read_counter)(void); + int64_t (*init)(struct platform_timesource *); void (*resume)(struct platform_timesource *); int counter_bits; }; @@ -368,7 +368,7 @@ static u64 cf_check read_pit_count(void) return count32; } -static s64 __init cf_check init_pit(struct platform_timesource *pts) +static int64_t __init cf_check init_pit(struct platform_timesource *pts) { u8 portb = inb(0x61); u64 start, end; @@ -610,7 +610,7 @@ static u64 cf_check read_pmtimer_count(v return inl(pmtmr_ioport); } -static s64 __init cf_check init_pmtimer(struct platform_timesource *pts) +static int64_t __init cf_check init_pmtimer(struct platform_timesource *pts) { if ( !pmtmr_ioport || (pmtmr_width != 24 && pmtmr_width != 32) ) return 0; @@ -655,7 +655,7 @@ static unsigned int __initdata tsc_flags * Called in verify_tsc_reliability() under reliable TSC conditions * thus reusing all the checks already performed there. */ -static s64 __init cf_check init_tsc(struct platform_timesource *pts) +static int64_t __init cf_check init_tsc(struct platform_timesource *pts) { u64 ret = pts->frequency; @@ -1034,9 +1034,9 @@ static void __init reset_platform_timer( spin_unlock_irq(&platform_timer_lock); } -static s64 __init try_platform_timer(struct platform_timesource *pts) +static int64_t __init try_platform_timer(struct platform_timesource *pts) { - s64 rc = pts->init(pts); + int64_t rc = pts->init(pts); if ( rc <= 0 ) return rc; @@ -1070,7 +1070,7 @@ static u64 __init init_platform_timer(vo struct platform_timesource *pts = NULL; unsigned int i; - s64 rc = -1; + int64_t rc = -1; /* clocksource=tsc is initialized via __initcalls (when CPUs are up). */ if ( (opt_clocksource[0] != '\0') && strcmp(opt_clocksource, "tsc") ) @@ -1923,7 +1923,7 @@ static void cf_check local_time_calibrat * Weirdness can happen if we lose sync with the platform timer. * We could be smarter here: resync platform timer with local timer? */ - if ( ((s64)stime_elapsed64 < (EPOCH / 2)) ) + if ( ((int64_t)stime_elapsed64 < (EPOCH / 2)) ) goto out; /* @@ -2398,7 +2398,7 @@ static void __init tsc_check_writability write_tsc(tsc | (1ULL << 32)); tmp = rdtsc(); - if ( ABS((s64)tmp - (s64)tmp2) < (1LL << 31) ) + if ( ABS((int64_t)tmp - (int64_t)tmp2) < (1LL << 31) ) what = "only partially"; } else --- a/xen/drivers/acpi/apei/erst.c +++ b/xen/drivers/acpi/apei/erst.c @@ -105,7 +105,7 @@ static inline int erst_errno(int command static int erst_timedout(u64 *t, u64 spin_unit) { - if ((s64)*t < spin_unit) { + if ((int64_t)*t < spin_unit) { printk(XENLOG_WARNING "Firmware does not respond in time\n"); return 1; } --- a/xen/include/acpi/actypes.h +++ b/xen/include/acpi/actypes.h @@ -147,8 +147,8 @@ typedef int INT32; /*! [End] no source code translation !*/ -typedef u64 acpi_native_uint; -typedef s64 acpi_native_int; +typedef uint64_t acpi_native_uint; +typedef int64_t acpi_native_int; typedef u64 acpi_io_address; typedef u64 acpi_physical_address; --- a/xen/include/xen/linux-compat.h +++ b/xen/include/xen/linux-compat.h @@ -15,7 +15,7 @@ typedef int8_t s8, __s8; typedef uint8_t __u8; typedef int16_t s16, __s16; typedef int32_t s32, __s32; -typedef int64_t __s64; +typedef int64_t s64, __s64; typedef paddr_t phys_addr_t; --- a/xen/include/xen/time.h +++ b/xen/include/xen/time.h @@ -29,7 +29,7 @@ struct vcpu; * of real time into system time */ -typedef s64 s_time_t; +typedef int64_t s_time_t; #define PRI_stime PRId64 s_time_t get_s_time_fixed(u64 at_tick); --- a/xen/include/xen/types.h +++ b/xen/include/xen/types.h @@ -8,7 +8,6 @@ typedef uint8_t u8; typedef uint16_t u16, __u16; typedef uint32_t u32, __u32; -typedef int64_t s64; typedef uint64_t u64, __u64; #include <asm/types.h> --- a/xen/lib/divmod.c +++ b/xen/lib/divmod.c @@ -46,8 +46,8 @@ * one or more of the following formats. */ union uu { - s64 q; /* as a (signed) quad */ - s64 uq; /* as an unsigned quad */ + int64_t q; /* as a (signed) quad */ + uint64_t uq; /* as an unsigned quad */ long sl[2]; /* as two signed longs */ unsigned long ul[2]; /* as two unsigned longs */ }; @@ -72,7 +72,7 @@ union uu { * and assembly. */ #define CHAR_BIT 8 /* number of bits in a char */ -#define QUAD_BITS (sizeof(s64) * CHAR_BIT) +#define QUAD_BITS (sizeof(int64_t) * CHAR_BIT) #define LONG_BITS (sizeof(long) * CHAR_BIT) #define HALF_BITS (sizeof(long) * CHAR_BIT / 2) @@ -324,7 +324,7 @@ u64 __qdivrem(u64 uq, u64 vq, u64 *arq) * Divide two signed quads. * Truncates towards zero, as required by C99. */ -s64 __divdi3(s64 a, s64 b) +int64_t __divdi3(int64_t a, int64_t b) { u64 ua, ub, uq; int neg = (a < 0) ^ (b < 0); @@ -361,7 +361,7 @@ u64 __umoddi3(u64 a, u64 b) * 11 % -5 = 1 * -11 % -5 = -1 */ -s64 __moddi3(s64 a, s64 b) +int64_t __moddi3(int64_t a, int64_t b) { u64 ua, ub, urem; int neg = (a < 0); @@ -374,7 +374,7 @@ s64 __moddi3(s64 a, s64 b) /* * Quotient and remainder of unsigned long long division */ -s64 __ldivmod_helper(s64 a, s64 b, s64 *r) +int64_t __ldivmod_helper(int64_t a, int64_t b, int64_t *r) { u64 ua, ub, rem, quot;
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |