[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [PATCH 1/8] x86/pdx: simplify calculation of domain struct allocation boundary
On Wed, Jun 11, 2025 at 06:58:31PM +0100, Andrew Cooper wrote: > On 11/06/2025 6:16 pm, Roger Pau Monne wrote: > > diff --git a/xen/arch/x86/domain.c b/xen/arch/x86/domain.c > > index 7536b6c8717e..2f438ce367cf 100644 > > --- a/xen/arch/x86/domain.c > > +++ b/xen/arch/x86/domain.c > > @@ -461,30 +461,6 @@ void domain_cpu_policy_changed(struct domain *d) > > } > > } > > > > -#if !defined(CONFIG_BIGMEM) && defined(CONFIG_PDX_COMPRESSION) > > -/* > > - * The hole may be at or above the 44-bit boundary, so we need to determine > > - * the total bit count until reaching 32 significant (not squashed out) > > bits > > - * in PFN representations. > > - * Note that the way "bits" gets initialized/updated/bounds-checked > > guarantees > > - * that the function will never return zero, and hence will never be called > > - * more than once (which is important due to it being deliberately placed > > in > > - * .init.text). > > - */ > > -static unsigned int __init noinline _domain_struct_bits(void) > > -{ > > - unsigned int bits = 32 + PAGE_SHIFT; > > - unsigned int sig = hweight32(~pfn_hole_mask); > > - unsigned int mask = pfn_hole_mask >> 32; > > - > > - for ( ; bits < BITS_PER_LONG && sig < 32; ++bits, mask >>= 1 ) > > - if ( !(mask & 1) ) > > - ++sig; > > - > > - return bits; > > -} > > -#endif > > - > > I'm very happy to see this disappear. Both because of a non-__init > function calling an __init function, and because this internal is just > horrible. > > > struct domain *alloc_domain_struct(void) > > { > > struct domain *d; > > @@ -498,14 +474,15 @@ struct domain *alloc_domain_struct(void) > > * On systems with CONFIG_BIGMEM there's no packing, and so there's no > > * such restriction. > > */ > > -#if defined(CONFIG_BIGMEM) || !defined(CONFIG_PDX_COMPRESSION) > > - const unsigned int bits = IS_ENABLED(CONFIG_BIGMEM) ? 0 : > > - 32 + PAGE_SHIFT; > > +#if defined(CONFIG_BIGMEM) > > + const unsigned int bits = 0; > > #else > > - static unsigned int __read_mostly bits; > > + static unsigned int __ro_after_init bits; > > > > if ( unlikely(!bits) ) > > - bits = _domain_struct_bits(); > > + bits = flsl(pfn_to_paddr(pdx_to_pfn( > > + 1UL << (sizeof(((struct page_info *)NULL)->v.inuse._domain) * > > 8)))) > > + - 1; > > I think this would benefit greatly by not being a oneliner. There's > sizeof_field() which helps a little. Oh, I missed that. I was expecting to find something like member_size() or similar. > But, isn't this UB with CONFIG_BIGMEM? You're shifting 1UL by 64. CONFIG_BIGMEM doesn't get here (see the #fidef above). > When __pdx_t is unsigned long, there's no bits restriction necessary. > Therefore, don't you want !bits && sizeof_field(...) < BYTES_PER_LONG as > the entry criteria? __pdx_t being unsigned long can only happen for CONFIG_BIGMEM, and that's still handled separately using a #if defined(CONFIG_BIGMEM) ... (which I should have converted to #ifdef instead). Thanks, Roger.
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |