|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [PATCH v3 1/2] x86/mm: add API for marking only part of a MMIO page read only
On 21.05.2024 04:54, Marek Marczykowski-Górecki wrote:
> --- a/xen/arch/x86/include/asm/mm.h
> +++ b/xen/arch/x86/include/asm/mm.h
> @@ -522,9 +522,27 @@ extern struct rangeset *mmio_ro_ranges;
> void memguard_guard_stack(void *p);
> void memguard_unguard_stack(void *p);
>
> +/*
> + * Add more precise r/o marking for a MMIO page. Range specified here
> + * will still be R/O, but the rest of the page (not marked as R/O via another
> + * call) will have writes passed through.
> + * The start address and the size must be aligned to SUBPAGE_MMIO_RO_ALIGN.
> + *
> + * This API cannot be used for overlapping ranges, nor for pages already
> added
> + * to mmio_ro_ranges separately.
> + *
> + * Return values:
> + * - negative: error
> + * - 0: success
> + */
> +#define SUBPAGE_MMIO_RO_ALIGN 8
This isn't just alignment, but also (and perhaps more importantly) granularity.
I think the name wants to express this.
> @@ -4910,6 +4921,260 @@ long arch_memory_op(unsigned long cmd,
> XEN_GUEST_HANDLE_PARAM(void) arg)
> return rc;
> }
>
> +/*
> + * Mark part of the page as R/O.
> + * Returns:
> + * - 0 on success - first range in the page
> + * - 1 on success - subsequent range in the page
> + * - <0 on error
> + *
> + * This needs subpage_ro_lock already taken */
Nit: Comment style (full stop and */ on its own line).
> +static int __init subpage_mmio_ro_add_page(
> + mfn_t mfn, unsigned int offset_s, unsigned int offset_e)
> +{
> + struct subpage_ro_range *entry = NULL, *iter;
> + unsigned int i;
> +
> + list_for_each_entry(iter, &subpage_ro_ranges, list)
> + {
> + if ( mfn_eq(iter->mfn, mfn) )
> + {
> + entry = iter;
> + break;
> + }
> + }
> + if ( !entry )
> + {
> + /* iter == NULL marks it was a newly allocated entry */
> + iter = NULL;
> + entry = xzalloc(struct subpage_ro_range);
> + if ( !entry )
> + return -ENOMEM;
> + entry->mfn = mfn;
> + }
> +
> + for ( i = offset_s; i <= offset_e; i += SUBPAGE_MMIO_RO_ALIGN )
> + {
> + int oldbit = __test_and_set_bit(i / SUBPAGE_MMIO_RO_ALIGN,
> + entry->ro_qwords);
Why int, not bool?
> + ASSERT(!oldbit);
> + }
> +
> + if ( !iter )
> + list_add(&entry->list, &subpage_ro_ranges);
> +
> + return iter ? 1 : 0;
> +}
> +
> +/* This needs subpage_ro_lock already taken */
> +static void __init subpage_mmio_ro_remove_page(
> + mfn_t mfn,
> + int offset_s,
> + int offset_e)
Can either of these be negative? The more that ...
> +{
> + struct subpage_ro_range *entry = NULL, *iter;
> + unsigned int i;
... this is used ...
> + list_for_each_entry(iter, &subpage_ro_ranges, list)
> + {
> + if ( mfn_eq(iter->mfn, mfn) )
> + {
> + entry = iter;
> + break;
> + }
> + }
> + if ( !entry )
> + return;
> +
> + for ( i = offset_s; i <= offset_e; i += SUBPAGE_MMIO_RO_ALIGN )
... with both of them?
> + __clear_bit(i / SUBPAGE_MMIO_RO_ALIGN, entry->ro_qwords);
> +
> + if ( !bitmap_empty(entry->ro_qwords, PAGE_SIZE / SUBPAGE_MMIO_RO_ALIGN) )
> + return;
> +
> + list_del(&entry->list);
> + if ( entry->mapped )
> + iounmap(entry->mapped);
> + xfree(entry);
> +}
> +
> +int __init subpage_mmio_ro_add(
> + paddr_t start,
> + size_t size)
> +{
> + mfn_t mfn_start = maddr_to_mfn(start);
> + paddr_t end = start + size - 1;
> + mfn_t mfn_end = maddr_to_mfn(end);
> + unsigned int offset_end = 0;
> + int rc;
> + bool subpage_start, subpage_end;
> +
> + ASSERT(IS_ALIGNED(start, SUBPAGE_MMIO_RO_ALIGN));
> + ASSERT(IS_ALIGNED(size, SUBPAGE_MMIO_RO_ALIGN));
> + if ( !IS_ALIGNED(size, SUBPAGE_MMIO_RO_ALIGN) )
> + size = ROUNDUP(size, SUBPAGE_MMIO_RO_ALIGN);
> +
> + if ( !size )
> + return 0;
> +
> + if ( mfn_eq(mfn_start, mfn_end) )
> + {
> + /* Both starting and ending parts handled at once */
> + subpage_start = PAGE_OFFSET(start) || PAGE_OFFSET(end) != PAGE_SIZE
> - 1;
> + subpage_end = false;
> + }
> + else
> + {
> + subpage_start = PAGE_OFFSET(start);
> + subpage_end = PAGE_OFFSET(end) != PAGE_SIZE - 1;
> + }
> +
> + spin_lock(&subpage_ro_lock);
> +
> + if ( subpage_start )
> + {
> + offset_end = mfn_eq(mfn_start, mfn_end) ?
> + PAGE_OFFSET(end) :
> + (PAGE_SIZE - 1);
> + rc = subpage_mmio_ro_add_page(mfn_start,
> + PAGE_OFFSET(start),
> + offset_end);
> + if ( rc < 0 )
> + goto err_unlock;
> + /* Check if not marking R/W part of a page intended to be fully R/O
> */
> + ASSERT(rc || !rangeset_contains_singleton(mmio_ro_ranges,
> + mfn_x(mfn_start)));
> + }
> +
> + if ( subpage_end )
> + {
> + rc = subpage_mmio_ro_add_page(mfn_end, 0, PAGE_OFFSET(end));
> + if ( rc < 0 )
> + goto err_unlock_remove;
> + /* Check if not marking R/W part of a page intended to be fully R/O
> */
> + ASSERT(rc || !rangeset_contains_singleton(mmio_ro_ranges,
> + mfn_x(mfn_end)));
> + }
> +
> + spin_unlock(&subpage_ro_lock);
> +
> + rc = rangeset_add_range(mmio_ro_ranges, mfn_x(mfn_start),
> mfn_x(mfn_end));
> + if ( rc )
> + goto err_remove;
> +
> + return 0;
> +
> + err_remove:
> + spin_lock(&subpage_ro_lock);
> + if ( subpage_end )
> + subpage_mmio_ro_remove_page(mfn_end, 0, PAGE_OFFSET(end));
> + err_unlock_remove:
> + if ( subpage_start )
> + subpage_mmio_ro_remove_page(mfn_start, PAGE_OFFSET(start),
> offset_end);
> + err_unlock:
> + spin_unlock(&subpage_ro_lock);
> + return rc;
> +}
> +
> +static void __iomem *subpage_mmio_get_page(struct subpage_ro_range *entry)
> +{
> + void __iomem *mapped_page;
> +
> + if ( entry->mapped )
> + return entry->mapped;
> +
> + mapped_page = ioremap(mfn_x(entry->mfn) << PAGE_SHIFT, PAGE_SIZE);
mfn_to_maddr() or some such?
> + spin_lock(&subpage_ro_lock);
> + /* Re-check under the lock */
> + if ( entry->mapped )
> + {
> + spin_unlock(&subpage_ro_lock);
> + iounmap(mapped_page);
What if ioremap() failed?
Jan
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |