|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH 04/22] xen/arm: p2m: Fix multi-lines coding style comments
On Wed, 20 Jul 2016, Julien Grall wrote:
> The start and end markers should be on separate lines.
>
> Signed-off-by: Julien Grall <julien.grall@xxxxxxx>
Reviewed-by: Stefano Stabellini <sstabellini@xxxxxxxxxx>
> xen/arch/arm/p2m.c | 35 ++++++++++++++++++++++------------
> xen/include/asm-arm/p2m.h | 48
> +++++++++++++++++++++++++++++++----------------
> 2 files changed, 55 insertions(+), 28 deletions(-)
>
> diff --git a/xen/arch/arm/p2m.c b/xen/arch/arm/p2m.c
> index 64d84cc..79095f1 100644
> --- a/xen/arch/arm/p2m.c
> +++ b/xen/arch/arm/p2m.c
> @@ -33,9 +33,11 @@ static bool_t p2m_valid(lpae_t pte)
> {
> return pte.p2m.valid;
> }
> -/* These two can only be used on L0..L2 ptes because L3 mappings set
> +/*
> + * These two can only be used on L0..L2 ptes because L3 mappings set
> * the table bit and therefore these would return the opposite to what
> - * you would expect. */
> + * you would expect.
> + */
> static bool_t p2m_table(lpae_t pte)
> {
> return p2m_valid(pte) && pte.p2m.table;
> @@ -119,7 +121,8 @@ void flush_tlb_domain(struct domain *d)
> {
> unsigned long flags = 0;
>
> - /* Update the VTTBR if necessary with the domain d. In this case,
> + /*
> + * Update the VTTBR if necessary with the domain d. In this case,
> * it's only necessary to flush TLBs on every CPUs with the current VMID
> * (our domain).
> */
> @@ -325,8 +328,10 @@ static lpae_t mfn_to_p2m_entry(unsigned long mfn,
> unsigned int mattr,
> p2m_type_t t, p2m_access_t a)
> {
> paddr_t pa = ((paddr_t) mfn) << PAGE_SHIFT;
> - /* sh, xn and write bit will be defined in the following switches
> - * based on mattr and t. */
> + /*
> + * sh, xn and write bit will be defined in the following switches
> + * based on mattr and t.
> + */
> lpae_t e = (lpae_t) {
> .p2m.af = 1,
> .p2m.read = 1,
> @@ -552,15 +557,17 @@ enum p2m_operation {
> MEMACCESS,
> };
>
> -/* Put any references on the single 4K page referenced by pte. TODO:
> - * Handle superpages, for now we only take special references for leaf
> +/*
> + * Put any references on the single 4K page referenced by pte.
> + * TODO: Handle superpages, for now we only take special references for leaf
> * pages (specifically foreign ones, which can't be super mapped today).
> */
> static void p2m_put_l3_page(const lpae_t pte)
> {
> ASSERT(p2m_valid(pte));
>
> - /* TODO: Handle other p2m types
> + /*
> + * TODO: Handle other p2m types
> *
> * It's safe to do the put_page here because page_alloc will
> * flush the TLBs if the page is reallocated before the end of
> @@ -932,7 +939,8 @@ static int apply_p2m_changes(struct domain *d,
> PAGE_LIST_HEAD(free_pages);
> struct page_info *pg;
>
> - /* Some IOMMU don't support coherent PT walk. When the p2m is
> + /*
> + * Some IOMMU don't support coherent PT walk. When the p2m is
> * shared with the CPU, Xen has to make sure that the PT changes have
> * reached the memory
> */
> @@ -1275,7 +1283,8 @@ int p2m_alloc_table(struct domain *d)
> d->arch.vttbr = page_to_maddr(p2m->root)
> | ((uint64_t)p2m->vmid&0xff)<<48;
>
> - /* Make sure that all TLBs corresponding to the new VMID are flushed
> + /*
> + * Make sure that all TLBs corresponding to the new VMID are flushed
> * before using it
> */
> flush_tlb_domain(d);
> @@ -1290,8 +1299,10 @@ int p2m_alloc_table(struct domain *d)
>
> static spinlock_t vmid_alloc_lock = SPIN_LOCK_UNLOCKED;
>
> -/* VTTBR_EL2 VMID field is 8 bits. Using a bitmap here limits us to
> - * 256 concurrent domains. */
> +/*
> + * VTTBR_EL2 VMID field is 8 bits. Using a bitmap here limits us to
> + * 256 concurrent domains.
> + */
> static DECLARE_BITMAP(vmid_mask, MAX_VMID);
>
> void p2m_vmid_allocator_init(void)
> diff --git a/xen/include/asm-arm/p2m.h b/xen/include/asm-arm/p2m.h
> index 34096bc..8fe78c1 100644
> --- a/xen/include/asm-arm/p2m.h
> +++ b/xen/include/asm-arm/p2m.h
> @@ -31,12 +31,14 @@ struct p2m_domain {
> /* Current VMID in use */
> uint8_t vmid;
>
> - /* Highest guest frame that's ever been mapped in the p2m
> + /*
> + * Highest guest frame that's ever been mapped in the p2m
> * Only takes into account ram and foreign mapping
> */
> gfn_t max_mapped_gfn;
>
> - /* Lowest mapped gfn in the p2m. When releasing mapped gfn's in a
> + /*
> + * Lowest mapped gfn in the p2m. When releasing mapped gfn's in a
> * preemptible manner this is update to track recall where to
> * resume the search. Apart from during teardown this can only
> * decrease. */
> @@ -51,24 +53,31 @@ struct p2m_domain {
> unsigned long shattered[4];
> } stats;
>
> - /* If true, and an access fault comes in and there is no vm_event
> listener,
> - * pause domain. Otherwise, remove access restrictions. */
> + /*
> + * If true, and an access fault comes in and there is no vm_event
> listener,
> + * pause domain. Otherwise, remove access restrictions.
> + */
> bool_t access_required;
>
> /* Defines if mem_access is in use for the domain. */
> bool_t mem_access_enabled;
>
> - /* Default P2M access type for each page in the the domain: new pages,
> + /*
> + * Default P2M access type for each page in the the domain: new pages,
> * swapped in pages, cleared pages, and pages that are ambiguously
> - * retyped get this access type. See definition of p2m_access_t. */
> + * retyped get this access type. See definition of p2m_access_t.
> + */
> p2m_access_t default_access;
>
> - /* Radix tree to store the p2m_access_t settings as the pte's don't have
> - * enough available bits to store this information. */
> + /*
> + * Radix tree to store the p2m_access_t settings as the pte's don't have
> + * enough available bits to store this information.
> + */
> struct radix_tree_root mem_access_settings;
> };
>
> -/* List of possible type for each page in the p2m entry.
> +/*
> + * List of possible type for each page in the p2m entry.
> * The number of available bit per page in the pte for this purpose is 4
> bits.
> * So it's possible to only have 16 fields. If we run out of value in the
> * future, it's possible to use higher value for pseudo-type and don't store
> @@ -116,13 +125,15 @@ int p2m_init(struct domain *d);
> /* Return all the p2m resources to Xen. */
> void p2m_teardown(struct domain *d);
>
> -/* Remove mapping refcount on each mapping page in the p2m
> +/*
> + * Remove mapping refcount on each mapping page in the p2m
> *
> * TODO: For the moment only foreign mappings are handled
> */
> int relinquish_p2m_mapping(struct domain *d);
>
> -/* Allocate a new p2m table for a domain.
> +/*
> + * Allocate a new p2m table for a domain.
> *
> * Returns 0 for success or -errno.
> */
> @@ -181,8 +192,10 @@ mfn_t gfn_to_mfn(struct domain *d, gfn_t gfn);
> * Populate-on-demand
> */
>
> -/* Call when decreasing memory reservation to handle PoD entries properly.
> - * Will return '1' if all entries were handled and nothing more need be
> done.*/
> +/*
> + * Call when decreasing memory reservation to handle PoD entries properly.
> + * Will return '1' if all entries were handled and nothing more need be done.
> + */
> int
> p2m_pod_decrease_reservation(struct domain *d,
> xen_pfn_t gpfn,
> @@ -210,7 +223,8 @@ static inline struct page_info *get_page_from_gfn(
> return NULL;
> page = mfn_to_page(mfn);
>
> - /* get_page won't work on foreign mapping because the page doesn't
> + /*
> + * get_page won't work on foreign mapping because the page doesn't
> * belong to the current domain.
> */
> if ( p2mt == p2m_map_foreign )
> @@ -257,8 +271,10 @@ static inline bool_t p2m_vm_event_sanity_check(struct
> domain *d)
> return 1;
> }
>
> -/* Send mem event based on the access. Boolean return value indicates if trap
> - * needs to be injected into guest. */
> +/*
> + * Send mem event based on the access. Boolean return value indicates if trap
> + * needs to be injected into guest.
> + */
> bool_t p2m_mem_access_check(paddr_t gpa, vaddr_t gla, const struct npfec
> npfec);
>
> #endif /* _XEN_P2M_H */
> --
> 1.9.1
>
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
https://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |