[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [PATCH v4 04/17] xen: mapcache: Refactor xen_map_cache for multi-instance
On Tue, 30 Apr 2024, Edgar E. Iglesias wrote: > From: "Edgar E. Iglesias" <edgar.iglesias@xxxxxxx> > > Make xen_map_cache take a MapCache as argument. This is in > prepaparation to support multiple map caches. > > No functional changes. > > Signed-off-by: Edgar E. Iglesias <edgar.iglesias@xxxxxxx> Reviewed-by: Stefano Stabellini <sstabellini@xxxxxxxxxx> > --- > hw/xen/xen-mapcache.c | 35 ++++++++++++++++++----------------- > 1 file changed, 18 insertions(+), 17 deletions(-) > > diff --git a/hw/xen/xen-mapcache.c b/hw/xen/xen-mapcache.c > index 3f11562075..896021d86f 100644 > --- a/hw/xen/xen-mapcache.c > +++ b/hw/xen/xen-mapcache.c > @@ -240,7 +240,8 @@ static void xen_remap_bucket(MapCacheEntry *entry, > g_free(err); > } > > -static uint8_t *xen_map_cache_unlocked(hwaddr phys_addr, hwaddr size, > +static uint8_t *xen_map_cache_unlocked(MapCache *mc, > + hwaddr phys_addr, hwaddr size, > uint8_t lock, bool dma) > { > MapCacheEntry *entry, *pentry = NULL, > @@ -269,16 +270,16 @@ tryagain: > test_bit_size = XC_PAGE_SIZE; > } > > - if (mapcache->last_entry != NULL && > - mapcache->last_entry->paddr_index == address_index && > + if (mc->last_entry != NULL && > + mc->last_entry->paddr_index == address_index && > !lock && !size && > test_bits(address_offset >> XC_PAGE_SHIFT, > test_bit_size >> XC_PAGE_SHIFT, > - mapcache->last_entry->valid_mapping)) { > + mc->last_entry->valid_mapping)) { > trace_xen_map_cache_return( > - mapcache->last_entry->vaddr_base + address_offset > + mc->last_entry->vaddr_base + address_offset > ); > - return mapcache->last_entry->vaddr_base + address_offset; > + return mc->last_entry->vaddr_base + address_offset; > } > > /* size is always a multiple of MCACHE_BUCKET_SIZE */ > @@ -291,7 +292,7 @@ tryagain: > cache_size = MCACHE_BUCKET_SIZE; > } > > - entry = &mapcache->entry[address_index % mapcache->nr_buckets]; > + entry = &mc->entry[address_index % mc->nr_buckets]; > > while (entry && (lock || entry->lock) && entry->vaddr_base && > (entry->paddr_index != address_index || entry->size != > cache_size || > @@ -326,10 +327,10 @@ tryagain: > if(!test_bits(address_offset >> XC_PAGE_SHIFT, > test_bit_size >> XC_PAGE_SHIFT, > entry->valid_mapping)) { > - mapcache->last_entry = NULL; > + mc->last_entry = NULL; > #ifdef XEN_COMPAT_PHYSMAP > - if (!translated && mapcache->phys_offset_to_gaddr) { > - phys_addr = mapcache->phys_offset_to_gaddr(phys_addr, size); > + if (!translated && mc->phys_offset_to_gaddr) { > + phys_addr = mc->phys_offset_to_gaddr(phys_addr, size); > translated = true; > goto tryagain; > } > @@ -342,7 +343,7 @@ tryagain: > return NULL; > } > > - mapcache->last_entry = entry; > + mc->last_entry = entry; > if (lock) { > MapCacheRev *reventry = g_new0(MapCacheRev, 1); > entry->lock++; > @@ -352,16 +353,16 @@ tryagain: > abort(); > } > reventry->dma = dma; > - reventry->vaddr_req = mapcache->last_entry->vaddr_base + > address_offset; > - reventry->paddr_index = mapcache->last_entry->paddr_index; > + reventry->vaddr_req = mc->last_entry->vaddr_base + address_offset; > + reventry->paddr_index = mc->last_entry->paddr_index; > reventry->size = entry->size; > - QTAILQ_INSERT_HEAD(&mapcache->locked_entries, reventry, next); > + QTAILQ_INSERT_HEAD(&mc->locked_entries, reventry, next); > } > > trace_xen_map_cache_return( > - mapcache->last_entry->vaddr_base + address_offset > + mc->last_entry->vaddr_base + address_offset > ); > - return mapcache->last_entry->vaddr_base + address_offset; > + return mc->last_entry->vaddr_base + address_offset; > } > > uint8_t *xen_map_cache(hwaddr phys_addr, hwaddr size, > @@ -370,7 +371,7 @@ uint8_t *xen_map_cache(hwaddr phys_addr, hwaddr size, > uint8_t *p; > > mapcache_lock(mapcache); > - p = xen_map_cache_unlocked(phys_addr, size, lock, dma); > + p = xen_map_cache_unlocked(mapcache, phys_addr, size, lock, dma); > mapcache_unlock(mapcache); > return p; > } > -- > 2.40.1 >
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |