[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH 1/4] xen/tmem: Switch to using bool
On Wed, Jun 28, 2017 at 12:16:19PM +0100, Andrew Cooper wrote: > * Drop redundant initialisers > * Style corrections while changing client_over_quota() > * Drop all write-only bools from do_tmem_op() s/write-only/useless write-only/ > > Signed-off-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx> Reviewed-by: Konrad Rzeszutek Wilk <konrad.wilk@xxxxxxxxxx> > --- > CC: Konrad Rzeszutek Wilk <konrad.wilk@xxxxxxxxxx> > --- > xen/common/tmem.c | 39 ++++++++++++++++----------------------- > xen/common/tmem_control.c | 16 ++++++++-------- > xen/common/tmem_xen.c | 12 ++++++------ > xen/include/xen/tmem_xen.h | 24 ++++++++++++------------ > 4 files changed, 42 insertions(+), 49 deletions(-) > > diff --git a/xen/common/tmem.c b/xen/common/tmem.c > index 306dda6..c955cf7 100644 > --- a/xen/common/tmem.c > +++ b/xen/common/tmem.c > @@ -71,7 +71,7 @@ struct tmem_page_descriptor { > pagesize_t size; /* 0 == PAGE_SIZE (pfp), -1 == data invalid, > else compressed data (cdata). */ > uint32_t index; > - bool_t eviction_attempted; /* CHANGE TO lifetimes? (settable). */ > + bool eviction_attempted; /* CHANGE TO lifetimes? (settable). */ > union { > struct page_info *pfp; /* Page frame pointer. */ > char *cdata; /* Compressed data. */ > @@ -884,39 +884,41 @@ static void client_flush(struct client *client) > client_free(client); > } > > -static bool_t client_over_quota(struct client *client) > +static bool client_over_quota(const struct client *client) > { > int total = _atomic_read(tmem_global.client_weight_total); > > ASSERT(client != NULL); > if ( (total == 0) || (client->info.weight == 0) || > (client->eph_count == 0) ) > - return 0; > - return ( ((tmem_global.eph_count*100L) / client->eph_count ) > > - ((total*100L) / client->info.weight) ); > + return false; > + > + return (((tmem_global.eph_count * 100L) / client->eph_count) > > + ((total * 100L) / client->info.weight)); > } > > /************ MEMORY REVOCATION ROUTINES *******************************/ > > -static bool_t tmem_try_to_evict_pgp(struct tmem_page_descriptor *pgp, bool_t > *hold_pool_rwlock) > +static bool tmem_try_to_evict_pgp(struct tmem_page_descriptor *pgp, > + bool *hold_pool_rwlock) > { > struct tmem_object_root *obj = pgp->us.obj; > struct tmem_pool *pool = obj->pool; > > if ( pool->is_dying ) > - return 0; > + return false; > if ( spin_trylock(&obj->obj_spinlock) ) > { > if ( obj->pgp_count > 1 ) > - return 1; > + return true; > if ( write_trylock(&pool->pool_rwlock) ) > { > *hold_pool_rwlock = 1; > - return 1; > + return true; > } > spin_unlock(&obj->obj_spinlock); > } > - return 0; > + return false; > } > > int tmem_evict(void) > @@ -926,7 +928,7 @@ int tmem_evict(void) > struct tmem_object_root *obj; > struct tmem_pool *pool; > int ret = 0; > - bool_t hold_pool_rwlock = 0; > + bool hold_pool_rwlock = false; > > tmem_stats.evict_attempts++; > spin_lock(&eph_lists_spinlock); > @@ -995,7 +997,7 @@ int tmem_evict(void) > * is a minimum amount of memory (1MB) available BEFORE any data structure > * locks are held. > */ > -static inline bool_t tmem_ensure_avail_pages(void) > +static inline bool tmem_ensure_avail_pages(void) > { > int failed_evict = 10; > unsigned long free_mem; > @@ -1004,12 +1006,12 @@ static inline bool_t tmem_ensure_avail_pages(void) > free_mem = (tmem_page_list_pages + total_free_pages()) > >> (20 - PAGE_SHIFT); > if ( free_mem ) > - return 1; > + return true; > if ( !tmem_evict() ) > failed_evict--; > } while ( failed_evict > 0 ); > > - return 0; > + return false; > } > > /************ TMEM CORE OPERATIONS ************************************/ > @@ -1879,9 +1881,6 @@ long do_tmem_op(tmem_cli_op_t uops) > struct tmem_pool *pool = NULL; > struct xen_tmem_oid *oidp; > int rc = 0; > - bool_t succ_get = 0, succ_put = 0; > - bool_t non_succ_get = 0, non_succ_put = 0; > - bool_t flush = 0, flush_obj = 0; > > if ( !tmem_initialized ) > return -ENODEV; > @@ -1965,22 +1964,16 @@ long do_tmem_op(tmem_cli_op_t uops) > tmem_cli_buf_null); > else > rc = -ENOMEM; > - if (rc == 1) succ_put = 1; > - else non_succ_put = 1; > break; > case TMEM_GET_PAGE: > rc = do_tmem_get(pool, oidp, op.u.gen.index, op.u.gen.cmfn, > tmem_cli_buf_null); > - if (rc == 1) succ_get = 1; > - else non_succ_get = 1; > break; > case TMEM_FLUSH_PAGE: > - flush = 1; > rc = do_tmem_flush_page(pool, oidp, op.u.gen.index); > break; > case TMEM_FLUSH_OBJECT: > rc = do_tmem_flush_object(pool, oidp); > - flush_obj = 1; > break; > default: > tmem_client_warn("tmem: op %d not implemented\n", op.cmd); > diff --git a/xen/common/tmem_control.c b/xen/common/tmem_control.c > index 2d980e3..30bf6fb 100644 > --- a/xen/common/tmem_control.c > +++ b/xen/common/tmem_control.c > @@ -19,8 +19,8 @@ > static int tmemc_freeze_pools(domid_t cli_id, int arg) > { > struct client *client; > - bool_t freeze = (arg == XEN_SYSCTL_TMEM_OP_FREEZE) ? 1 : 0; > - bool_t destroy = (arg == XEN_SYSCTL_TMEM_OP_DESTROY) ? 1 : 0; > + bool freeze = arg == XEN_SYSCTL_TMEM_OP_FREEZE; > + bool destroy = arg == XEN_SYSCTL_TMEM_OP_DESTROY; > char *s; > > s = destroy ? "destroyed" : ( freeze ? "frozen" : "thawed" ); > @@ -96,12 +96,12 @@ static int tmemc_flush_mem(domid_t cli_id, uint32_t kb) > #define BSIZE 1024 > > static int tmemc_list_client(struct client *c, tmem_cli_va_param_t buf, > - int off, uint32_t len, bool_t use_long) > + int off, uint32_t len, bool use_long) > { > char info[BSIZE]; > int i, n = 0, sum = 0; > struct tmem_pool *p; > - bool_t s; > + bool s; > > n = scnprintf(info,BSIZE,"C=CI:%d,ww:%d,co:%d,fr:%d," > "Tc:%"PRIu64",Ge:%ld,Pp:%ld,Gp:%ld%c", > @@ -149,7 +149,7 @@ static int tmemc_list_client(struct client *c, > tmem_cli_va_param_t buf, > } > > static int tmemc_list_shared(tmem_cli_va_param_t buf, int off, uint32_t len, > - bool_t use_long) > + bool use_long) > { > char info[BSIZE]; > int i, n = 0, sum = 0; > @@ -188,7 +188,7 @@ static int tmemc_list_shared(tmem_cli_va_param_t buf, int > off, uint32_t len, > } > > static int tmemc_list_global_perf(tmem_cli_va_param_t buf, int off, > - uint32_t len, bool_t use_long) > + uint32_t len, bool use_long) > { > char info[BSIZE]; > int n = 0, sum = 0; > @@ -204,7 +204,7 @@ static int tmemc_list_global_perf(tmem_cli_va_param_t > buf, int off, > } > > static int tmemc_list_global(tmem_cli_va_param_t buf, int off, uint32_t len, > - bool_t use_long) > + bool use_long) > { > char info[BSIZE]; > int n = 0, sum = off; > @@ -238,7 +238,7 @@ static int tmemc_list_global(tmem_cli_va_param_t buf, int > off, uint32_t len, > } > > static int tmemc_list(domid_t cli_id, tmem_cli_va_param_t buf, uint32_t len, > - bool_t use_long) > + bool use_long) > { > struct client *client; > int off = 0; > diff --git a/xen/common/tmem_xen.c b/xen/common/tmem_xen.c > index 725ae93..20f74b2 100644 > --- a/xen/common/tmem_xen.c > +++ b/xen/common/tmem_xen.c > @@ -14,10 +14,10 @@ > #include <xen/cpu.h> > #include <xen/init.h> > > -bool_t __read_mostly opt_tmem = 0; > +bool __read_mostly opt_tmem; > boolean_param("tmem", opt_tmem); > > -bool_t __read_mostly opt_tmem_compress = 0; > +bool __read_mostly opt_tmem_compress; > boolean_param("tmem_compress", opt_tmem_compress); > > atomic_t freeable_page_count = ATOMIC_INIT(0); > @@ -32,14 +32,14 @@ static DEFINE_PER_CPU_READ_MOSTLY(void *, scratch_page); > > #if defined(CONFIG_ARM) > static inline void *cli_get_page(xen_pfn_t cmfn, unsigned long *pcli_mfn, > - struct page_info **pcli_pfp, bool_t > cli_write) > + struct page_info **pcli_pfp, bool cli_write) > { > ASSERT_UNREACHABLE(); > return NULL; > } > > static inline void cli_put_page(void *cli_va, struct page_info *cli_pfp, > - unsigned long cli_mfn, bool_t mark_dirty) > + unsigned long cli_mfn, bool mark_dirty) > { > ASSERT_UNREACHABLE(); > } > @@ -47,7 +47,7 @@ static inline void cli_put_page(void *cli_va, struct > page_info *cli_pfp, > #include <asm/p2m.h> > > static inline void *cli_get_page(xen_pfn_t cmfn, unsigned long *pcli_mfn, > - struct page_info **pcli_pfp, bool_t > cli_write) > + struct page_info **pcli_pfp, bool cli_write) > { > p2m_type_t t; > struct page_info *page; > @@ -72,7 +72,7 @@ static inline void *cli_get_page(xen_pfn_t cmfn, unsigned > long *pcli_mfn, > } > > static inline void cli_put_page(void *cli_va, struct page_info *cli_pfp, > - unsigned long cli_mfn, bool_t mark_dirty) > + unsigned long cli_mfn, bool mark_dirty) > { > if ( mark_dirty ) > { > diff --git a/xen/include/xen/tmem_xen.h b/xen/include/xen/tmem_xen.h > index dc5888c..542c0b3 100644 > --- a/xen/include/xen/tmem_xen.h > +++ b/xen/include/xen/tmem_xen.h > @@ -35,27 +35,27 @@ extern atomic_t freeable_page_count; > extern int tmem_init(void); > #define tmem_hash hash_long > > -extern bool_t opt_tmem_compress; > -static inline bool_t tmem_compression_enabled(void) > +extern bool opt_tmem_compress; > +static inline bool tmem_compression_enabled(void) > { > return opt_tmem_compress; > } > > #ifdef CONFIG_TMEM > -extern bool_t opt_tmem; > -static inline bool_t tmem_enabled(void) > +extern bool opt_tmem; > +static inline bool tmem_enabled(void) > { > return opt_tmem; > } > > static inline void tmem_disable(void) > { > - opt_tmem = 0; > + opt_tmem = false; > } > #else > -static inline bool_t tmem_enabled(void) > +static inline bool tmem_enabled(void) > { > - return 0; > + return false; > } > > static inline void tmem_disable(void) > @@ -266,7 +266,7 @@ struct tmem_global { > struct list_head ephemeral_page_list; /* All pages in ephemeral pools. > */ > struct list_head client_list; > struct tmem_pool *shared_pools[MAX_GLOBAL_SHARED_POOLS]; > - bool_t shared_auth; > + bool shared_auth; > long eph_count; /* Atomicity depends on eph_lists_spinlock. */ > atomic_t client_weight_total; > }; > @@ -286,7 +286,7 @@ struct client { > domid_t cli_id; > xen_tmem_client_t info; > /* For save/restore/migration. */ > - bool_t was_frozen; > + bool was_frozen; > struct list_head persistent_invalidated_list; > struct tmem_page_descriptor *cur_pgp; > /* Statistics collection. */ > @@ -307,9 +307,9 @@ struct client { > #define is_shared(_p) (_p->shared) > > struct tmem_pool { > - bool_t shared; > - bool_t persistent; > - bool_t is_dying; > + bool shared; > + bool persistent; > + bool is_dying; > struct client *client; > uint64_t uuid[2]; /* 0 for private, non-zero for shared. */ > uint32_t pool_id; > -- > 2.1.4 > _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx https://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |