[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH 1/4] xen/tmem: Switch to using bool
* Drop redundant initialisers * Style corrections while changing client_over_quota() * Drop all write-only bools from do_tmem_op() Signed-off-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx> --- CC: Konrad Rzeszutek Wilk <konrad.wilk@xxxxxxxxxx> --- xen/common/tmem.c | 39 ++++++++++++++++----------------------- xen/common/tmem_control.c | 16 ++++++++-------- xen/common/tmem_xen.c | 12 ++++++------ xen/include/xen/tmem_xen.h | 24 ++++++++++++------------ 4 files changed, 42 insertions(+), 49 deletions(-) diff --git a/xen/common/tmem.c b/xen/common/tmem.c index 306dda6..c955cf7 100644 --- a/xen/common/tmem.c +++ b/xen/common/tmem.c @@ -71,7 +71,7 @@ struct tmem_page_descriptor { pagesize_t size; /* 0 == PAGE_SIZE (pfp), -1 == data invalid, else compressed data (cdata). */ uint32_t index; - bool_t eviction_attempted; /* CHANGE TO lifetimes? (settable). */ + bool eviction_attempted; /* CHANGE TO lifetimes? (settable). */ union { struct page_info *pfp; /* Page frame pointer. */ char *cdata; /* Compressed data. */ @@ -884,39 +884,41 @@ static void client_flush(struct client *client) client_free(client); } -static bool_t client_over_quota(struct client *client) +static bool client_over_quota(const struct client *client) { int total = _atomic_read(tmem_global.client_weight_total); ASSERT(client != NULL); if ( (total == 0) || (client->info.weight == 0) || (client->eph_count == 0) ) - return 0; - return ( ((tmem_global.eph_count*100L) / client->eph_count ) > - ((total*100L) / client->info.weight) ); + return false; + + return (((tmem_global.eph_count * 100L) / client->eph_count) > + ((total * 100L) / client->info.weight)); } /************ MEMORY REVOCATION ROUTINES *******************************/ -static bool_t tmem_try_to_evict_pgp(struct tmem_page_descriptor *pgp, bool_t *hold_pool_rwlock) +static bool tmem_try_to_evict_pgp(struct tmem_page_descriptor *pgp, + bool *hold_pool_rwlock) { struct tmem_object_root *obj = pgp->us.obj; struct tmem_pool *pool = obj->pool; if ( pool->is_dying ) - return 0; + return false; if ( spin_trylock(&obj->obj_spinlock) ) { if ( obj->pgp_count > 1 ) - return 1; + return true; if ( write_trylock(&pool->pool_rwlock) ) { *hold_pool_rwlock = 1; - return 1; + return true; } spin_unlock(&obj->obj_spinlock); } - return 0; + return false; } int tmem_evict(void) @@ -926,7 +928,7 @@ int tmem_evict(void) struct tmem_object_root *obj; struct tmem_pool *pool; int ret = 0; - bool_t hold_pool_rwlock = 0; + bool hold_pool_rwlock = false; tmem_stats.evict_attempts++; spin_lock(&eph_lists_spinlock); @@ -995,7 +997,7 @@ int tmem_evict(void) * is a minimum amount of memory (1MB) available BEFORE any data structure * locks are held. */ -static inline bool_t tmem_ensure_avail_pages(void) +static inline bool tmem_ensure_avail_pages(void) { int failed_evict = 10; unsigned long free_mem; @@ -1004,12 +1006,12 @@ static inline bool_t tmem_ensure_avail_pages(void) free_mem = (tmem_page_list_pages + total_free_pages()) >> (20 - PAGE_SHIFT); if ( free_mem ) - return 1; + return true; if ( !tmem_evict() ) failed_evict--; } while ( failed_evict > 0 ); - return 0; + return false; } /************ TMEM CORE OPERATIONS ************************************/ @@ -1879,9 +1881,6 @@ long do_tmem_op(tmem_cli_op_t uops) struct tmem_pool *pool = NULL; struct xen_tmem_oid *oidp; int rc = 0; - bool_t succ_get = 0, succ_put = 0; - bool_t non_succ_get = 0, non_succ_put = 0; - bool_t flush = 0, flush_obj = 0; if ( !tmem_initialized ) return -ENODEV; @@ -1965,22 +1964,16 @@ long do_tmem_op(tmem_cli_op_t uops) tmem_cli_buf_null); else rc = -ENOMEM; - if (rc == 1) succ_put = 1; - else non_succ_put = 1; break; case TMEM_GET_PAGE: rc = do_tmem_get(pool, oidp, op.u.gen.index, op.u.gen.cmfn, tmem_cli_buf_null); - if (rc == 1) succ_get = 1; - else non_succ_get = 1; break; case TMEM_FLUSH_PAGE: - flush = 1; rc = do_tmem_flush_page(pool, oidp, op.u.gen.index); break; case TMEM_FLUSH_OBJECT: rc = do_tmem_flush_object(pool, oidp); - flush_obj = 1; break; default: tmem_client_warn("tmem: op %d not implemented\n", op.cmd); diff --git a/xen/common/tmem_control.c b/xen/common/tmem_control.c index 2d980e3..30bf6fb 100644 --- a/xen/common/tmem_control.c +++ b/xen/common/tmem_control.c @@ -19,8 +19,8 @@ static int tmemc_freeze_pools(domid_t cli_id, int arg) { struct client *client; - bool_t freeze = (arg == XEN_SYSCTL_TMEM_OP_FREEZE) ? 1 : 0; - bool_t destroy = (arg == XEN_SYSCTL_TMEM_OP_DESTROY) ? 1 : 0; + bool freeze = arg == XEN_SYSCTL_TMEM_OP_FREEZE; + bool destroy = arg == XEN_SYSCTL_TMEM_OP_DESTROY; char *s; s = destroy ? "destroyed" : ( freeze ? "frozen" : "thawed" ); @@ -96,12 +96,12 @@ static int tmemc_flush_mem(domid_t cli_id, uint32_t kb) #define BSIZE 1024 static int tmemc_list_client(struct client *c, tmem_cli_va_param_t buf, - int off, uint32_t len, bool_t use_long) + int off, uint32_t len, bool use_long) { char info[BSIZE]; int i, n = 0, sum = 0; struct tmem_pool *p; - bool_t s; + bool s; n = scnprintf(info,BSIZE,"C=CI:%d,ww:%d,co:%d,fr:%d," "Tc:%"PRIu64",Ge:%ld,Pp:%ld,Gp:%ld%c", @@ -149,7 +149,7 @@ static int tmemc_list_client(struct client *c, tmem_cli_va_param_t buf, } static int tmemc_list_shared(tmem_cli_va_param_t buf, int off, uint32_t len, - bool_t use_long) + bool use_long) { char info[BSIZE]; int i, n = 0, sum = 0; @@ -188,7 +188,7 @@ static int tmemc_list_shared(tmem_cli_va_param_t buf, int off, uint32_t len, } static int tmemc_list_global_perf(tmem_cli_va_param_t buf, int off, - uint32_t len, bool_t use_long) + uint32_t len, bool use_long) { char info[BSIZE]; int n = 0, sum = 0; @@ -204,7 +204,7 @@ static int tmemc_list_global_perf(tmem_cli_va_param_t buf, int off, } static int tmemc_list_global(tmem_cli_va_param_t buf, int off, uint32_t len, - bool_t use_long) + bool use_long) { char info[BSIZE]; int n = 0, sum = off; @@ -238,7 +238,7 @@ static int tmemc_list_global(tmem_cli_va_param_t buf, int off, uint32_t len, } static int tmemc_list(domid_t cli_id, tmem_cli_va_param_t buf, uint32_t len, - bool_t use_long) + bool use_long) { struct client *client; int off = 0; diff --git a/xen/common/tmem_xen.c b/xen/common/tmem_xen.c index 725ae93..20f74b2 100644 --- a/xen/common/tmem_xen.c +++ b/xen/common/tmem_xen.c @@ -14,10 +14,10 @@ #include <xen/cpu.h> #include <xen/init.h> -bool_t __read_mostly opt_tmem = 0; +bool __read_mostly opt_tmem; boolean_param("tmem", opt_tmem); -bool_t __read_mostly opt_tmem_compress = 0; +bool __read_mostly opt_tmem_compress; boolean_param("tmem_compress", opt_tmem_compress); atomic_t freeable_page_count = ATOMIC_INIT(0); @@ -32,14 +32,14 @@ static DEFINE_PER_CPU_READ_MOSTLY(void *, scratch_page); #if defined(CONFIG_ARM) static inline void *cli_get_page(xen_pfn_t cmfn, unsigned long *pcli_mfn, - struct page_info **pcli_pfp, bool_t cli_write) + struct page_info **pcli_pfp, bool cli_write) { ASSERT_UNREACHABLE(); return NULL; } static inline void cli_put_page(void *cli_va, struct page_info *cli_pfp, - unsigned long cli_mfn, bool_t mark_dirty) + unsigned long cli_mfn, bool mark_dirty) { ASSERT_UNREACHABLE(); } @@ -47,7 +47,7 @@ static inline void cli_put_page(void *cli_va, struct page_info *cli_pfp, #include <asm/p2m.h> static inline void *cli_get_page(xen_pfn_t cmfn, unsigned long *pcli_mfn, - struct page_info **pcli_pfp, bool_t cli_write) + struct page_info **pcli_pfp, bool cli_write) { p2m_type_t t; struct page_info *page; @@ -72,7 +72,7 @@ static inline void *cli_get_page(xen_pfn_t cmfn, unsigned long *pcli_mfn, } static inline void cli_put_page(void *cli_va, struct page_info *cli_pfp, - unsigned long cli_mfn, bool_t mark_dirty) + unsigned long cli_mfn, bool mark_dirty) { if ( mark_dirty ) { diff --git a/xen/include/xen/tmem_xen.h b/xen/include/xen/tmem_xen.h index dc5888c..542c0b3 100644 --- a/xen/include/xen/tmem_xen.h +++ b/xen/include/xen/tmem_xen.h @@ -35,27 +35,27 @@ extern atomic_t freeable_page_count; extern int tmem_init(void); #define tmem_hash hash_long -extern bool_t opt_tmem_compress; -static inline bool_t tmem_compression_enabled(void) +extern bool opt_tmem_compress; +static inline bool tmem_compression_enabled(void) { return opt_tmem_compress; } #ifdef CONFIG_TMEM -extern bool_t opt_tmem; -static inline bool_t tmem_enabled(void) +extern bool opt_tmem; +static inline bool tmem_enabled(void) { return opt_tmem; } static inline void tmem_disable(void) { - opt_tmem = 0; + opt_tmem = false; } #else -static inline bool_t tmem_enabled(void) +static inline bool tmem_enabled(void) { - return 0; + return false; } static inline void tmem_disable(void) @@ -266,7 +266,7 @@ struct tmem_global { struct list_head ephemeral_page_list; /* All pages in ephemeral pools. */ struct list_head client_list; struct tmem_pool *shared_pools[MAX_GLOBAL_SHARED_POOLS]; - bool_t shared_auth; + bool shared_auth; long eph_count; /* Atomicity depends on eph_lists_spinlock. */ atomic_t client_weight_total; }; @@ -286,7 +286,7 @@ struct client { domid_t cli_id; xen_tmem_client_t info; /* For save/restore/migration. */ - bool_t was_frozen; + bool was_frozen; struct list_head persistent_invalidated_list; struct tmem_page_descriptor *cur_pgp; /* Statistics collection. */ @@ -307,9 +307,9 @@ struct client { #define is_shared(_p) (_p->shared) struct tmem_pool { - bool_t shared; - bool_t persistent; - bool_t is_dying; + bool shared; + bool persistent; + bool is_dying; struct client *client; uint64_t uuid[2]; /* 0 for private, non-zero for shared. */ uint32_t pool_id; -- 2.1.4 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx https://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |