diff options
author | Vsevolod Stakhov <vsevolod@highsecure.ru> | 2014-09-18 13:17:43 +0100 |
---|---|---|
committer | Vsevolod Stakhov <vsevolod@highsecure.ru> | 2014-09-18 13:17:43 +0100 |
commit | 6e280e579b2d510fd8c2d3abe3ca652448cdc4a5 (patch) | |
tree | f2da8b73716b85baaac0f8f7d0ef3f3657bf446b /src | |
parent | 8cf474069601dbb376de86dffae49988029794ee (diff) | |
download | rspamd-6e280e579b2d510fd8c2d3abe3ca652448cdc4a5.tar.gz rspamd-6e280e579b2d510fd8c2d3abe3ca652448cdc4a5.zip |
Some fixes to memory pools operations.
Diffstat (limited to 'src')
-rw-r--r-- | src/libutil/mem_pool.c | 70 | ||||
-rw-r--r-- | src/libutil/mem_pool.h | 14 |
2 files changed, 29 insertions, 55 deletions
diff --git a/src/libutil/mem_pool.c b/src/libutil/mem_pool.c index a40578e83..810c85d0c 100644 --- a/src/libutil/mem_pool.c +++ b/src/libutil/mem_pool.c @@ -33,15 +33,6 @@ #define MUTEX_SLEEP_TIME 10000000L #define MUTEX_SPIN_COUNT 100 -#ifdef _THREAD_SAFE -pthread_mutex_t stat_mtx = PTHREAD_MUTEX_INITIALIZER; -# define STAT_LOCK() do { pthread_mutex_lock (&stat_mtx); } while (0) -# define STAT_UNLOCK() do { pthread_mutex_unlock (&stat_mtx); } while (0) -#else -# define STAT_LOCK() do {} while (0) -# define STAT_UNLOCK() do {} while (0) -#endif - #define POOL_MTX_LOCK() do { rspamd_mutex_lock (pool->mtx); } while (0) #define POOL_MTX_UNLOCK() do { rspamd_mutex_unlock (pool->mtx); } while (0) @@ -95,10 +86,8 @@ pool_chain_new (gsize size) chain->pos = align_ptr (chain->begin, MEM_ALIGNMENT); chain->len = size; chain->next = NULL; - STAT_LOCK (); - mem_pool_stat->bytes_allocated += size; - mem_pool_stat->chunks_allocated++; - STAT_UNLOCK (); + g_atomic_int_add (&mem_pool_stat->bytes_allocated, size); + g_atomic_int_inc (&mem_pool_stat->chunks_allocated); return chain; } @@ -151,10 +140,9 @@ pool_chain_new_shared (gsize size) chain->len = size; chain->lock = NULL; chain->next = NULL; - STAT_LOCK (); - mem_pool_stat->shared_chunks_allocated++; - mem_pool_stat->bytes_allocated += size; - STAT_UNLOCK (); + + g_atomic_int_inc (&mem_pool_stat->shared_chunks_allocated); + g_atomic_int_add (&mem_pool_stat->bytes_allocated, size); return chain; } @@ -414,17 +402,14 @@ rspamd_mempool_alloc_shared (rspamd_mempool_t * pool, gsize size) /* Attach new pool to chain */ cur->next = new; new->pos += size; - STAT_LOCK (); - mem_pool_stat->bytes_allocated += size; - STAT_UNLOCK (); - POOL_MTX_UNLOCK () - ; + g_atomic_int_add (&mem_pool_stat->bytes_allocated, size); + + POOL_MTX_UNLOCK (); return new->begin; } tmp = align_ptr (cur->pos, MEM_ALIGNMENT); cur->pos = tmp + size; - POOL_MTX_UNLOCK () - ; + POOL_MTX_UNLOCK (); return tmp; } return NULL; @@ -517,13 +502,10 @@ __mutex_spin (rspamd_mempool_mutex_t * mutex) /* Spin again */ g_atomic_int_set (&mutex->spin, MUTEX_SPIN_COUNT); } -#ifdef HAVE_ASM_PAUSE - __asm __volatile ("pause"); -#elif defined(HAVE_SCHED_YIELD) - (void)sched_yield (); -#endif -#if defined(HAVE_NANOSLEEP) +#ifdef HAVE_SCHED_YIELD + (void)sched_yield (); +#elif defined(HAVE_NANOSLEEP) struct timespec ts; ts.tv_sec = 0; ts.tv_nsec = MUTEX_SLEEP_TIME; @@ -648,10 +630,8 @@ rspamd_mempool_delete (rspamd_mempool_t * pool) while (cur) { tmp = cur; cur = cur->next; - STAT_LOCK (); - mem_pool_stat->chunks_freed++; - mem_pool_stat->bytes_allocated -= tmp->len; - STAT_UNLOCK (); + g_atomic_int_inc (&mem_pool_stat->chunks_freed); + g_atomic_int_add (&mem_pool_stat->bytes_allocated, -tmp->len); g_slice_free1 (tmp->len, tmp->begin); g_slice_free (struct _pool_chain, tmp); } @@ -660,10 +640,8 @@ rspamd_mempool_delete (rspamd_mempool_t * pool) while (cur) { tmp = cur; cur = cur->next; - STAT_LOCK (); - mem_pool_stat->chunks_freed++; - mem_pool_stat->bytes_allocated -= tmp->len; - STAT_UNLOCK (); + g_atomic_int_inc (&mem_pool_stat->chunks_freed); + g_atomic_int_add (&mem_pool_stat->bytes_allocated, -tmp->len); g_slice_free1 (tmp->len, tmp->begin); g_slice_free (struct _pool_chain, tmp); } @@ -671,10 +649,8 @@ rspamd_mempool_delete (rspamd_mempool_t * pool) while (cur_shared) { tmp_shared = cur_shared; cur_shared = cur_shared->next; - STAT_LOCK (); - mem_pool_stat->chunks_freed++; - mem_pool_stat->bytes_allocated -= tmp_shared->len; - STAT_UNLOCK (); + g_atomic_int_inc (&mem_pool_stat->chunks_freed); + g_atomic_int_add (&mem_pool_stat->bytes_allocated, -tmp_shared->len); munmap ((void *)tmp_shared, tmp_shared->len + sizeof (struct _pool_chain_shared)); } @@ -682,7 +658,7 @@ rspamd_mempool_delete (rspamd_mempool_t * pool) g_hash_table_destroy (pool->variables); } - mem_pool_stat->pools_freed++; + g_atomic_int_inc (&mem_pool_stat->pools_freed); POOL_MTX_UNLOCK (); rspamd_mutex_free (pool->mtx); g_slice_free (rspamd_mempool_t, pool); @@ -698,14 +674,12 @@ rspamd_mempool_cleanup_tmp (rspamd_mempool_t * pool) while (cur) { tmp = cur; cur = cur->next; - STAT_LOCK (); - mem_pool_stat->chunks_freed++; - mem_pool_stat->bytes_allocated -= tmp->len; - STAT_UNLOCK (); + g_atomic_int_inc (&mem_pool_stat->chunks_freed); + g_atomic_int_add (&mem_pool_stat->bytes_allocated, -tmp->len); g_slice_free1 (tmp->len, tmp->begin); g_slice_free (struct _pool_chain, tmp); } - mem_pool_stat->pools_freed++; + g_atomic_int_inc (&mem_pool_stat->pools_freed); POOL_MTX_UNLOCK (); } diff --git a/src/libutil/mem_pool.h b/src/libutil/mem_pool.h index 2a8546ef0..6d59e27c7 100644 --- a/src/libutil/mem_pool.h +++ b/src/libutil/mem_pool.h @@ -86,13 +86,13 @@ typedef struct memory_pool_s { * Statistics structure */ typedef struct memory_pool_stat_s { - gsize pools_allocated; /**< total number of allocated pools */ - gsize pools_freed; /**< number of freed pools */ - gsize bytes_allocated; /**< bytes that are allocated with pool allocator */ - gsize chunks_allocated; /**< number of chunks that are allocated */ - gsize shared_chunks_allocated; /**< shared chunks allocated */ - gsize chunks_freed; /**< chunks freed */ - gsize oversized_chunks; /**< oversized chunks */ + guint pools_allocated; /**< total number of allocated pools */ + guint pools_freed; /**< number of freed pools */ + guint bytes_allocated; /**< bytes that are allocated with pool allocator */ + guint chunks_allocated; /**< number of chunks that are allocated */ + guint shared_chunks_allocated; /**< shared chunks allocated */ + guint chunks_freed; /**< chunks freed */ + guint oversized_chunks; /**< oversized chunks */ } rspamd_mempool_stat_t; /** |