From cf75aea218579006262409fd50a239d7b56ce897 Mon Sep 17 00:00:00 2001 From: Campbell Barton Date: Sat, 2 Mar 2019 18:35:20 +1100 Subject: Cleanup: code style & comments --- source/blender/blenlib/intern/BLI_mempool.c | 102 +++++++++++++++++----------- 1 file changed, 62 insertions(+), 40 deletions(-) (limited to 'source/blender/blenlib') diff --git a/source/blender/blenlib/intern/BLI_mempool.c b/source/blender/blenlib/intern/BLI_mempool.c index cfb96897381..a96d4c8b5dd 100644 --- a/source/blender/blenlib/intern/BLI_mempool.c +++ b/source/blender/blenlib/intern/BLI_mempool.c @@ -94,7 +94,8 @@ static bool mempool_debug_memset = false; */ typedef struct BLI_freenode { struct BLI_freenode *next; - intptr_t freeword; /* used to identify this as a freed node */ + /** Used to identify this as a freed node. */ + intptr_t freeword; } BLI_freenode; /** @@ -109,22 +110,30 @@ typedef struct BLI_mempool_chunk { * The mempool, stores and tracks memory \a chunks and elements within those chunks \a free. */ struct BLI_mempool { - BLI_mempool_chunk *chunks; /* single linked list of allocated chunks */ - /* keep a pointer to the last, so we can append new chunks there - * this is needed for iteration so we can loop over chunks in the order added */ + /** Single linked list of allocated chunks. */ + BLI_mempool_chunk *chunks; + /** Keep a pointer to the last, so we can append new chunks there + * this is needed for iteration so we can loop over chunks in the order added. */ BLI_mempool_chunk *chunk_tail; - uint esize; /* element size in bytes */ - uint csize; /* chunk size in bytes */ - uint pchunk; /* number of elements per chunk */ + /** Element size in bytes. */ + uint esize; + /** Chunk size in bytes. */ + uint csize; + /** Number of elements per chunk. */ + uint pchunk; uint flag; /* keeps aligned to 16 bits */ - BLI_freenode *free; /* free element list. Interleaved into chunk datas. */ - uint maxchunks; /* use to know how many chunks to keep for BLI_mempool_clear */ - uint totused; /* number of elements currently in use */ + /** Free element list. Interleaved into chunk datas. */ + BLI_freenode *free; + /** Use to know how many chunks to keep for #BLI_mempool_clear. */ + uint maxchunks; + /** Number of elements currently in use. */ + uint totused; #ifdef USE_TOTALLOC - uint totalloc; /* number of elements allocated in total */ + /** Number of elements allocated in total. */ + uint totalloc; #endif }; @@ -135,7 +144,7 @@ struct BLI_mempool { #define NODE_STEP_NEXT(node) ((void *)((char *)(node) + esize)) #define NODE_STEP_PREV(node) ((void *)((char *)(node) - esize)) -/* extra bytes implicitly used for every chunk alloc */ +/** Extra bytes implicitly used for every chunk alloc. */ #define CHUNK_OVERHEAD (uint)(MEM_SIZE_OVERHEAD) #ifdef USE_CHUNK_POW2 @@ -180,12 +189,13 @@ static BLI_mempool_chunk *mempool_chunk_alloc(BLI_mempool *pool) * * \param pool: The pool to add the chunk into. * \param mpchunk: The new uninitialized chunk (can be malloc'd) - * \param lasttail: The last element of the previous chunk + * \param last_tail: The last element of the previous chunk * (used when building free chunks initially) * \return The last chunk, */ -static BLI_freenode *mempool_chunk_add(BLI_mempool *pool, BLI_mempool_chunk *mpchunk, - BLI_freenode *lasttail) +static BLI_freenode *mempool_chunk_add( + BLI_mempool *pool, BLI_mempool_chunk *mpchunk, + BLI_freenode *last_tail) { const uint esize = pool->esize; BLI_freenode *curnode = CHUNK_DATA(mpchunk); @@ -224,7 +234,7 @@ static BLI_freenode *mempool_chunk_add(BLI_mempool *pool, BLI_mempool_chunk *mpc } /* terminate the list (rewind one) - * will be overwritten if 'curnode' gets passed in again as 'lasttail' */ + * will be overwritten if 'curnode' gets passed in again as 'last_tail' */ curnode = NODE_STEP_PREV(curnode); curnode->next = NULL; @@ -233,8 +243,8 @@ static BLI_freenode *mempool_chunk_add(BLI_mempool *pool, BLI_mempool_chunk *mpc #endif /* final pointer in the previously allocated chunk is wrong */ - if (lasttail) { - lasttail->next = CHUNK_DATA(mpchunk); + if (last_tail) { + last_tail->next = CHUNK_DATA(mpchunk); } return curnode; @@ -255,11 +265,12 @@ static void mempool_chunk_free_all(BLI_mempool_chunk *mpchunk) } } -BLI_mempool *BLI_mempool_create(uint esize, uint totelem, - uint pchunk, uint flag) +BLI_mempool *BLI_mempool_create( + uint esize, uint totelem, + uint pchunk, uint flag) { BLI_mempool *pool; - BLI_freenode *lasttail = NULL; + BLI_freenode *last_tail = NULL; uint i, maxchunks; /* allocate the pool structure */ @@ -282,7 +293,7 @@ BLI_mempool *BLI_mempool_create(uint esize, uint totelem, pool->csize = esize * pchunk; - /* Optimize chunk size to powers of 2, accounting for slop-space */ + /* Optimize chunk size to powers of 2, accounting for slop-space. */ #ifdef USE_CHUNK_POW2 { BLI_assert(pool->csize > CHUNK_OVERHEAD); @@ -302,10 +313,10 @@ BLI_mempool *BLI_mempool_create(uint esize, uint totelem, pool->totused = 0; if (totelem) { - /* allocate the actual chunks */ + /* Allocate the actual chunks. */ for (i = 0; i < maxchunks; i++) { BLI_mempool_chunk *mpchunk = mempool_chunk_alloc(pool); - lasttail = mempool_chunk_add(pool, mpchunk, lasttail); + last_tail = mempool_chunk_add(pool, mpchunk, last_tail); } } @@ -321,7 +332,7 @@ void *BLI_mempool_alloc(BLI_mempool *pool) BLI_freenode *free_pop; if (UNLIKELY(pool->free == NULL)) { - /* need to allocate a new chunk */ + /* Need to allocate a new chunk. */ BLI_mempool_chunk *mpchunk = mempool_chunk_alloc(pool); mempool_chunk_add(pool, mpchunk, NULL); } @@ -375,7 +386,7 @@ void BLI_mempool_free(BLI_mempool *pool, void *addr) } } - /* enable for debugging */ + /* Enable for debugging. */ if (UNLIKELY(mempool_debug_memset)) { memset(addr, 255, pool->esize); } @@ -383,7 +394,7 @@ void BLI_mempool_free(BLI_mempool *pool, void *addr) if (pool->flag & BLI_MEMPOOL_ALLOW_ITER) { #ifndef NDEBUG - /* this will detect double free's */ + /* This will detect double free's. */ BLI_assert(newhead->freeword != FREEWORD); #endif newhead->freeword = FREEWORD; @@ -398,7 +409,7 @@ void BLI_mempool_free(BLI_mempool *pool, void *addr) VALGRIND_MEMPOOL_FREE(pool, addr); #endif - /* nothing is in use; free all the chunks except the first */ + /* Nothing is in use; free all the chunks except the first. */ if (UNLIKELY(pool->totused == 0) && (pool->chunks->next)) { @@ -416,7 +427,7 @@ void BLI_mempool_free(BLI_mempool *pool, void *addr) pool->totalloc = pool->pchunk; #endif - /* temp alloc so valgrind doesn't complain when setting free'd blocks 'next' */ + /* Temp alloc so valgrind doesn't complain when setting free'd blocks 'next'. */ #ifdef WITH_MEM_VALGRIND VALGRIND_MEMPOOL_ALLOC(pool, CHUNK_DATA(first), pool->csize); #endif @@ -448,12 +459,12 @@ void *BLI_mempool_findelem(BLI_mempool *pool, uint index) BLI_assert(pool->flag & BLI_MEMPOOL_ALLOW_ITER); if (index < pool->totused) { - /* we could have some faster mem chunk stepping code inline */ + /* We could have some faster mem chunk stepping code inline. */ BLI_mempool_iter iter; void *elem; BLI_mempool_iternew(pool, &iter); for (elem = BLI_mempool_iterstep(&iter); index-- != 0; elem = BLI_mempool_iterstep(&iter)) { - /* do nothing */ + /* pass */ } return elem; } @@ -519,7 +530,7 @@ void *BLI_mempool_as_arrayN(BLI_mempool *pool, const char *allocstr) } /** - * Initialize a new mempool iterator, \a BLI_MEMPOOL_ALLOW_ITER flag must be set. + * Initialize a new mempool iterator, #BLI_MEMPOOL_ALLOW_ITER flag must be set. */ void BLI_mempool_iternew(BLI_mempool *pool, BLI_mempool_iter *iter) { @@ -533,7 +544,7 @@ void BLI_mempool_iternew(BLI_mempool *pool, BLI_mempool_iter *iter) } /** - * Initialize an array of mempool iterators, \a BLI_MEMPOOL_ALLOW_ITER flag must be set. + * Initialize an array of mempool iterators, #BLI_MEMPOOL_ALLOW_ITER flag must be set. * * This is used in threaded code, to generate as much iterators as needed (each task should have its own), * such that each iterator goes over its own single chunk, and only getting the next chunk to iterate over has to be @@ -557,7 +568,8 @@ BLI_mempool_iter *BLI_mempool_iter_threadsafe_create(BLI_mempool *pool, const si for (size_t i = 1; i < num_iter; i++) { iter_arr[i] = iter_arr[0]; - *curchunk_threaded_shared = iter_arr[i].curchunk = (*curchunk_threaded_shared) ? (*curchunk_threaded_shared)->next : NULL; + *curchunk_threaded_shared = iter_arr[i].curchunk = ( + (*curchunk_threaded_shared) ? (*curchunk_threaded_shared)->next : NULL); } return iter_arr; @@ -594,7 +606,11 @@ static void *bli_mempool_iternext(BLI_mempool_iter *iter) if (iter->curchunk == NULL) { return ret; } - if (atomic_cas_ptr((void **)iter->curchunk_threaded_shared, iter->curchunk, iter->curchunk->next) == iter->curchunk) { + if (atomic_cas_ptr( + (void **)iter->curchunk_threaded_shared, + iter->curchunk, + iter->curchunk->next) == iter->curchunk) + { break; } } @@ -643,8 +659,14 @@ void *BLI_mempool_iterstep(BLI_mempool_iter *iter) if (iter->curchunk_threaded_shared) { for (iter->curchunk = *iter->curchunk_threaded_shared; (iter->curchunk != NULL) && - (atomic_cas_ptr((void **)iter->curchunk_threaded_shared, iter->curchunk, iter->curchunk->next) != iter->curchunk); - iter->curchunk = *iter->curchunk_threaded_shared); + (atomic_cas_ptr( + (void **)iter->curchunk_threaded_shared, + iter->curchunk, + iter->curchunk->next) != iter->curchunk); + iter->curchunk = *iter->curchunk_threaded_shared) + { + /* pass. */ + } if (UNLIKELY(iter->curchunk == NULL)) { return (ret->freeword == FREEWORD) ? NULL : ret; @@ -676,7 +698,7 @@ void BLI_mempool_clear_ex(BLI_mempool *pool, const int totelem_reserve) uint maxchunks; BLI_mempool_chunk *chunks_temp; - BLI_freenode *lasttail = NULL; + BLI_freenode *last_tail = NULL; #ifdef WITH_MEM_VALGRIND VALGRIND_DESTROY_MEMPOOL(pool); @@ -690,7 +712,7 @@ void BLI_mempool_clear_ex(BLI_mempool *pool, const int totelem_reserve) maxchunks = mempool_maxchunks((uint)totelem_reserve, pool->pchunk); } - /* free all after pool->maxchunks */ + /* Free all after 'pool->maxchunks'. */ mpchunk = mempool_chunk_find(pool->chunks, maxchunks - 1); if (mpchunk && mpchunk->next) { /* terminate */ @@ -717,7 +739,7 @@ void BLI_mempool_clear_ex(BLI_mempool *pool, const int totelem_reserve) while ((mpchunk = chunks_temp)) { chunks_temp = mpchunk->next; - lasttail = mempool_chunk_add(pool, mpchunk, lasttail); + last_tail = mempool_chunk_add(pool, mpchunk, last_tail); } } -- cgit v1.2.3