[Bf-blender-cvs] [cf75aea2185] master: Cleanup: code style & comments

Campbell Barton noreply at git.blender.org
Sat Mar 2 08:38:35 CET 2019


Commit: cf75aea218579006262409fd50a239d7b56ce897
Author: Campbell Barton
Date:   Sat Mar 2 18:35:20 2019 +1100
Branches: master
https://developer.blender.org/rBcf75aea218579006262409fd50a239d7b56ce897

Cleanup: code style & comments

===================================================================

M	source/blender/blenlib/intern/BLI_mempool.c

===================================================================

diff --git a/source/blender/blenlib/intern/BLI_mempool.c b/source/blender/blenlib/intern/BLI_mempool.c
index cfb96897381..a96d4c8b5dd 100644
--- a/source/blender/blenlib/intern/BLI_mempool.c
+++ b/source/blender/blenlib/intern/BLI_mempool.c
@@ -94,7 +94,8 @@ static bool mempool_debug_memset = false;
  */
 typedef struct BLI_freenode {
 	struct BLI_freenode *next;
-	intptr_t freeword; /* used to identify this as a freed node */
+	/** Used to identify this as a freed node. */
+	intptr_t freeword;
 } BLI_freenode;
 
 /**
@@ -109,22 +110,30 @@ typedef struct BLI_mempool_chunk {
  * The mempool, stores and tracks memory \a chunks and elements within those chunks \a free.
  */
 struct BLI_mempool {
-	BLI_mempool_chunk *chunks;  /* single linked list of allocated chunks */
-	/* keep a pointer to the last, so we can append new chunks there
-	 * this is needed for iteration so we can loop over chunks in the order added */
+	/** Single linked list of allocated chunks. */
+	BLI_mempool_chunk *chunks;
+	/** Keep a pointer to the last, so we can append new chunks there
+	 * this is needed for iteration so we can loop over chunks in the order added. */
 	BLI_mempool_chunk *chunk_tail;
 
-	uint esize;         /* element size in bytes */
-	uint csize;         /* chunk size in bytes */
-	uint pchunk;        /* number of elements per chunk */
+	/** Element size in bytes. */
+	uint esize;
+	/** Chunk size in bytes. */
+	uint csize;
+	/** Number of elements per chunk. */
+	uint pchunk;
 	uint flag;
 	/* keeps aligned to 16 bits */
 
-	BLI_freenode *free;         /* free element list. Interleaved into chunk datas. */
-	uint maxchunks;     /* use to know how many chunks to keep for BLI_mempool_clear */
-	uint totused;       /* number of elements currently in use */
+	/** Free element list. Interleaved into chunk datas. */
+	BLI_freenode *free;
+	/** Use to know how many chunks to keep for #BLI_mempool_clear. */
+	uint maxchunks;
+	/** Number of elements currently in use. */
+	uint totused;
 #ifdef USE_TOTALLOC
-	uint totalloc;          /* number of elements allocated in total */
+	/** Number of elements allocated in total. */
+	uint totalloc;
 #endif
 };
 
@@ -135,7 +144,7 @@ struct BLI_mempool {
 #define NODE_STEP_NEXT(node)  ((void *)((char *)(node) + esize))
 #define NODE_STEP_PREV(node)  ((void *)((char *)(node) - esize))
 
-/* extra bytes implicitly used for every chunk alloc */
+/** Extra bytes implicitly used for every chunk alloc. */
 #define CHUNK_OVERHEAD (uint)(MEM_SIZE_OVERHEAD)
 
 #ifdef USE_CHUNK_POW2
@@ -180,12 +189,13 @@ static BLI_mempool_chunk *mempool_chunk_alloc(BLI_mempool *pool)
  *
  * \param pool: The pool to add the chunk into.
  * \param mpchunk: The new uninitialized chunk (can be malloc'd)
- * \param lasttail: The last element of the previous chunk
+ * \param last_tail: The last element of the previous chunk
  * (used when building free chunks initially)
  * \return The last chunk,
  */
-static BLI_freenode *mempool_chunk_add(BLI_mempool *pool, BLI_mempool_chunk *mpchunk,
-                                       BLI_freenode *lasttail)
+static BLI_freenode *mempool_chunk_add(
+        BLI_mempool *pool, BLI_mempool_chunk *mpchunk,
+        BLI_freenode *last_tail)
 {
 	const uint esize = pool->esize;
 	BLI_freenode *curnode = CHUNK_DATA(mpchunk);
@@ -224,7 +234,7 @@ static BLI_freenode *mempool_chunk_add(BLI_mempool *pool, BLI_mempool_chunk *mpc
 	}
 
 	/* terminate the list (rewind one)
-	 * will be overwritten if 'curnode' gets passed in again as 'lasttail' */
+	 * will be overwritten if 'curnode' gets passed in again as 'last_tail' */
 	curnode = NODE_STEP_PREV(curnode);
 	curnode->next = NULL;
 
@@ -233,8 +243,8 @@ static BLI_freenode *mempool_chunk_add(BLI_mempool *pool, BLI_mempool_chunk *mpc
 #endif
 
 	/* final pointer in the previously allocated chunk is wrong */
-	if (lasttail) {
-		lasttail->next = CHUNK_DATA(mpchunk);
+	if (last_tail) {
+		last_tail->next = CHUNK_DATA(mpchunk);
 	}
 
 	return curnode;
@@ -255,11 +265,12 @@ static void mempool_chunk_free_all(BLI_mempool_chunk *mpchunk)
 	}
 }
 
-BLI_mempool *BLI_mempool_create(uint esize, uint totelem,
-                                uint pchunk, uint flag)
+BLI_mempool *BLI_mempool_create(
+        uint esize, uint totelem,
+        uint pchunk, uint flag)
 {
 	BLI_mempool *pool;
-	BLI_freenode *lasttail = NULL;
+	BLI_freenode *last_tail = NULL;
 	uint i, maxchunks;
 
 	/* allocate the pool structure */
@@ -282,7 +293,7 @@ BLI_mempool *BLI_mempool_create(uint esize, uint totelem,
 	pool->csize = esize * pchunk;
 
 
-	/* Optimize chunk size to powers of 2, accounting for slop-space */
+	/* Optimize chunk size to powers of 2, accounting for slop-space. */
 #ifdef USE_CHUNK_POW2
 	{
 		BLI_assert(pool->csize > CHUNK_OVERHEAD);
@@ -302,10 +313,10 @@ BLI_mempool *BLI_mempool_create(uint esize, uint totelem,
 	pool->totused = 0;
 
 	if (totelem) {
-		/* allocate the actual chunks */
+		/* Allocate the actual chunks. */
 		for (i = 0; i < maxchunks; i++) {
 			BLI_mempool_chunk *mpchunk = mempool_chunk_alloc(pool);
-			lasttail = mempool_chunk_add(pool, mpchunk, lasttail);
+			last_tail = mempool_chunk_add(pool, mpchunk, last_tail);
 		}
 	}
 
@@ -321,7 +332,7 @@ void *BLI_mempool_alloc(BLI_mempool *pool)
 	BLI_freenode *free_pop;
 
 	if (UNLIKELY(pool->free == NULL)) {
-		/* need to allocate a new chunk */
+		/* Need to allocate a new chunk. */
 		BLI_mempool_chunk *mpchunk = mempool_chunk_alloc(pool);
 		mempool_chunk_add(pool, mpchunk, NULL);
 	}
@@ -375,7 +386,7 @@ void BLI_mempool_free(BLI_mempool *pool, void *addr)
 		}
 	}
 
-	/* enable for debugging */
+	/* Enable for debugging. */
 	if (UNLIKELY(mempool_debug_memset)) {
 		memset(addr, 255, pool->esize);
 	}
@@ -383,7 +394,7 @@ void BLI_mempool_free(BLI_mempool *pool, void *addr)
 
 	if (pool->flag & BLI_MEMPOOL_ALLOW_ITER) {
 #ifndef NDEBUG
-		/* this will detect double free's */
+		/* This will detect double free's. */
 		BLI_assert(newhead->freeword != FREEWORD);
 #endif
 		newhead->freeword = FREEWORD;
@@ -398,7 +409,7 @@ void BLI_mempool_free(BLI_mempool *pool, void *addr)
 	VALGRIND_MEMPOOL_FREE(pool, addr);
 #endif
 
-	/* nothing is in use; free all the chunks except the first */
+	/* Nothing is in use; free all the chunks except the first. */
 	if (UNLIKELY(pool->totused == 0) &&
 	    (pool->chunks->next))
 	{
@@ -416,7 +427,7 @@ void BLI_mempool_free(BLI_mempool *pool, void *addr)
 		pool->totalloc = pool->pchunk;
 #endif
 
-		/* temp alloc so valgrind doesn't complain when setting free'd blocks 'next' */
+		/* Temp alloc so valgrind doesn't complain when setting free'd blocks 'next'. */
 #ifdef WITH_MEM_VALGRIND
 		VALGRIND_MEMPOOL_ALLOC(pool, CHUNK_DATA(first), pool->csize);
 #endif
@@ -448,12 +459,12 @@ void *BLI_mempool_findelem(BLI_mempool *pool, uint index)
 	BLI_assert(pool->flag & BLI_MEMPOOL_ALLOW_ITER);
 
 	if (index < pool->totused) {
-		/* we could have some faster mem chunk stepping code inline */
+		/* We could have some faster mem chunk stepping code inline. */
 		BLI_mempool_iter iter;
 		void *elem;
 		BLI_mempool_iternew(pool, &iter);
 		for (elem = BLI_mempool_iterstep(&iter); index-- != 0; elem = BLI_mempool_iterstep(&iter)) {
-			/* do nothing */
+			/* pass */
 		}
 		return elem;
 	}
@@ -519,7 +530,7 @@ void *BLI_mempool_as_arrayN(BLI_mempool *pool, const char *allocstr)
 }
 
 /**
- * Initialize a new mempool iterator, \a BLI_MEMPOOL_ALLOW_ITER flag must be set.
+ * Initialize a new mempool iterator, #BLI_MEMPOOL_ALLOW_ITER flag must be set.
  */
 void BLI_mempool_iternew(BLI_mempool *pool, BLI_mempool_iter *iter)
 {
@@ -533,7 +544,7 @@ void BLI_mempool_iternew(BLI_mempool *pool, BLI_mempool_iter *iter)
 }
 
 /**
- * Initialize an array of mempool iterators, \a BLI_MEMPOOL_ALLOW_ITER flag must be set.
+ * Initialize an array of mempool iterators, #BLI_MEMPOOL_ALLOW_ITER flag must be set.
  *
  * This is used in threaded code, to generate as much iterators as needed (each task should have its own),
  * such that each iterator goes over its own single chunk, and only getting the next chunk to iterate over has to be
@@ -557,7 +568,8 @@ BLI_mempool_iter *BLI_mempool_iter_threadsafe_create(BLI_mempool *pool, const si
 
 	for (size_t i = 1; i < num_iter; i++) {
 		iter_arr[i] = iter_arr[0];
-		*curchunk_threaded_shared = iter_arr[i].curchunk = (*curchunk_threaded_shared) ? (*curchunk_threaded_shared)->next : NULL;
+		*curchunk_threaded_shared = iter_arr[i].curchunk = (
+		        (*curchunk_threaded_shared) ? (*curchunk_threaded_shared)->next : NULL);
 	}
 
 	return iter_arr;
@@ -594,7 +606,11 @@ static void *bli_mempool_iternext(BLI_mempool_iter *iter)
 				if (iter->curchunk == NULL) {
 					return ret;
 				}
-				if (atomic_cas_ptr((void **)iter->curchunk_threaded_shared, iter->curchunk, iter->curchunk->next) == iter->curchunk) {
+				if (atomic_cas_ptr(
+				            (void **)iter->curchunk_threaded_shared,
+				            iter->curchunk,
+				            iter->curchunk->next) == iter->curchunk)
+				{
 					break;
 				}
 			}
@@ -643,8 +659,14 @@ void *BLI_mempool_iterstep(BLI_mempool_iter *iter)
 			if (iter->curchunk_threaded_shared) {
 				for (iter->curchunk = *iter->curchunk_threaded_shared;
 				     (iter->curchunk != NULL) &&
-				     (atomic_cas_ptr((void **)iter->curchunk_threaded_shared, iter->curchunk, iter->curchunk->next) != iter->curchunk);
-				     iter->curchunk = *iter->curchunk_threaded_shared);
+				     (atomic_cas_ptr(
+				             (void **)iter->curchunk_threaded_shared,
+				             iter->curchunk,
+				             iter->curchunk->next) != iter->curchunk);
+				     iter->curchunk = *iter->curchunk_threaded_shared)
+				{
+					/* pass. */
+				}
 
 				if (UNLIKELY(iter->curchunk == NULL)) {
 					return (ret->freeword == FREEWORD) ? NULL : ret;
@@ -676,7 +698,7 @@ void BLI_mempool_clear_ex(BLI_mempool *pool, const int totelem_reserve)
 	uint maxchunks;
 
 	BLI_mempool_chunk *chunks_temp;
-	BLI_freenode *lasttail = NULL;
+	BLI_freenode *last_tail = NULL;
 
 #ifdef WITH_MEM_VALGRIND
 	VALGRIND_DESTROY_MEMPOOL(pool);
@@ -690,7 +712,7 @@ void BLI_mempool_clear_ex(BLI_mempool *pool, const int totelem_reserve)
 		maxchunks = mempool_maxchunks((uint)totele

@@ Diff output truncated at 10240 characters. @@



More information about the Bf-blender-cvs mailing list