From e12c08e8d170b7ca40f204a5b0423c23a9fbc2c1 Mon Sep 17 00:00:00 2001 From: Campbell Barton Date: Wed, 17 Apr 2019 06:17:24 +0200 Subject: ClangFormat: apply to source, most of intern Apply clang format as proposed in T53211. For details on usage and instructions for migrating branches without conflicts, see: https://wiki.blender.org/wiki/Tools/ClangFormat --- .../blender/draw/intern/draw_cache_impl_metaball.c | 304 ++++++++++----------- 1 file changed, 152 insertions(+), 152 deletions(-) (limited to 'source/blender/draw/intern/draw_cache_impl_metaball.c') diff --git a/source/blender/draw/intern/draw_cache_impl_metaball.c b/source/blender/draw/intern/draw_cache_impl_metaball.c index 46c247d67ea..2c06d536f3d 100644 --- a/source/blender/draw/intern/draw_cache_impl_metaball.c +++ b/source/blender/draw/intern/draw_cache_impl_metaball.c @@ -35,9 +35,7 @@ #include "GPU_batch.h" - -#include "draw_cache_impl.h" /* own include */ - +#include "draw_cache_impl.h" /* own include */ static void metaball_batch_cache_clear(MetaBall *mb); @@ -45,127 +43,128 @@ static void metaball_batch_cache_clear(MetaBall *mb); /* MetaBall GPUBatch Cache */ typedef struct MetaBallBatchCache { - GPUBatch *batch; - GPUBatch **shaded_triangles; + GPUBatch *batch; + GPUBatch **shaded_triangles; - int mat_len; + int mat_len; - /* Shared */ - GPUVertBuf *pos_nor_in_order; + /* Shared */ + GPUVertBuf *pos_nor_in_order; - /* Wireframe */ - struct { - GPUBatch *batch; - } face_wire; + /* Wireframe */ + struct { + GPUBatch *batch; + } face_wire; - /* Edge detection */ - GPUBatch *edge_detection; - GPUIndexBuf *edges_adj_lines; + /* Edge detection */ + GPUBatch *edge_detection; + GPUIndexBuf *edges_adj_lines; - /* settings to determine if cache is invalid */ - bool is_dirty; + /* settings to determine if cache is invalid */ + bool is_dirty; - /* Valid only if edge_detection is up to date. */ - bool is_manifold; + /* Valid only if edge_detection is up to date. */ + bool is_manifold; } MetaBallBatchCache; /* GPUBatch cache management. */ static bool metaball_batch_cache_valid(MetaBall *mb) { - MetaBallBatchCache *cache = mb->batch_cache; + MetaBallBatchCache *cache = mb->batch_cache; - if (cache == NULL) { - return false; - } + if (cache == NULL) { + return false; + } - return cache->is_dirty == false; + return cache->is_dirty == false; } static void metaball_batch_cache_init(MetaBall *mb) { - MetaBallBatchCache *cache = mb->batch_cache; - - if (!cache) { - cache = mb->batch_cache = MEM_mallocN(sizeof(*cache), __func__); - } - cache->batch = NULL; - cache->mat_len = 0; - cache->shaded_triangles = NULL; - cache->is_dirty = false; - cache->pos_nor_in_order = NULL; - cache->face_wire.batch = NULL; - cache->edge_detection = NULL; - cache->edges_adj_lines = NULL; - cache->is_manifold = false; + MetaBallBatchCache *cache = mb->batch_cache; + + if (!cache) { + cache = mb->batch_cache = MEM_mallocN(sizeof(*cache), __func__); + } + cache->batch = NULL; + cache->mat_len = 0; + cache->shaded_triangles = NULL; + cache->is_dirty = false; + cache->pos_nor_in_order = NULL; + cache->face_wire.batch = NULL; + cache->edge_detection = NULL; + cache->edges_adj_lines = NULL; + cache->is_manifold = false; } static MetaBallBatchCache *metaball_batch_cache_get(MetaBall *mb) { - if (!metaball_batch_cache_valid(mb)) { - metaball_batch_cache_clear(mb); - metaball_batch_cache_init(mb); - } - return mb->batch_cache; + if (!metaball_batch_cache_valid(mb)) { + metaball_batch_cache_clear(mb); + metaball_batch_cache_init(mb); + } + return mb->batch_cache; } void DRW_mball_batch_cache_dirty_tag(MetaBall *mb, int mode) { - MetaBallBatchCache *cache = mb->batch_cache; - if (cache == NULL) { - return; - } - switch (mode) { - case BKE_MBALL_BATCH_DIRTY_ALL: - cache->is_dirty = true; - break; - default: - BLI_assert(0); - } + MetaBallBatchCache *cache = mb->batch_cache; + if (cache == NULL) { + return; + } + switch (mode) { + case BKE_MBALL_BATCH_DIRTY_ALL: + cache->is_dirty = true; + break; + default: + BLI_assert(0); + } } static void metaball_batch_cache_clear(MetaBall *mb) { - MetaBallBatchCache *cache = mb->batch_cache; - if (!cache) { - return; - } - - GPU_BATCH_DISCARD_SAFE(cache->face_wire.batch); - GPU_BATCH_DISCARD_SAFE(cache->batch); - GPU_BATCH_DISCARD_SAFE(cache->edge_detection); - GPU_VERTBUF_DISCARD_SAFE(cache->pos_nor_in_order); - GPU_INDEXBUF_DISCARD_SAFE(cache->edges_adj_lines); - /* Note: shaded_triangles[0] is already freed by cache->batch */ - MEM_SAFE_FREE(cache->shaded_triangles); - cache->mat_len = 0; - cache->is_manifold = false; + MetaBallBatchCache *cache = mb->batch_cache; + if (!cache) { + return; + } + + GPU_BATCH_DISCARD_SAFE(cache->face_wire.batch); + GPU_BATCH_DISCARD_SAFE(cache->batch); + GPU_BATCH_DISCARD_SAFE(cache->edge_detection); + GPU_VERTBUF_DISCARD_SAFE(cache->pos_nor_in_order); + GPU_INDEXBUF_DISCARD_SAFE(cache->edges_adj_lines); + /* Note: shaded_triangles[0] is already freed by cache->batch */ + MEM_SAFE_FREE(cache->shaded_triangles); + cache->mat_len = 0; + cache->is_manifold = false; } void DRW_mball_batch_cache_free(MetaBall *mb) { - metaball_batch_cache_clear(mb); - MEM_SAFE_FREE(mb->batch_cache); + metaball_batch_cache_clear(mb); + MEM_SAFE_FREE(mb->batch_cache); } static GPUVertBuf *mball_batch_cache_get_pos_and_normals(Object *ob, MetaBallBatchCache *cache) { - if (cache->pos_nor_in_order == NULL) { - ListBase *lb = &ob->runtime.curve_cache->disp; - cache->pos_nor_in_order = MEM_callocN(sizeof(GPUVertBuf), __func__); - DRW_displist_vertbuf_create_pos_and_nor(lb, cache->pos_nor_in_order); - } - return cache->pos_nor_in_order; + if (cache->pos_nor_in_order == NULL) { + ListBase *lb = &ob->runtime.curve_cache->disp; + cache->pos_nor_in_order = MEM_callocN(sizeof(GPUVertBuf), __func__); + DRW_displist_vertbuf_create_pos_and_nor(lb, cache->pos_nor_in_order); + } + return cache->pos_nor_in_order; } static GPUIndexBuf *mball_batch_cache_get_edges_adj_lines(Object *ob, MetaBallBatchCache *cache) { - if (cache->edges_adj_lines == NULL) { - ListBase *lb = &ob->runtime.curve_cache->disp; - cache->edges_adj_lines = MEM_callocN(sizeof(GPUVertBuf), __func__); - DRW_displist_indexbuf_create_edges_adjacency_lines(lb, cache->edges_adj_lines, &cache->is_manifold); - } - return cache->edges_adj_lines; + if (cache->edges_adj_lines == NULL) { + ListBase *lb = &ob->runtime.curve_cache->disp; + cache->edges_adj_lines = MEM_callocN(sizeof(GPUVertBuf), __func__); + DRW_displist_indexbuf_create_edges_adjacency_lines( + lb, cache->edges_adj_lines, &cache->is_manifold); + } + return cache->edges_adj_lines; } /* -------------------------------------------------------------------- */ @@ -174,95 +173,96 @@ static GPUIndexBuf *mball_batch_cache_get_edges_adj_lines(Object *ob, MetaBallBa GPUBatch *DRW_metaball_batch_cache_get_triangles_with_normals(Object *ob) { - if (!BKE_mball_is_basis(ob)) { - return NULL; - } - - MetaBall *mb = ob->data; - MetaBallBatchCache *cache = metaball_batch_cache_get(mb); - - if (cache->batch == NULL) { - ListBase *lb = &ob->runtime.curve_cache->disp; - GPUIndexBuf *ibo = MEM_callocN(sizeof(GPUIndexBuf), __func__); - DRW_displist_indexbuf_create_triangles_in_order(lb, ibo); - cache->batch = GPU_batch_create_ex( - GPU_PRIM_TRIS, - mball_batch_cache_get_pos_and_normals(ob, cache), - ibo, - GPU_BATCH_OWNS_INDEX); - } - - return cache->batch; + if (!BKE_mball_is_basis(ob)) { + return NULL; + } + + MetaBall *mb = ob->data; + MetaBallBatchCache *cache = metaball_batch_cache_get(mb); + + if (cache->batch == NULL) { + ListBase *lb = &ob->runtime.curve_cache->disp; + GPUIndexBuf *ibo = MEM_callocN(sizeof(GPUIndexBuf), __func__); + DRW_displist_indexbuf_create_triangles_in_order(lb, ibo); + cache->batch = GPU_batch_create_ex(GPU_PRIM_TRIS, + mball_batch_cache_get_pos_and_normals(ob, cache), + ibo, + GPU_BATCH_OWNS_INDEX); + } + + return cache->batch; } -GPUBatch **DRW_metaball_batch_cache_get_surface_shaded(Object *ob, MetaBall *mb, struct GPUMaterial **UNUSED(gpumat_array), uint gpumat_array_len) +GPUBatch **DRW_metaball_batch_cache_get_surface_shaded(Object *ob, + MetaBall *mb, + struct GPUMaterial **UNUSED(gpumat_array), + uint gpumat_array_len) { - if (!BKE_mball_is_basis(ob)) { - return NULL; - } - - MetaBallBatchCache *cache = metaball_batch_cache_get(mb); - if (cache->shaded_triangles == NULL) { - cache->mat_len = gpumat_array_len; - cache->shaded_triangles = MEM_callocN(sizeof(*cache->shaded_triangles) * cache->mat_len, __func__); - cache->shaded_triangles[0] = DRW_metaball_batch_cache_get_triangles_with_normals(ob); - for (int i = 1; i < cache->mat_len; ++i) { - cache->shaded_triangles[i] = NULL; - } - } - return cache->shaded_triangles; - + if (!BKE_mball_is_basis(ob)) { + return NULL; + } + + MetaBallBatchCache *cache = metaball_batch_cache_get(mb); + if (cache->shaded_triangles == NULL) { + cache->mat_len = gpumat_array_len; + cache->shaded_triangles = MEM_callocN(sizeof(*cache->shaded_triangles) * cache->mat_len, + __func__); + cache->shaded_triangles[0] = DRW_metaball_batch_cache_get_triangles_with_normals(ob); + for (int i = 1; i < cache->mat_len; ++i) { + cache->shaded_triangles[i] = NULL; + } + } + return cache->shaded_triangles; } GPUBatch *DRW_metaball_batch_cache_get_wireframes_face(Object *ob) { - if (!BKE_mball_is_basis(ob)) { - return NULL; - } + if (!BKE_mball_is_basis(ob)) { + return NULL; + } - MetaBall *mb = ob->data; - MetaBallBatchCache *cache = metaball_batch_cache_get(mb); + MetaBall *mb = ob->data; + MetaBallBatchCache *cache = metaball_batch_cache_get(mb); - if (cache->face_wire.batch == NULL) { - ListBase *lb = &ob->runtime.curve_cache->disp; + if (cache->face_wire.batch == NULL) { + ListBase *lb = &ob->runtime.curve_cache->disp; - GPUVertBuf *vbo_wiredata = MEM_callocN(sizeof(GPUVertBuf), __func__); - DRW_displist_vertbuf_create_wiredata(lb, vbo_wiredata); + GPUVertBuf *vbo_wiredata = MEM_callocN(sizeof(GPUVertBuf), __func__); + DRW_displist_vertbuf_create_wiredata(lb, vbo_wiredata); - GPUIndexBuf *ibo = MEM_callocN(sizeof(GPUIndexBuf), __func__); - DRW_displist_indexbuf_create_lines_in_order(lb, ibo); + GPUIndexBuf *ibo = MEM_callocN(sizeof(GPUIndexBuf), __func__); + DRW_displist_indexbuf_create_lines_in_order(lb, ibo); - cache->face_wire.batch = GPU_batch_create_ex( - GPU_PRIM_LINES, - mball_batch_cache_get_pos_and_normals(ob, cache), - ibo, - GPU_BATCH_OWNS_INDEX); + cache->face_wire.batch = GPU_batch_create_ex(GPU_PRIM_LINES, + mball_batch_cache_get_pos_and_normals(ob, cache), + ibo, + GPU_BATCH_OWNS_INDEX); - GPU_batch_vertbuf_add_ex(cache->face_wire.batch, vbo_wiredata, true); - } + GPU_batch_vertbuf_add_ex(cache->face_wire.batch, vbo_wiredata, true); + } - return cache->face_wire.batch; + return cache->face_wire.batch; } -struct GPUBatch *DRW_metaball_batch_cache_get_edge_detection(struct Object *ob, bool *r_is_manifold) +struct GPUBatch *DRW_metaball_batch_cache_get_edge_detection(struct Object *ob, + bool *r_is_manifold) { - if (!BKE_mball_is_basis(ob)) { - return NULL; - } + if (!BKE_mball_is_basis(ob)) { + return NULL; + } - MetaBall *mb = ob->data; - MetaBallBatchCache *cache = metaball_batch_cache_get(mb); + MetaBall *mb = ob->data; + MetaBallBatchCache *cache = metaball_batch_cache_get(mb); - if (cache->edge_detection == NULL) { - cache->edge_detection = GPU_batch_create( - GPU_PRIM_LINES_ADJ, - mball_batch_cache_get_pos_and_normals(ob, cache), - mball_batch_cache_get_edges_adj_lines(ob, cache)); - } + if (cache->edge_detection == NULL) { + cache->edge_detection = GPU_batch_create(GPU_PRIM_LINES_ADJ, + mball_batch_cache_get_pos_and_normals(ob, cache), + mball_batch_cache_get_edges_adj_lines(ob, cache)); + } - if (r_is_manifold) { - *r_is_manifold = cache->is_manifold; - } + if (r_is_manifold) { + *r_is_manifold = cache->is_manifold; + } - return cache->edge_detection; + return cache->edge_detection; } \ No newline at end of file -- cgit v1.2.3