/* SPDX-License-Identifier: GPL-2.0-or-later */ /** \file * \ingroup edmesh */ #include "MEM_guardedalloc.h" #include "CLG_log.h" #include "DNA_key_types.h" #include "DNA_layer_types.h" #include "DNA_mesh_types.h" #include "DNA_meshdata_types.h" #include "DNA_object_types.h" #include "DNA_scene_types.h" #include "BLI_array_utils.h" #include "BLI_listbase.h" #include "BKE_context.h" #include "BKE_customdata.h" #include "BKE_editmesh.h" #include "BKE_key.h" #include "BKE_layer.h" #include "BKE_lib_id.h" #include "BKE_main.h" #include "BKE_mesh.h" #include "BKE_object.h" #include "BKE_undo_system.h" #include "DEG_depsgraph.h" #include "ED_mesh.h" #include "ED_object.h" #include "ED_undo.h" #include "ED_util.h" #include "WM_api.h" #include "WM_types.h" #define USE_ARRAY_STORE #ifdef USE_ARRAY_STORE // # define DEBUG_PRINT // # define DEBUG_TIME # ifdef DEBUG_TIME # include "PIL_time_utildefines.h" # endif # include "BLI_array_store.h" # include "BLI_array_store_utils.h" /* check on best size later... */ # define ARRAY_CHUNK_SIZE 256 # define USE_ARRAY_STORE_THREAD #endif #ifdef USE_ARRAY_STORE_THREAD # include "BLI_task.h" #endif /** We only need this locally. */ static CLG_LogRef LOG = {"ed.undo.mesh"}; /* -------------------------------------------------------------------- */ /** \name Undo Conversion * \{ */ #ifdef USE_ARRAY_STORE /* Single linked list of layers stored per type */ typedef struct BArrayCustomData { struct BArrayCustomData *next; CustomDataType type; int states_len; /* number of layers for each type */ BArrayState *states[0]; } BArrayCustomData; #endif typedef struct UndoMesh { /** * This undo-meshes in `um_arraystore.local_links`. * Not to be confused with the next and previous undo steps. */ struct UndoMesh *local_next, *local_prev; Mesh me; int selectmode; /** \note * This isn't a perfect solution, if you edit keys and change shapes this works well * (fixing T32442), but editing shape keys, going into object mode, removing or changing their * order, then go back into editmode and undo will give issues - where the old index will be * out of sync with the new object index. * * There are a few ways this could be made to work but for now its a known limitation with mixing * object and editmode operations - Campbell. */ int shapenr; #ifdef USE_ARRAY_STORE /* NULL arrays are considered empty */ struct { /* most data is stored as 'custom' data */ BArrayCustomData *vdata, *edata, *ldata, *pdata; BArrayState **keyblocks; BArrayState *mselect; } store; #endif /* USE_ARRAY_STORE */ size_t undo_size; } UndoMesh; #ifdef USE_ARRAY_STORE /* -------------------------------------------------------------------- */ /** \name Array Store * \{ */ static struct { struct BArrayStore_AtSize bs_stride; int users; /** * A list of #UndoMesh items ordered from oldest to newest * used to access previous undo data for a mesh. */ ListBase local_links; # ifdef USE_ARRAY_STORE_THREAD TaskPool *task_pool; # endif } um_arraystore = {{NULL}}; static void um_arraystore_cd_compact(struct CustomData *cdata, const size_t data_len, bool create, const BArrayCustomData *bcd_reference, BArrayCustomData **r_bcd_first) { if (data_len == 0) { if (create) { *r_bcd_first = NULL; } } const BArrayCustomData *bcd_reference_current = bcd_reference; BArrayCustomData *bcd = NULL, *bcd_first = NULL, *bcd_prev = NULL; for (int layer_start = 0, layer_end; layer_start < cdata->totlayer; layer_start = layer_end) { const CustomDataType type = cdata->layers[layer_start].type; /* Perform a full copy on dynamic layers. * * Unfortunately we can't compare dynamic layer types as they contain allocated pointers, * which burns CPU cycles looking for duplicate data that doesn't exist. * The array data isn't comparable once copied from the mesh, * this bottlenecks on high poly meshes, see T84114. * * Notes: * * - Ideally the data would be expanded into a format that could be de-duplicated effectively, * this would require a flat representation of each dynamic custom-data layer. * * - The data in the layer could be kept as-is to save on the extra copy, * it would complicate logic in this function. */ const bool layer_type_is_dynamic = CustomData_layertype_is_dynamic(type); layer_end = layer_start + 1; while ((layer_end < cdata->totlayer) && (type == cdata->layers[layer_end].type)) { layer_end++; } const int stride = CustomData_sizeof(type); BArrayStore *bs = create ? BLI_array_store_at_size_ensure( &um_arraystore.bs_stride, stride, ARRAY_CHUNK_SIZE) : NULL; const int layer_len = layer_end - layer_start; if (create) { if (bcd_reference_current && (bcd_reference_current->type == type)) { /* common case, the reference is aligned */ } else { bcd_reference_current = NULL; /* Do a full lookup when unaligned. */ if (bcd_reference) { const BArrayCustomData *bcd_iter = bcd_reference; while (bcd_iter) { if (bcd_iter->type == type) { bcd_reference_current = bcd_iter; break; } bcd_iter = bcd_iter->next; } } } } if (create) { bcd = MEM_callocN(sizeof(BArrayCustomData) + (layer_len * sizeof(BArrayState *)), __func__); bcd->next = NULL; bcd->type = type; bcd->states_len = layer_end - layer_start; if (bcd_prev) { bcd_prev->next = bcd; bcd_prev = bcd; } else { bcd_first = bcd; bcd_prev = bcd; } } CustomDataLayer *layer = &cdata->layers[layer_start]; for (int i = 0; i < layer_len; i++, layer++) { if (create) { if (layer->data) { BArrayState *state_reference = (bcd_reference_current && i < bcd_reference_current->states_len) ? bcd_reference_current->states[i] : NULL; /* See comment on `layer_type_is_dynamic` above. */ if (layer_type_is_dynamic) { state_reference = NULL; } bcd->states[i] = BLI_array_store_state_add( bs, layer->data, (size_t)data_len * stride, state_reference); } else { bcd->states[i] = NULL; } } if (layer->data) { MEM_freeN(layer->data); layer->data = NULL; } } if (create) { if (bcd_reference_current) { bcd_reference_current = bcd_reference_current->next; } } } if (create) { *r_bcd_first = bcd_first; } } /** * \note There is no room for data going out of sync here. * The layers and the states are stored together so this can be kept working. */ static void um_arraystore_cd_expand(const BArrayCustomData *bcd, struct CustomData *cdata, const size_t data_len) { CustomDataLayer *layer = cdata->layers; while (bcd) { const int stride = CustomData_sizeof(bcd->type); for (int i = 0; i < bcd->states_len; i++) { BLI_assert(bcd->type == layer->type); if (bcd->states[i]) { size_t state_len; layer->data = BLI_array_store_state_data_get_alloc(bcd->states[i], &state_len); BLI_assert(stride * data_len == state_len); UNUSED_VARS_NDEBUG(stride, data_len); } else { layer->data = NULL; } layer++; } bcd = bcd->next; } } static void um_arraystore_cd_free(BArrayCustomData *bcd) { while (bcd) { BArrayCustomData *bcd_next = bcd->next; const int stride = CustomData_sizeof(bcd->type); BArrayStore *bs = BLI_array_store_at_size_get(&um_arraystore.bs_stride, stride); for (int i = 0; i < bcd->states_len; i++) { if (bcd->states[i]) { BLI_array_store_state_remove(bs, bcd->states[i]); } } MEM_freeN(bcd); bcd = bcd_next; } } /** * \param create: When false, only free the arrays. * This is done since when reading from an undo state, they must be temporarily expanded. * then discarded afterwards, having this argument avoids having 2x code paths. */ static void um_arraystore_compact_ex(UndoMesh *um, const UndoMesh *um_ref, bool create) { Mesh *me = &um->me; um_arraystore_cd_compact( &me->vdata, me->totvert, create, um_ref ? um_ref->store.vdata : NULL, &um->store.vdata); um_arraystore_cd_compact( &me->edata, me->totedge, create, um_ref ? um_ref->store.edata : NULL, &um->store.edata); um_arraystore_cd_compact( &me->ldata, me->totloop, create, um_ref ? um_ref->store.ldata : NULL, &um->store.ldata); um_arraystore_cd_compact( &me->pdata, me->totpoly, create, um_ref ? um_ref->store.pdata : NULL, &um->store.pdata); if (me->key && me->key->totkey) { const size_t stride = me->key->elemsize; BArrayStore *bs = create ? BLI_array_store_at_size_ensure( &um_arraystore.bs_stride, stride, ARRAY_CHUNK_SIZE) : NULL; if (create) { um->store.keyblocks = MEM_mallocN(me->key->totkey * sizeof(*um->store.keyblocks), __func__); } KeyBlock *keyblock = me->key->block.first; for (int i = 0; i < me->key->totkey; i++, keyblock = keyblock->next) { if (create) { BArrayState *state_reference = (um_ref && um_ref->me.key && (i < um_ref->me.key->totkey)) ? um_ref->store.keyblocks[i] : NULL; um->store.keyblocks[i] = BLI_array_store_state_add( bs, keyblock->data, (size_t)keyblock->totelem * stride, state_reference); } if (keyblock->data) { MEM_freeN(keyblock->data); keyblock->data = NULL; } } } if (me->mselect && me->totselect) { BLI_assert(create == (um->store.mselect == NULL)); if (create) { BArrayState *state_reference = um_ref ? um_ref->store.mselect : NULL; const size_t stride = sizeof(*me->mselect); BArrayStore *bs = BLI_array_store_at_size_ensure( &um_arraystore.bs_stride, stride, ARRAY_CHUNK_SIZE); um->store.mselect = BLI_array_store_state_add( bs, me->mselect, (size_t)me->totselect * stride, state_reference); } /* keep me->totselect for validation */ MEM_freeN(me->mselect); me->mselect = NULL; } if (create) { um_arraystore.users += 1; } BKE_mesh_update_customdata_pointers(me, false); } /** * Move data from allocated arrays to de-duplicated states and clear arrays. */ static void um_arraystore_compact(UndoMesh *um, const UndoMesh *um_ref) { um_arraystore_compact_ex(um, um_ref, true); } static void um_arraystore_compact_with_info(UndoMesh *um, const UndoMesh *um_ref) { # ifdef DEBUG_PRINT size_t size_expanded_prev, size_compacted_prev; BLI_array_store_at_size_calc_memory_usage( &um_arraystore.bs_stride, &size_expanded_prev, &size_compacted_prev); # endif # ifdef DEBUG_TIME TIMEIT_START(mesh_undo_compact); # endif um_arraystore_compact(um, um_ref); # ifdef DEBUG_TIME TIMEIT_END(mesh_undo_compact); # endif # ifdef DEBUG_PRINT { size_t size_expanded, size_compacted; BLI_array_store_at_size_calc_memory_usage( &um_arraystore.bs_stride, &size_expanded, &size_compacted); const double percent_total = size_expanded ? (((double)size_compacted / (double)size_expanded) * 100.0) : -1.0; size_t size_expanded_step = size_expanded - size_expanded_prev; size_t size_compacted_step = size_compacted - size_compacted_prev; const double percent_step = size_expanded_step ? (((double)size_compacted_step / (double)size_expanded_step) * 100.0) : -1.0; printf("overall memory use: %.8f%% of expanded size\n", percent_total); printf("step memory use: %.8f%% of expanded size\n", percent_step); } # endif } # ifdef USE_ARRAY_STORE_THREAD struct UMArrayData { UndoMesh *um; const UndoMesh *um_ref; /* can be NULL */ }; static void um_arraystore_compact_cb(TaskPool *__restrict UNUSED(pool), void *taskdata) { struct UMArrayData *um_data = taskdata; um_arraystore_compact_with_info(um_data->um, um_data->um_ref); } # endif /* USE_ARRAY_STORE_THREAD */ /** * Remove data we only expanded for temporary use. */ static void um_arraystore_expand_clear(UndoMesh *um) { um_arraystore_compact_ex(um, NULL, false); } static void um_arraystore_expand(UndoMesh *um) { Mesh *me = &um->me; um_arraystore_cd_expand(um->store.vdata, &me->vdata, me->totvert); um_arraystore_cd_expand(um->store.edata, &me->edata, me->totedge); um_arraystore_cd_expand(um->store.ldata, &me->ldata, me->totloop); um_arraystore_cd_expand(um->store.pdata, &me->pdata, me->totpoly); if (um->store.keyblocks) { const size_t stride = me->key->elemsize; KeyBlock *keyblock = me->key->block.first; for (int i = 0; i < me->key->totkey; i++, keyblock = keyblock->next) { BArrayState *state = um->store.keyblocks[i]; size_t state_len; keyblock->data = BLI_array_store_state_data_get_alloc(state, &state_len); BLI_assert(keyblock->totelem == (state_len / stride)); UNUSED_VARS_NDEBUG(stride); } } if (um->store.mselect) { const size_t stride = sizeof(*me->mselect); BArrayState *state = um->store.mselect; size_t state_len; me->mselect = BLI_array_store_state_data_get_alloc(state, &state_len); BLI_assert(me->totselect == (state_len / stride)); UNUSED_VARS_NDEBUG(stride); } /* not essential, but prevents accidental dangling pointer access */ BKE_mesh_update_customdata_pointers(me, false); } static void um_arraystore_free(UndoMesh *um) { Mesh *me = &um->me; um_arraystore_cd_free(um->store.vdata); um_arraystore_cd_free(um->store.edata); um_arraystore_cd_free(um->store.ldata); um_arraystore_cd_free(um->store.pdata); if (um->store.keyblocks) { const size_t stride = me->key->elemsize; BArrayStore *bs = BLI_array_store_at_size_get(&um_arraystore.bs_stride, stride); for (int i = 0; i < me->key->totkey; i++) { BArrayState *state = um->store.keyblocks[i]; BLI_array_store_state_remove(bs, state); } MEM_freeN(um->store.keyblocks); um->store.keyblocks = NULL; } if (um->store.mselect) { const size_t stride = sizeof(*me->mselect); BArrayStore *bs = BLI_array_store_at_size_get(&um_arraystore.bs_stride, stride); BArrayState *state = um->store.mselect; BLI_array_store_state_remove(bs, state); um->store.mselect = NULL; } um_arraystore.users -= 1; BLI_assert(um_arraystore.users >= 0); if (um_arraystore.users == 0) { # ifdef DEBUG_PRINT printf("mesh undo store: freeing all data!\n"); # endif BLI_array_store_at_size_clear(&um_arraystore.bs_stride); # ifdef USE_ARRAY_STORE_THREAD BLI_task_pool_free(um_arraystore.task_pool); um_arraystore.task_pool = NULL; # endif } } /** \} */ /* -------------------------------------------------------------------- */ /** \name Array Store Utilities * \{ */ /** * Create an array of #UndoMesh from `objects`. * * where each element in the resulting array is the most recently created * undo-mesh for the object's mesh. * When no undo-mesh can be found that array index is NULL. * * This is used for de-duplicating memory between undo steps, * failure to find the undo step will store a full duplicate in memory. * define `DEBUG_PRINT` to check memory is de-duplicating as expected. */ static UndoMesh **mesh_undostep_reference_elems_from_objects(Object **object, int object_len) { /* Map: `Mesh.id.session_uuid` -> `UndoMesh`. */ GHash *uuid_map = BLI_ghash_ptr_new_ex(__func__, object_len); UndoMesh **um_references = MEM_callocN(sizeof(UndoMesh *) * object_len, __func__); for (int i = 0; i < object_len; i++) { const Mesh *me = object[i]->data; BLI_ghash_insert(uuid_map, POINTER_FROM_INT(me->id.session_uuid), &um_references[i]); } int uuid_map_len = object_len; /* Loop backwards over all previous mesh undo data until either: * - All elements have been found (where `um_references` we'll have every element set). * - There are no undo steps left to look for. */ UndoMesh *um_iter = um_arraystore.local_links.last; while (um_iter && (uuid_map_len != 0)) { UndoMesh **um_p; if ((um_p = BLI_ghash_popkey(uuid_map, POINTER_FROM_INT(um_iter->me.id.session_uuid), NULL))) { *um_p = um_iter; uuid_map_len--; } um_iter = um_iter->local_prev; } BLI_assert(uuid_map_len == BLI_ghash_len(uuid_map)); BLI_ghash_free(uuid_map, NULL, NULL); if (uuid_map_len == object_len) { MEM_freeN(um_references); um_references = NULL; } return um_references; } /** \} */ #endif /* USE_ARRAY_STORE */ /* for callbacks */ /* undo simply makes copies of a bmesh */ /** * \param um_ref: The reference to use for de-duplicating memory between undo-steps. */ static void *undomesh_from_editmesh(UndoMesh *um, BMEditMesh *em, Key *key, UndoMesh *um_ref) { BLI_assert(BLI_array_is_zeroed(um, 1)); #ifdef USE_ARRAY_STORE_THREAD /* changes this waits is low, but must have finished */ if (um_arraystore.task_pool) { BLI_task_pool_work_and_wait(um_arraystore.task_pool); } #endif /* make sure shape keys work */ if (key != NULL) { um->me.key = (Key *)BKE_id_copy_ex( NULL, &key->id, NULL, LIB_ID_COPY_LOCALIZE | LIB_ID_COPY_NO_ANIMDATA); } else { um->me.key = NULL; } /* BM_mesh_validate(em->bm); */ /* for troubleshooting */ BM_mesh_bm_to_me( NULL, em->bm, &um->me, (&(struct BMeshToMeshParams){ /* Undo code should not be manipulating 'G_MAIN->object' hooks/vertex-parent. */ .calc_object_remap = false, .update_shapekey_indices = false, .cd_mask_extra = {.vmask = CD_MASK_SHAPE_KEYINDEX}, })); um->selectmode = em->selectmode; um->shapenr = em->bm->shapenr; #ifdef USE_ARRAY_STORE { /* Add ourselves. */ BLI_addtail(&um_arraystore.local_links, um); # ifdef USE_ARRAY_STORE_THREAD if (um_arraystore.task_pool == NULL) { um_arraystore.task_pool = BLI_task_pool_create_background(NULL, TASK_PRIORITY_LOW); } struct UMArrayData *um_data = MEM_mallocN(sizeof(*um_data), __func__); um_data->um = um; um_data->um_ref = um_ref; BLI_task_pool_push(um_arraystore.task_pool, um_arraystore_compact_cb, um_data, true, NULL); # else um_arraystore_compact_with_info(um, um_ref); # endif } #else UNUSED_VARS(um_ref); #endif return um; } static void undomesh_to_editmesh(UndoMesh *um, Object *ob, BMEditMesh *em, Key *key) { BMEditMesh *em_tmp; BMesh *bm; #ifdef USE_ARRAY_STORE # ifdef USE_ARRAY_STORE_THREAD /* changes this waits is low, but must have finished */ BLI_task_pool_work_and_wait(um_arraystore.task_pool); # endif # ifdef DEBUG_TIME TIMEIT_START(mesh_undo_expand); # endif um_arraystore_expand(um); # ifdef DEBUG_TIME TIMEIT_END(mesh_undo_expand); # endif #endif /* USE_ARRAY_STORE */ const BMAllocTemplate allocsize = BMALLOC_TEMPLATE_FROM_ME(&um->me); em->bm->shapenr = um->shapenr; EDBM_mesh_free_data(em); bm = BM_mesh_create(&allocsize, &((struct BMeshCreateParams){ .use_toolflags = true, })); BM_mesh_bm_from_me(bm, &um->me, (&(struct BMeshFromMeshParams){ /* Handled with tessellation. */ .calc_face_normal = false, .active_shapekey = um->shapenr, })); em_tmp = BKE_editmesh_create(bm); *em = *em_tmp; /* Normals should not be stored in the undo mesh, so recalculate them. The edit * mesh is expected to have valid normals and there is no tracked dirty state. */ BLI_assert(BKE_mesh_vertex_normals_are_dirty(&um->me)); /* Calculate face normals and tessellation at once since it's multi-threaded. */ BKE_editmesh_looptri_and_normals_calc(em); em->selectmode = um->selectmode; bm->selectmode = um->selectmode; bm->spacearr_dirty = BM_SPACEARR_DIRTY_ALL; /* T35170: Restore the active key on the RealMesh. Otherwise 'fake' offset propagation happens * if the active is a basis for any other. */ if (key && (key->type == KEY_RELATIVE)) { /* Since we can't add, remove or reorder keyblocks in editmode, it's safe to assume * shapenr from restored bmesh and keyblock indices are in sync. */ const int kb_act_idx = ob->shapenr - 1; /* If it is, let's patch the current mesh key block to its restored value. * Else, the offsets won't be computed and it won't matter. */ if (BKE_keyblock_is_basis(key, kb_act_idx)) { KeyBlock *kb_act = BLI_findlink(&key->block, kb_act_idx); if (kb_act->totelem != um->me.totvert) { /* The current mesh has some extra/missing verts compared to the undo, adjust. */ MEM_SAFE_FREE(kb_act->data); kb_act->data = MEM_mallocN((size_t)(key->elemsize) * bm->totvert, __func__); kb_act->totelem = um->me.totvert; } BKE_keyblock_update_from_mesh(&um->me, kb_act); } } ob->shapenr = um->shapenr; MEM_freeN(em_tmp); #ifdef USE_ARRAY_STORE um_arraystore_expand_clear(um); #endif } static void undomesh_free_data(UndoMesh *um) { Mesh *me = &um->me; #ifdef USE_ARRAY_STORE # ifdef USE_ARRAY_STORE_THREAD /* changes this waits is low, but must have finished */ BLI_task_pool_work_and_wait(um_arraystore.task_pool); # endif /* we need to expand so any allocations in custom-data are freed with the mesh */ um_arraystore_expand(um); BLI_assert(BLI_findindex(&um_arraystore.local_links, um) != -1); BLI_remlink(&um_arraystore.local_links, um); um_arraystore_free(um); #endif if (me->key) { BKE_key_free_data(me->key); MEM_freeN(me->key); } BKE_mesh_free_data_for_undo(me); } static Object *editmesh_object_from_context(bContext *C) { ViewLayer *view_layer = CTX_data_view_layer(C); Object *obedit = OBEDIT_FROM_VIEW_LAYER(view_layer); if (obedit && obedit->type == OB_MESH) { Mesh *me = obedit->data; if (me->edit_mesh != NULL) { return obedit; } } return NULL; } /** \} */ /* -------------------------------------------------------------------- */ /** \name Implements ED Undo System * * \note This is similar for all edit-mode types. * \{ */ typedef struct MeshUndoStep_Elem { UndoRefID_Object obedit_ref; UndoMesh data; } MeshUndoStep_Elem; typedef struct MeshUndoStep { UndoStep step; MeshUndoStep_Elem *elems; uint elems_len; } MeshUndoStep; static bool mesh_undosys_poll(bContext *C) { return editmesh_object_from_context(C) != NULL; } static bool mesh_undosys_step_encode(struct bContext *C, struct Main *bmain, UndoStep *us_p) { MeshUndoStep *us = (MeshUndoStep *)us_p; /* Important not to use the 3D view when getting objects because all objects * outside of this list will be moved out of edit-mode when reading back undo steps. */ ViewLayer *view_layer = CTX_data_view_layer(C); uint objects_len = 0; Object **objects = ED_undo_editmode_objects_from_view_layer(view_layer, &objects_len); us->elems = MEM_callocN(sizeof(*us->elems) * objects_len, __func__); us->elems_len = objects_len; UndoMesh **um_references = NULL; #ifdef USE_ARRAY_STORE um_references = mesh_undostep_reference_elems_from_objects(objects, objects_len); #endif for (uint i = 0; i < objects_len; i++) { Object *ob = objects[i]; MeshUndoStep_Elem *elem = &us->elems[i]; elem->obedit_ref.ptr = ob; Mesh *me = elem->obedit_ref.ptr->data; BMEditMesh *em = me->edit_mesh; undomesh_from_editmesh( &elem->data, me->edit_mesh, me->key, um_references ? um_references[i] : NULL); em->needs_flush_to_id = 1; us->step.data_size += elem->data.undo_size; #ifdef USE_ARRAY_STORE /** As this is only data storage it is safe to set the session ID here. */ elem->data.me.id.session_uuid = me->id.session_uuid; #endif } MEM_freeN(objects); if (um_references != NULL) { MEM_freeN(um_references); } bmain->is_memfile_undo_flush_needed = true; return true; } static void mesh_undosys_step_decode(struct bContext *C, struct Main *bmain, UndoStep *us_p, const eUndoStepDir UNUSED(dir), bool UNUSED(is_final)) { MeshUndoStep *us = (MeshUndoStep *)us_p; ED_undo_object_editmode_restore_helper( C, &us->elems[0].obedit_ref.ptr, us->elems_len, sizeof(*us->elems)); BLI_assert(BKE_object_is_in_editmode(us->elems[0].obedit_ref.ptr)); for (uint i = 0; i < us->elems_len; i++) { MeshUndoStep_Elem *elem = &us->elems[i]; Object *obedit = elem->obedit_ref.ptr; Mesh *me = obedit->data; if (me->edit_mesh == NULL) { /* Should never fail, may not crash but can give odd behavior. */ CLOG_ERROR(&LOG, "name='%s', failed to enter edit-mode for object '%s', undo state invalid", us_p->name, obedit->id.name); continue; } BMEditMesh *em = me->edit_mesh; undomesh_to_editmesh(&elem->data, obedit, em, me->key); em->needs_flush_to_id = 1; DEG_id_tag_update(&me->id, ID_RECALC_GEOMETRY); } /* The first element is always active */ ED_undo_object_set_active_or_warn( CTX_data_scene(C), CTX_data_view_layer(C), us->elems[0].obedit_ref.ptr, us_p->name, &LOG); /* Check after setting active. */ BLI_assert(mesh_undosys_poll(C)); Scene *scene = CTX_data_scene(C); scene->toolsettings->selectmode = us->elems[0].data.selectmode; bmain->is_memfile_undo_flush_needed = true; WM_event_add_notifier(C, NC_GEOM | ND_DATA, NULL); } static void mesh_undosys_step_free(UndoStep *us_p) { MeshUndoStep *us = (MeshUndoStep *)us_p; for (uint i = 0; i < us->elems_len; i++) { MeshUndoStep_Elem *elem = &us->elems[i]; undomesh_free_data(&elem->data); } MEM_freeN(us->elems); } static void mesh_undosys_foreach_ID_ref(UndoStep *us_p, UndoTypeForEachIDRefFn foreach_ID_ref_fn, void *user_data) { MeshUndoStep *us = (MeshUndoStep *)us_p; for (uint i = 0; i < us->elems_len; i++) { MeshUndoStep_Elem *elem = &us->elems[i]; foreach_ID_ref_fn(user_data, ((UndoRefID *)&elem->obedit_ref)); } } void ED_mesh_undosys_type(UndoType *ut) { ut->name = "Edit Mesh"; ut->poll = mesh_undosys_poll; ut->step_encode = mesh_undosys_step_encode; ut->step_decode = mesh_undosys_step_decode; ut->step_free = mesh_undosys_step_free; ut->step_foreach_ID_ref = mesh_undosys_foreach_ID_ref; ut->flags = UNDOTYPE_FLAG_NEED_CONTEXT_FOR_ENCODE; ut->step_size = sizeof(MeshUndoStep); } /** \} */