Welcome to mirror list, hosted at ThFree Co, Russian Federation.

git.blender.org/blender.git - Unnamed repository; edit this file 'description' to name the repository.
summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
Diffstat (limited to 'source/blender/blenloader/intern/undofile.cc')
-rw-r--r--source/blender/blenloader/intern/undofile.cc349
1 files changed, 349 insertions, 0 deletions
diff --git a/source/blender/blenloader/intern/undofile.cc b/source/blender/blenloader/intern/undofile.cc
new file mode 100644
index 00000000000..99d1ca336fe
--- /dev/null
+++ b/source/blender/blenloader/intern/undofile.cc
@@ -0,0 +1,349 @@
+/* SPDX-License-Identifier: GPL-2.0-or-later
+ * Copyright 2004 Blender Foundation. All rights reserved. */
+
+/** \file
+ * \ingroup blenloader
+ */
+
+#include <errno.h>
+#include <fcntl.h>
+#include <math.h>
+#include <stdio.h>
+#include <stdlib.h>
+#include <string.h>
+
+/* open/close */
+#ifndef _WIN32
+# include <unistd.h>
+#else
+# include <io.h>
+#endif
+
+#include "MEM_guardedalloc.h"
+
+#include "DNA_listBase.h"
+
+#include "BLI_blenlib.h"
+#include "BLI_ghash.h"
+
+#include "BLO_readfile.h"
+#include "BLO_undofile.h"
+
+#include "BKE_lib_id.h"
+#include "BKE_main.h"
+#include "BKE_undo_system.h"
+
+/* keep last */
+#include "BLI_strict_flags.h"
+
+/* **************** support for memory-write, for undo buffers *************** */
+
+void BLO_memfile_free(MemFile *memfile)
+{
+ MemFileChunk *chunk;
+
+ while ((chunk = static_cast<MemFileChunk *>(BLI_pophead(&memfile->chunks)))) {
+ if (chunk->is_identical == false) {
+ MEM_freeN((void *)chunk->buf);
+ }
+ MEM_freeN(chunk);
+ }
+ memfile->size = 0;
+}
+
+void BLO_memfile_merge(MemFile *first, MemFile *second)
+{
+ /* We use this mapping to store the memory buffers from second memfile chunks which are not owned
+ * by it (i.e. shared with some previous memory steps). */
+ GHash *buffer_to_second_memchunk = BLI_ghash_new(
+ BLI_ghashutil_ptrhash, BLI_ghashutil_ptrcmp, __func__);
+
+ /* First, detect all memchunks in second memfile that are not owned by it. */
+ for (MemFileChunk *sc = static_cast<MemFileChunk *>(second->chunks.first); sc != nullptr;
+ sc = static_cast<MemFileChunk *>(sc->next)) {
+ if (sc->is_identical) {
+ BLI_ghash_insert(buffer_to_second_memchunk, (void *)sc->buf, sc);
+ }
+ }
+
+ /* Now, check all chunks from first memfile (the one we are removing), and if a memchunk owned by
+ * it is also used by the second memfile, transfer the ownership. */
+ for (MemFileChunk *fc = static_cast<MemFileChunk *>(first->chunks.first); fc != nullptr;
+ fc = static_cast<MemFileChunk *>(fc->next)) {
+ if (!fc->is_identical) {
+ MemFileChunk *sc = static_cast<MemFileChunk *>(
+ BLI_ghash_lookup(buffer_to_second_memchunk, fc->buf));
+ if (sc != nullptr) {
+ BLI_assert(sc->is_identical);
+ sc->is_identical = false;
+ fc->is_identical = true;
+ }
+ /* Note that if the second memfile does not use that chunk, we assume that the first one
+ * fully owns it without sharing it with any other memfile, and hence it should be freed with
+ * it. */
+ }
+ }
+
+ BLI_ghash_free(buffer_to_second_memchunk, nullptr, nullptr);
+
+ BLO_memfile_free(first);
+}
+
+void BLO_memfile_clear_future(MemFile *memfile)
+{
+ LISTBASE_FOREACH (MemFileChunk *, chunk, &memfile->chunks) {
+ chunk->is_identical_future = false;
+ }
+}
+
+void BLO_memfile_write_init(MemFileWriteData *mem_data,
+ MemFile *written_memfile,
+ MemFile *reference_memfile)
+{
+ mem_data->written_memfile = written_memfile;
+ mem_data->reference_memfile = reference_memfile;
+ mem_data->reference_current_chunk = reference_memfile ? static_cast<MemFileChunk *>(
+ reference_memfile->chunks.first) :
+ nullptr;
+
+ /* If we have a reference memfile, we generate a mapping between the session_uuid's of the
+ * IDs stored in that previous undo step, and its first matching memchunk. This will allow
+ * us to easily find the existing undo memory storage of IDs even when some re-ordering in
+ * current Main data-base broke the order matching with the memchunks from previous step.
+ */
+ if (reference_memfile != nullptr) {
+ mem_data->id_session_uuid_mapping = BLI_ghash_new(
+ BLI_ghashutil_inthash_p_simple, BLI_ghashutil_intcmp, __func__);
+ uint current_session_uuid = MAIN_ID_SESSION_UUID_UNSET;
+ LISTBASE_FOREACH (MemFileChunk *, mem_chunk, &reference_memfile->chunks) {
+ if (!ELEM(mem_chunk->id_session_uuid, MAIN_ID_SESSION_UUID_UNSET, current_session_uuid)) {
+ current_session_uuid = mem_chunk->id_session_uuid;
+ void **entry;
+ if (!BLI_ghash_ensure_p(mem_data->id_session_uuid_mapping,
+ POINTER_FROM_UINT(current_session_uuid),
+ &entry)) {
+ *entry = mem_chunk;
+ }
+ else {
+ BLI_assert_unreachable();
+ }
+ }
+ }
+ }
+}
+
+void BLO_memfile_write_finalize(MemFileWriteData *mem_data)
+{
+ if (mem_data->id_session_uuid_mapping != nullptr) {
+ BLI_ghash_free(mem_data->id_session_uuid_mapping, nullptr, nullptr);
+ }
+}
+
+void BLO_memfile_chunk_add(MemFileWriteData *mem_data, const char *buf, size_t size)
+{
+ MemFile *memfile = mem_data->written_memfile;
+ MemFileChunk **compchunk_step = &mem_data->reference_current_chunk;
+
+ MemFileChunk *curchunk = static_cast<MemFileChunk *>(
+ MEM_mallocN(sizeof(MemFileChunk), "MemFileChunk"));
+ curchunk->size = size;
+ curchunk->buf = nullptr;
+ curchunk->is_identical = false;
+ /* This is unsafe in the sense that an app handler or other code that does not
+ * perform an undo push may make changes after the last undo push that
+ * will then not be undo. Though it's not entirely clear that is wrong behavior. */
+ curchunk->is_identical_future = true;
+ curchunk->id_session_uuid = mem_data->current_id_session_uuid;
+ BLI_addtail(&memfile->chunks, curchunk);
+
+ /* we compare compchunk with buf */
+ if (*compchunk_step != nullptr) {
+ MemFileChunk *compchunk = *compchunk_step;
+ if (compchunk->size == curchunk->size) {
+ if (memcmp(compchunk->buf, buf, size) == 0) {
+ curchunk->buf = compchunk->buf;
+ curchunk->is_identical = true;
+ compchunk->is_identical_future = true;
+ }
+ }
+ *compchunk_step = static_cast<MemFileChunk *>(compchunk->next);
+ }
+
+ /* not equal... */
+ if (curchunk->buf == nullptr) {
+ char *buf_new = static_cast<char *>(MEM_mallocN(size, "Chunk buffer"));
+ memcpy(buf_new, buf, size);
+ curchunk->buf = buf_new;
+ memfile->size += size;
+ }
+}
+
+struct Main *BLO_memfile_main_get(struct MemFile *memfile,
+ struct Main *bmain,
+ struct Scene **r_scene)
+{
+ struct Main *bmain_undo = nullptr;
+ BlendFileReadParams read_params{};
+ BlendFileData *bfd = BLO_read_from_memfile(
+ bmain, BKE_main_blendfile_path(bmain), memfile, &read_params, nullptr);
+
+ if (bfd) {
+ bmain_undo = bfd->main;
+ if (r_scene) {
+ *r_scene = bfd->curscene;
+ }
+
+ MEM_freeN(bfd);
+ }
+
+ return bmain_undo;
+}
+
+bool BLO_memfile_write_file(struct MemFile *memfile, const char *filepath)
+{
+ MemFileChunk *chunk;
+ int file, oflags;
+
+ /* NOTE: This is currently used for autosave and 'quit.blend',
+ * where _not_ following symlinks is OK,
+ * however if this is ever executed explicitly by the user,
+ * we may want to allow writing to symlinks.
+ */
+
+ oflags = O_BINARY | O_WRONLY | O_CREAT | O_TRUNC;
+#ifdef O_NOFOLLOW
+ /* use O_NOFOLLOW to avoid writing to a symlink - use 'O_EXCL' (CVE-2008-1103) */
+ oflags |= O_NOFOLLOW;
+#else
+ /* TODO(sergey): How to deal with symlinks on windows? */
+# ifndef _MSC_VER
+# warning "Symbolic links will be followed on undo save, possibly causing CVE-2008-1103"
+# endif
+#endif
+ file = BLI_open(filepath, oflags, 0666);
+
+ if (file == -1) {
+ fprintf(stderr,
+ "Unable to save '%s': %s\n",
+ filepath,
+ errno ? strerror(errno) : "Unknown error opening file");
+ return false;
+ }
+
+ for (chunk = static_cast<MemFileChunk *>(memfile->chunks.first); chunk;
+ chunk = static_cast<MemFileChunk *>(chunk->next)) {
+#ifdef _WIN32
+ if (size_t(write(file, chunk->buf, uint(chunk->size))) != chunk->size)
+#else
+ if (size_t(write(file, chunk->buf, chunk->size)) != chunk->size)
+#endif
+ {
+ break;
+ }
+ }
+
+ close(file);
+
+ if (chunk) {
+ fprintf(stderr,
+ "Unable to save '%s': %s\n",
+ filepath,
+ errno ? strerror(errno) : "Unknown error writing file");
+ return false;
+ }
+ return true;
+}
+
+static ssize_t undo_read(FileReader *reader, void *buffer, size_t size)
+{
+ UndoReader *undo = (UndoReader *)reader;
+
+ static size_t seek = SIZE_MAX; /* The current position. */
+ static size_t offset = 0; /* Size of previous chunks. */
+ static MemFileChunk *chunk = nullptr;
+ size_t chunkoffset, readsize, totread;
+
+ undo->memchunk_identical = true;
+
+ if (size == 0) {
+ return 0;
+ }
+
+ if (seek != size_t(undo->reader.offset)) {
+ chunk = static_cast<MemFileChunk *>(undo->memfile->chunks.first);
+ seek = 0;
+
+ while (chunk) {
+ if (seek + chunk->size > size_t(undo->reader.offset)) {
+ break;
+ }
+ seek += chunk->size;
+ chunk = static_cast<MemFileChunk *>(chunk->next);
+ }
+ offset = seek;
+ seek = size_t(undo->reader.offset);
+ }
+
+ if (chunk) {
+ totread = 0;
+
+ do {
+ /* First check if it's on the end if current chunk. */
+ if (seek - offset == chunk->size) {
+ offset += chunk->size;
+ chunk = static_cast<MemFileChunk *>(chunk->next);
+ }
+
+ /* Debug, should never happen. */
+ if (chunk == nullptr) {
+ printf("illegal read, chunk zero\n");
+ return 0;
+ }
+
+ chunkoffset = seek - offset;
+ readsize = size - totread;
+
+ /* Data can be spread over multiple chunks, so clamp size
+ * to within this chunk, and then it will read further in
+ * the next chunk. */
+ if (chunkoffset + readsize > chunk->size) {
+ readsize = chunk->size - chunkoffset;
+ }
+
+ memcpy(POINTER_OFFSET(buffer, totread), chunk->buf + chunkoffset, readsize);
+ totread += readsize;
+ undo->reader.offset += (off64_t)readsize;
+ seek += readsize;
+
+ /* `is_identical` of current chunk represents whether it changed compared to previous undo
+ * step. this is fine in redo case, but not in undo case, where we need an extra flag
+ * defined when saving the next (future) step after the one we want to restore, as we are
+ * supposed to 'come from' that future undo step, and not the one before current one. */
+ undo->memchunk_identical &= undo->undo_direction == STEP_REDO ? chunk->is_identical :
+ chunk->is_identical_future;
+ } while (totread < size);
+
+ return ssize_t(totread);
+ }
+
+ return 0;
+}
+
+static void undo_close(FileReader *reader)
+{
+ MEM_freeN(reader);
+}
+
+FileReader *BLO_memfile_new_filereader(MemFile *memfile, int undo_direction)
+{
+ UndoReader *undo = static_cast<UndoReader *>(MEM_callocN(sizeof(UndoReader), __func__));
+
+ undo->memfile = memfile;
+ undo->undo_direction = undo_direction;
+
+ undo->reader.read = undo_read;
+ undo->reader.seek = nullptr;
+ undo->reader.close = undo_close;
+
+ return (FileReader *)undo;
+}