Welcome to mirror list, hosted at ThFree Co, Russian Federation.

github.com/littlefs-project/littlefs.git - Unnamed repository; edit this file 'description' to name the repository.
summaryrefslogtreecommitdiff
path: root/lfs.c
diff options
context:
space:
mode:
authorChristopher Haster <geky@geky.net>2024-01-19 21:51:11 +0300
committerGitHub <noreply@github.com>2024-01-19 21:51:11 +0300
commit09972a1710a68f245ed2c8cd6f229748edfc1994 (patch)
tree013119c2b25c47bfe92943dcec8a43c7fbc54557 /lfs.c
parented7bd054357094688462aca9e302ce98ab24f35c (diff)
parentb5cd957f42d33f5960014dddbca7414502638a2d (diff)
Merge pull request #913 from littlefs-project/gc-compactions
Extend lfs_fs_gc to compact metadata, compact_thresh
Diffstat (limited to 'lfs.c')
-rw-r--r--lfs.c106
1 files changed, 83 insertions, 23 deletions
diff --git a/lfs.c b/lfs.c
index df6a230..44a8261 100644
--- a/lfs.c
+++ b/lfs.c
@@ -593,19 +593,6 @@ static int lfs_rawunmount(lfs_t *lfs);
/// Block allocator ///
-#ifndef LFS_READONLY
-static int lfs_alloc_lookahead(void *p, lfs_block_t block) {
- lfs_t *lfs = (lfs_t*)p;
- lfs_block_t off = ((block - lfs->lookahead.start)
- + lfs->block_count) % lfs->block_count;
-
- if (off < lfs->lookahead.size) {
- lfs->lookahead.buffer[off / 8] |= 1U << (off % 8);
- }
-
- return 0;
-}
-#endif
// allocations should call this when all allocated blocks are committed to
// the filesystem
@@ -624,7 +611,21 @@ static void lfs_alloc_drop(lfs_t *lfs) {
}
#ifndef LFS_READONLY
-static int lfs_fs_rawgc(lfs_t *lfs) {
+static int lfs_alloc_lookahead(void *p, lfs_block_t block) {
+ lfs_t *lfs = (lfs_t*)p;
+ lfs_block_t off = ((block - lfs->lookahead.start)
+ + lfs->block_count) % lfs->block_count;
+
+ if (off < lfs->lookahead.size) {
+ lfs->lookahead.buffer[off / 8] |= 1U << (off % 8);
+ }
+
+ return 0;
+}
+#endif
+
+#ifndef LFS_READONLY
+static int lfs_alloc_scan(lfs_t *lfs) {
// move lookahead buffer to the first unused block
//
// note we limit the lookahead buffer to at most the amount of blocks
@@ -693,7 +694,7 @@ static int lfs_alloc(lfs_t *lfs, lfs_block_t *block) {
// No blocks in our lookahead buffer, we need to scan the filesystem for
// unused blocks in the next lookahead window.
- int err = lfs_fs_rawgc(lfs);
+ int err = lfs_alloc_scan(lfs);
if(err) {
return err;
}
@@ -4189,6 +4190,14 @@ static int lfs_init(lfs_t *lfs, const struct lfs_config *cfg) {
// wear-leveling.
LFS_ASSERT(lfs->cfg->block_cycles != 0);
+ // check that compact_thresh makes sense
+ //
+ // metadata can't be compacted below block_size/2, and metadata can't
+ // exceed a block_size
+ LFS_ASSERT(lfs->cfg->compact_thresh == 0
+ || lfs->cfg->compact_thresh >= lfs->cfg->block_size/2);
+ LFS_ASSERT(lfs->cfg->compact_thresh == (lfs_size_t)-1
+ || lfs->cfg->compact_thresh <= lfs->cfg->block_size);
// setup read cache
if (lfs->cfg->read_buffer) {
@@ -5080,6 +5089,57 @@ static lfs_ssize_t lfs_fs_rawsize(lfs_t *lfs) {
return size;
}
+// explicit garbage collection
+#ifndef LFS_READONLY
+static int lfs_fs_rawgc(lfs_t *lfs) {
+ // force consistency, even if we're not necessarily going to write,
+ // because this function is supposed to take care of janitorial work
+ // isn't it?
+ int err = lfs_fs_forceconsistency(lfs);
+ if (err) {
+ return err;
+ }
+
+ // try to compact metadata pairs, note we can't really accomplish
+ // anything if compact_thresh doesn't at least leave a prog_size
+ // available
+ if (lfs->cfg->compact_thresh
+ < lfs->cfg->block_size - lfs->cfg->prog_size) {
+ // iterate over all mdirs
+ lfs_mdir_t mdir = {.tail = {0, 1}};
+ while (!lfs_pair_isnull(mdir.tail)) {
+ err = lfs_dir_fetch(lfs, &mdir, mdir.tail);
+ if (err) {
+ return err;
+ }
+
+ // not erased? exceeds our compaction threshold?
+ if (!mdir.erased || ((lfs->cfg->compact_thresh == 0)
+ ? mdir.off > lfs->cfg->block_size - lfs->cfg->block_size/8
+ : mdir.off > lfs->cfg->compact_thresh)) {
+ // the easiest way to trigger a compaction is to mark
+ // the mdir as unerased and add an empty commit
+ mdir.erased = false;
+ err = lfs_dir_commit(lfs, &mdir, NULL, 0);
+ if (err) {
+ return err;
+ }
+ }
+ }
+ }
+
+ // try to populate the lookahead buffer, unless it's already full
+ if (lfs->lookahead.size < 8*lfs->cfg->lookahead_size) {
+ err = lfs_alloc_scan(lfs);
+ if (err) {
+ return err;
+ }
+ }
+
+ return 0;
+}
+#endif
+
#ifndef LFS_READONLY
static int lfs_fs_rawgrow(lfs_t *lfs, lfs_size_t block_count) {
// shrinking is not supported
@@ -6286,32 +6346,32 @@ int lfs_fs_traverse(lfs_t *lfs, int (*cb)(void *, lfs_block_t), void *data) {
}
#ifndef LFS_READONLY
-int lfs_fs_gc(lfs_t *lfs) {
+int lfs_fs_mkconsistent(lfs_t *lfs) {
int err = LFS_LOCK(lfs->cfg);
if (err) {
return err;
}
- LFS_TRACE("lfs_fs_gc(%p)", (void*)lfs);
+ LFS_TRACE("lfs_fs_mkconsistent(%p)", (void*)lfs);
- err = lfs_fs_rawgc(lfs);
+ err = lfs_fs_rawmkconsistent(lfs);
- LFS_TRACE("lfs_fs_gc -> %d", err);
+ LFS_TRACE("lfs_fs_mkconsistent -> %d", err);
LFS_UNLOCK(lfs->cfg);
return err;
}
#endif
#ifndef LFS_READONLY
-int lfs_fs_mkconsistent(lfs_t *lfs) {
+int lfs_fs_gc(lfs_t *lfs) {
int err = LFS_LOCK(lfs->cfg);
if (err) {
return err;
}
- LFS_TRACE("lfs_fs_mkconsistent(%p)", (void*)lfs);
+ LFS_TRACE("lfs_fs_gc(%p)", (void*)lfs);
- err = lfs_fs_rawmkconsistent(lfs);
+ err = lfs_fs_rawgc(lfs);
- LFS_TRACE("lfs_fs_mkconsistent -> %d", err);
+ LFS_TRACE("lfs_fs_gc -> %d", err);
LFS_UNLOCK(lfs->cfg);
return err;
}