From 0b8a21ee1668c3ab2e3682f1ae997ad7fb2227ae Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Ra=C3=BAl=20Kripalani?= Date: Sun, 1 Nov 2020 17:26:37 +0000 Subject: [PATCH] badger: restore prev. max table size (64MiB; default). --- node/repo/blockstore_opts.go | 9 ++++----- 1 file changed, 4 insertions(+), 5 deletions(-) diff --git a/node/repo/blockstore_opts.go b/node/repo/blockstore_opts.go index 3744d7647..2f9bc5047 100644 --- a/node/repo/blockstore_opts.go +++ b/node/repo/blockstore_opts.go @@ -35,13 +35,12 @@ func BadgerBlockstoreOptions(domain BlockstoreDomain, path string, readonly bool opts.ValueLogLoadingMode = badgerbs.FileIO opts.TableLoadingMode = badgerbs.MemoryMap - // Embed only values < 128 bytes in the LSM; larger values in value logs. + // Embed only values < 128 bytes in the LSM tree; larger values are stored + // in value logs. opts.ValueThreshold = 128 - // Reduce this from 64MiB to 16MiB. That means badger will hold on to - // 20MiB by default instead of 80MiB. This does not appear to have a - // significant performance hit. - opts.MaxTableSize = 16 << 20 + // Default table size is already 64MiB. This is here to make it explicit. + opts.MaxTableSize = 64 << 20 // NOTE: The chain blockstore doesn't require any GC (blocks are never // deleted). This will change if we move to a tiered blockstore.