mirror of
https://source.quilibrium.com/quilibrium/ceremonyclient.git
synced 2024-12-24 23:55:18 +00:00
1087 lines
41 KiB
Go
1087 lines
41 KiB
Go
// Copyright 2020 The LevelDB-Go and Pebble Authors. All rights reserved. Use
|
||
// of this source code is governed by a BSD-style license that can be found in
|
||
// the LICENSE file.
|
||
|
||
package pebble
|
||
|
||
import (
|
||
"fmt"
|
||
"math"
|
||
|
||
"github.com/cockroachdb/errors"
|
||
"github.com/cockroachdb/pebble/internal/base"
|
||
"github.com/cockroachdb/pebble/internal/keyspan"
|
||
"github.com/cockroachdb/pebble/internal/manifest"
|
||
"github.com/cockroachdb/pebble/sstable"
|
||
)
|
||
|
||
// In-memory statistics about tables help inform compaction picking, but may
|
||
// be expensive to calculate or load from disk. Every time a database is
|
||
// opened, these statistics must be reloaded or recalculated. To minimize
|
||
// impact on user activity and compactions, we load these statistics
|
||
// asynchronously in the background and store loaded statistics in each
|
||
// table's *FileMetadata.
|
||
//
|
||
// This file implements the asynchronous loading of statistics by maintaining
|
||
// a list of files that require statistics, alongside their LSM levels.
|
||
// Whenever new files are added to the LSM, the files are appended to
|
||
// d.mu.tableStats.pending. If a stats collection job is not currently
|
||
// running, one is started in a separate goroutine.
|
||
//
|
||
// The stats collection job grabs and clears the pending list, computes table
|
||
// statistics relative to the current readState and updates the tables' file
|
||
// metadata. New pending files may accumulate during a stats collection job,
|
||
// so a completing job triggers a new job if necessary. Only one job runs at a
|
||
// time.
|
||
//
|
||
// When an existing database is opened, all files lack in-memory statistics.
|
||
// These files' stats are loaded incrementally whenever the pending list is
|
||
// empty by scanning a current readState for files missing statistics. Once a
|
||
// job completes a scan without finding any remaining files without
|
||
// statistics, it flips a `loadedInitial` flag. From then on, the stats
|
||
// collection job only needs to load statistics for new files appended to the
|
||
// pending list.
|
||
|
||
func (d *DB) maybeCollectTableStatsLocked() {
|
||
if d.shouldCollectTableStatsLocked() {
|
||
go d.collectTableStats()
|
||
}
|
||
}
|
||
|
||
// updateTableStatsLocked is called when new files are introduced, after the
|
||
// read state has been updated. It may trigger a new stat collection.
|
||
// DB.mu must be locked when calling.
|
||
func (d *DB) updateTableStatsLocked(newFiles []manifest.NewFileEntry) {
|
||
var needStats bool
|
||
for _, nf := range newFiles {
|
||
if !nf.Meta.StatsValid() {
|
||
needStats = true
|
||
break
|
||
}
|
||
}
|
||
if !needStats {
|
||
return
|
||
}
|
||
|
||
d.mu.tableStats.pending = append(d.mu.tableStats.pending, newFiles...)
|
||
d.maybeCollectTableStatsLocked()
|
||
}
|
||
|
||
func (d *DB) shouldCollectTableStatsLocked() bool {
|
||
return !d.mu.tableStats.loading &&
|
||
d.closed.Load() == nil &&
|
||
!d.opts.private.disableTableStats &&
|
||
(len(d.mu.tableStats.pending) > 0 || !d.mu.tableStats.loadedInitial)
|
||
}
|
||
|
||
// collectTableStats runs a table stats collection job, returning true if the
|
||
// invocation did the collection work, false otherwise (e.g. if another job was
|
||
// already running).
|
||
func (d *DB) collectTableStats() bool {
|
||
const maxTableStatsPerScan = 50
|
||
|
||
d.mu.Lock()
|
||
if !d.shouldCollectTableStatsLocked() {
|
||
d.mu.Unlock()
|
||
return false
|
||
}
|
||
|
||
pending := d.mu.tableStats.pending
|
||
d.mu.tableStats.pending = nil
|
||
d.mu.tableStats.loading = true
|
||
jobID := d.mu.nextJobID
|
||
d.mu.nextJobID++
|
||
loadedInitial := d.mu.tableStats.loadedInitial
|
||
// Drop DB.mu before performing IO.
|
||
d.mu.Unlock()
|
||
|
||
// Every run of collectTableStats either collects stats from the pending
|
||
// list (if non-empty) or from scanning the version (loadedInitial is
|
||
// false). This job only runs if at least one of those conditions holds.
|
||
|
||
// Grab a read state to scan for tables.
|
||
rs := d.loadReadState()
|
||
var collected []collectedStats
|
||
var hints []deleteCompactionHint
|
||
if len(pending) > 0 {
|
||
collected, hints = d.loadNewFileStats(rs, pending)
|
||
} else {
|
||
var moreRemain bool
|
||
var buf [maxTableStatsPerScan]collectedStats
|
||
collected, hints, moreRemain = d.scanReadStateTableStats(rs, buf[:0])
|
||
loadedInitial = !moreRemain
|
||
}
|
||
rs.unref()
|
||
|
||
// Update the FileMetadata with the loaded stats while holding d.mu.
|
||
d.mu.Lock()
|
||
defer d.mu.Unlock()
|
||
d.mu.tableStats.loading = false
|
||
if loadedInitial && !d.mu.tableStats.loadedInitial {
|
||
d.mu.tableStats.loadedInitial = loadedInitial
|
||
d.opts.EventListener.TableStatsLoaded(TableStatsInfo{
|
||
JobID: jobID,
|
||
})
|
||
}
|
||
|
||
maybeCompact := false
|
||
for _, c := range collected {
|
||
c.fileMetadata.Stats = c.TableStats
|
||
maybeCompact = maybeCompact || fileCompensation(c.fileMetadata) > 0
|
||
c.fileMetadata.StatsMarkValid()
|
||
}
|
||
d.mu.tableStats.cond.Broadcast()
|
||
d.maybeCollectTableStatsLocked()
|
||
if len(hints) > 0 && !d.opts.private.disableDeleteOnlyCompactions {
|
||
// Verify that all of the hint tombstones' files still exist in the
|
||
// current version. Otherwise, the tombstone itself may have been
|
||
// compacted into L6 and more recent keys may have had their sequence
|
||
// numbers zeroed.
|
||
//
|
||
// Note that it's possible that the tombstone file is being compacted
|
||
// presently. In that case, the file will be present in v. When the
|
||
// compaction finishes compacting the tombstone file, it will detect
|
||
// and clear the hint.
|
||
//
|
||
// See DB.maybeUpdateDeleteCompactionHints.
|
||
v := d.mu.versions.currentVersion()
|
||
keepHints := hints[:0]
|
||
for _, h := range hints {
|
||
if v.Contains(h.tombstoneLevel, d.cmp, h.tombstoneFile) {
|
||
keepHints = append(keepHints, h)
|
||
}
|
||
}
|
||
d.mu.compact.deletionHints = append(d.mu.compact.deletionHints, keepHints...)
|
||
}
|
||
if maybeCompact {
|
||
d.maybeScheduleCompaction()
|
||
}
|
||
return true
|
||
}
|
||
|
||
type collectedStats struct {
|
||
*fileMetadata
|
||
manifest.TableStats
|
||
}
|
||
|
||
func (d *DB) loadNewFileStats(
|
||
rs *readState, pending []manifest.NewFileEntry,
|
||
) ([]collectedStats, []deleteCompactionHint) {
|
||
var hints []deleteCompactionHint
|
||
collected := make([]collectedStats, 0, len(pending))
|
||
for _, nf := range pending {
|
||
// A file's stats might have been populated by an earlier call to
|
||
// loadNewFileStats if the file was moved.
|
||
// NB: We're not holding d.mu which protects f.Stats, but only
|
||
// collectTableStats updates f.Stats for active files, and we
|
||
// ensure only one goroutine runs it at a time through
|
||
// d.mu.tableStats.loading.
|
||
if nf.Meta.StatsValid() {
|
||
continue
|
||
}
|
||
|
||
// The file isn't guaranteed to still be live in the readState's
|
||
// version. It may have been deleted or moved. Skip it if it's not in
|
||
// the expected level.
|
||
if !rs.current.Contains(nf.Level, d.cmp, nf.Meta) {
|
||
continue
|
||
}
|
||
|
||
stats, newHints, err := d.loadTableStats(
|
||
rs.current, nf.Level,
|
||
nf.Meta,
|
||
)
|
||
if err != nil {
|
||
d.opts.EventListener.BackgroundError(err)
|
||
continue
|
||
}
|
||
// NB: We don't update the FileMetadata yet, because we aren't
|
||
// holding DB.mu. We'll copy it to the FileMetadata after we're
|
||
// finished with IO.
|
||
collected = append(collected, collectedStats{
|
||
fileMetadata: nf.Meta,
|
||
TableStats: stats,
|
||
})
|
||
hints = append(hints, newHints...)
|
||
}
|
||
return collected, hints
|
||
}
|
||
|
||
// scanReadStateTableStats is run by an active stat collection job when there
|
||
// are no pending new files, but there might be files that existed at Open for
|
||
// which we haven't loaded table stats.
|
||
func (d *DB) scanReadStateTableStats(
|
||
rs *readState, fill []collectedStats,
|
||
) ([]collectedStats, []deleteCompactionHint, bool) {
|
||
moreRemain := false
|
||
var hints []deleteCompactionHint
|
||
sizesChecked := make(map[base.DiskFileNum]struct{})
|
||
for l, levelMetadata := range rs.current.Levels {
|
||
iter := levelMetadata.Iter()
|
||
for f := iter.First(); f != nil; f = iter.Next() {
|
||
// NB: We're not holding d.mu which protects f.Stats, but only the
|
||
// active stats collection job updates f.Stats for active files,
|
||
// and we ensure only one goroutine runs it at a time through
|
||
// d.mu.tableStats.loading. This makes it safe to read validity
|
||
// through f.Stats.ValidLocked despite not holding d.mu.
|
||
if f.StatsValid() {
|
||
continue
|
||
}
|
||
|
||
// Limit how much work we do per read state. The older the read
|
||
// state is, the higher the likelihood files are no longer being
|
||
// used in the current version. If we've exhausted our allowance,
|
||
// return true for the last return value to signal there's more
|
||
// work to do.
|
||
if len(fill) == cap(fill) {
|
||
moreRemain = true
|
||
return fill, hints, moreRemain
|
||
}
|
||
|
||
// If the file is remote and not SharedForeign, we should check if its size
|
||
// matches. This is because checkConsistency skips over remote files.
|
||
//
|
||
// SharedForeign and External files are skipped as their sizes are allowed
|
||
// to have a mismatch; the size stored in the FileBacking is just the part
|
||
// of the file that is referenced by this Pebble instance, not the size of
|
||
// the whole object.
|
||
objMeta, err := d.objProvider.Lookup(fileTypeTable, f.FileBacking.DiskFileNum)
|
||
if err != nil {
|
||
// Set `moreRemain` so we'll try again.
|
||
moreRemain = true
|
||
d.opts.EventListener.BackgroundError(err)
|
||
continue
|
||
}
|
||
|
||
shouldCheckSize := objMeta.IsRemote() &&
|
||
!d.objProvider.IsSharedForeign(objMeta) &&
|
||
!objMeta.IsExternal()
|
||
if _, ok := sizesChecked[f.FileBacking.DiskFileNum]; !ok && shouldCheckSize {
|
||
size, err := d.objProvider.Size(objMeta)
|
||
fileSize := f.FileBacking.Size
|
||
if err != nil {
|
||
moreRemain = true
|
||
d.opts.EventListener.BackgroundError(err)
|
||
continue
|
||
}
|
||
if size != int64(fileSize) {
|
||
err := errors.Errorf(
|
||
"during consistency check in loadTableStats: L%d: %s: object size mismatch (%s): %d (provider) != %d (MANIFEST)",
|
||
errors.Safe(l), f.FileNum, d.objProvider.Path(objMeta),
|
||
errors.Safe(size), errors.Safe(fileSize))
|
||
d.opts.EventListener.BackgroundError(err)
|
||
d.opts.Logger.Fatalf("%s", err)
|
||
}
|
||
|
||
sizesChecked[f.FileBacking.DiskFileNum] = struct{}{}
|
||
}
|
||
|
||
stats, newHints, err := d.loadTableStats(
|
||
rs.current, l, f,
|
||
)
|
||
if err != nil {
|
||
// Set `moreRemain` so we'll try again.
|
||
moreRemain = true
|
||
d.opts.EventListener.BackgroundError(err)
|
||
continue
|
||
}
|
||
fill = append(fill, collectedStats{
|
||
fileMetadata: f,
|
||
TableStats: stats,
|
||
})
|
||
hints = append(hints, newHints...)
|
||
}
|
||
}
|
||
return fill, hints, moreRemain
|
||
}
|
||
|
||
func (d *DB) loadTableStats(
|
||
v *version, level int, meta *fileMetadata,
|
||
) (manifest.TableStats, []deleteCompactionHint, error) {
|
||
var stats manifest.TableStats
|
||
var compactionHints []deleteCompactionHint
|
||
err := d.tableCache.withCommonReader(
|
||
meta, func(r sstable.CommonReader) (err error) {
|
||
props := r.CommonProperties()
|
||
stats.NumEntries = props.NumEntries
|
||
stats.NumDeletions = props.NumDeletions
|
||
if props.NumPointDeletions() > 0 {
|
||
if err = d.loadTablePointKeyStats(props, v, level, meta, &stats); err != nil {
|
||
return
|
||
}
|
||
}
|
||
if props.NumRangeDeletions > 0 || props.NumRangeKeyDels > 0 {
|
||
if compactionHints, err = d.loadTableRangeDelStats(
|
||
r, v, level, meta, &stats,
|
||
); err != nil {
|
||
return
|
||
}
|
||
}
|
||
// TODO(travers): Once we have real-world data, consider collecting
|
||
// additional stats that may provide improved heuristics for compaction
|
||
// picking.
|
||
stats.NumRangeKeySets = props.NumRangeKeySets
|
||
stats.ValueBlocksSize = props.ValueBlocksSize
|
||
return
|
||
})
|
||
if err != nil {
|
||
return stats, nil, err
|
||
}
|
||
return stats, compactionHints, nil
|
||
}
|
||
|
||
// loadTablePointKeyStats calculates the point key statistics for the given
|
||
// table. The provided manifest.TableStats are updated.
|
||
func (d *DB) loadTablePointKeyStats(
|
||
props *sstable.CommonProperties,
|
||
v *version,
|
||
level int,
|
||
meta *fileMetadata,
|
||
stats *manifest.TableStats,
|
||
) error {
|
||
// TODO(jackson): If the file has a wide keyspace, the average
|
||
// value size beneath the entire file might not be representative
|
||
// of the size of the keys beneath the point tombstones.
|
||
// We could write the ranges of 'clusters' of point tombstones to
|
||
// a sstable property and call averageValueSizeBeneath for each of
|
||
// these narrower ranges to improve the estimate.
|
||
avgValLogicalSize, compressionRatio, err := d.estimateSizesBeneath(v, level, meta, props)
|
||
if err != nil {
|
||
return err
|
||
}
|
||
stats.PointDeletionsBytesEstimate =
|
||
pointDeletionsBytesEstimate(meta.Size, props, avgValLogicalSize, compressionRatio)
|
||
return nil
|
||
}
|
||
|
||
// loadTableRangeDelStats calculates the range deletion and range key deletion
|
||
// statistics for the given table.
|
||
func (d *DB) loadTableRangeDelStats(
|
||
r sstable.CommonReader, v *version, level int, meta *fileMetadata, stats *manifest.TableStats,
|
||
) ([]deleteCompactionHint, error) {
|
||
iter, err := newCombinedDeletionKeyspanIter(d.opts.Comparer, r, meta)
|
||
if err != nil {
|
||
return nil, err
|
||
}
|
||
defer iter.Close()
|
||
var compactionHints []deleteCompactionHint
|
||
// We iterate over the defragmented range tombstones and range key deletions,
|
||
// which ensures we don't double count ranges deleted at different sequence
|
||
// numbers. Also, merging abutting tombstones reduces the number of calls to
|
||
// estimateReclaimedSizeBeneath which is costly, and improves the accuracy of
|
||
// our overall estimate.
|
||
for s := iter.First(); s != nil; s = iter.Next() {
|
||
start, end := s.Start, s.End
|
||
// We only need to consider deletion size estimates for tables that contain
|
||
// RANGEDELs.
|
||
var maxRangeDeleteSeqNum uint64
|
||
for _, k := range s.Keys {
|
||
if k.Kind() == base.InternalKeyKindRangeDelete && maxRangeDeleteSeqNum < k.SeqNum() {
|
||
maxRangeDeleteSeqNum = k.SeqNum()
|
||
break
|
||
}
|
||
}
|
||
|
||
// If the file is in the last level of the LSM, there is no data beneath
|
||
// it. The fact that there is still a range tombstone in a bottommost file
|
||
// indicates two possibilites:
|
||
// 1. an open snapshot kept the tombstone around, and the data the
|
||
// tombstone deletes is contained within the file itself.
|
||
// 2. the file was ingested.
|
||
// In the first case, we'd like to estimate disk usage within the file
|
||
// itself since compacting the file will drop that covered data. In the
|
||
// second case, we expect that compacting the file will NOT drop any
|
||
// data and rewriting the file is a waste of write bandwidth. We can
|
||
// distinguish these cases by looking at the file metadata's sequence
|
||
// numbers. A file's range deletions can only delete data within the
|
||
// file at lower sequence numbers. All keys in an ingested sstable adopt
|
||
// the same sequence number, preventing tombstones from deleting keys
|
||
// within the same file. We check here if the largest RANGEDEL sequence
|
||
// number is greater than the file's smallest sequence number. If it is,
|
||
// the RANGEDEL could conceivably (although inconclusively) delete data
|
||
// within the same file.
|
||
//
|
||
// Note that this heuristic is imperfect. If a table containing a range
|
||
// deletion is ingested into L5 and subsequently compacted into L6 but
|
||
// an open snapshot prevents elision of covered keys in L6, the
|
||
// resulting RangeDeletionsBytesEstimate will incorrectly include all
|
||
// covered keys.
|
||
//
|
||
// TODO(jackson): We could prevent the above error in the heuristic by
|
||
// computing the file's RangeDeletionsBytesEstimate during the
|
||
// compaction itself. It's unclear how common this is.
|
||
//
|
||
// NOTE: If the span `s` wholly contains a table containing range keys,
|
||
// the returned size estimate will be slightly inflated by the range key
|
||
// block. However, in practice, range keys are expected to be rare, and
|
||
// the size of the range key block relative to the overall size of the
|
||
// table is expected to be small.
|
||
if level == numLevels-1 && meta.SmallestSeqNum < maxRangeDeleteSeqNum {
|
||
size, err := r.EstimateDiskUsage(start, end)
|
||
if err != nil {
|
||
return nil, err
|
||
}
|
||
stats.RangeDeletionsBytesEstimate += size
|
||
|
||
// As the file is in the bottommost level, there is no need to collect a
|
||
// deletion hint.
|
||
continue
|
||
}
|
||
|
||
// While the size estimates for point keys should only be updated if this
|
||
// span contains a range del, the sequence numbers are required for the
|
||
// hint. Unconditionally descend, but conditionally update the estimates.
|
||
hintType := compactionHintFromKeys(s.Keys)
|
||
estimate, hintSeqNum, err := d.estimateReclaimedSizeBeneath(v, level, start, end, hintType)
|
||
if err != nil {
|
||
return nil, err
|
||
}
|
||
stats.RangeDeletionsBytesEstimate += estimate
|
||
|
||
// If any files were completely contained with the range,
|
||
// hintSeqNum is the smallest sequence number contained in any
|
||
// such file.
|
||
if hintSeqNum == math.MaxUint64 {
|
||
continue
|
||
}
|
||
hint := deleteCompactionHint{
|
||
hintType: hintType,
|
||
start: make([]byte, len(start)),
|
||
end: make([]byte, len(end)),
|
||
tombstoneFile: meta,
|
||
tombstoneLevel: level,
|
||
tombstoneLargestSeqNum: s.LargestSeqNum(),
|
||
tombstoneSmallestSeqNum: s.SmallestSeqNum(),
|
||
fileSmallestSeqNum: hintSeqNum,
|
||
}
|
||
copy(hint.start, start)
|
||
copy(hint.end, end)
|
||
compactionHints = append(compactionHints, hint)
|
||
}
|
||
return compactionHints, err
|
||
}
|
||
|
||
func (d *DB) estimateSizesBeneath(
|
||
v *version, level int, meta *fileMetadata, fileProps *sstable.CommonProperties,
|
||
) (avgValueLogicalSize, compressionRatio float64, err error) {
|
||
// Find all files in lower levels that overlap with meta,
|
||
// summing their value sizes and entry counts.
|
||
file := meta
|
||
var fileSum, keySum, valSum, entryCount uint64
|
||
// Include the file itself. This is important because in some instances, the
|
||
// computed compression ratio is applied to the tombstones contained within
|
||
// `meta` itself. If there are no files beneath `meta` in the LSM, we would
|
||
// calculate a compression ratio of 0 which is not accurate for the file's
|
||
// own tombstones.
|
||
fileSum += file.Size
|
||
entryCount += fileProps.NumEntries
|
||
keySum += fileProps.RawKeySize
|
||
valSum += fileProps.RawValueSize
|
||
|
||
addPhysicalTableStats := func(r *sstable.Reader) (err error) {
|
||
fileSum += file.Size
|
||
entryCount += r.Properties.NumEntries
|
||
keySum += r.Properties.RawKeySize
|
||
valSum += r.Properties.RawValueSize
|
||
return nil
|
||
}
|
||
addVirtualTableStats := func(v sstable.VirtualReader) (err error) {
|
||
fileSum += file.Size
|
||
entryCount += file.Stats.NumEntries
|
||
keySum += v.Properties.RawKeySize
|
||
valSum += v.Properties.RawValueSize
|
||
return nil
|
||
}
|
||
|
||
for l := level + 1; l < numLevels; l++ {
|
||
overlaps := v.Overlaps(l, d.cmp, meta.Smallest.UserKey,
|
||
meta.Largest.UserKey, meta.Largest.IsExclusiveSentinel())
|
||
iter := overlaps.Iter()
|
||
for file = iter.First(); file != nil; file = iter.Next() {
|
||
var err error
|
||
if file.Virtual {
|
||
err = d.tableCache.withVirtualReader(file.VirtualMeta(), addVirtualTableStats)
|
||
} else {
|
||
err = d.tableCache.withReader(file.PhysicalMeta(), addPhysicalTableStats)
|
||
}
|
||
if err != nil {
|
||
return 0, 0, err
|
||
}
|
||
}
|
||
}
|
||
if entryCount == 0 {
|
||
return 0, 0, nil
|
||
}
|
||
// RawKeySize and RawValueSize are uncompressed totals. We'll need to scale
|
||
// the value sum according to the data size to account for compression,
|
||
// index blocks and metadata overhead. Eg:
|
||
//
|
||
// Compression rate × Average uncompressed value size
|
||
//
|
||
// ↓
|
||
//
|
||
// FileSize RawValueSize
|
||
// ----------------------- × ------------
|
||
// RawKeySize+RawValueSize NumEntries
|
||
//
|
||
// We return the average logical value size plus the compression ratio,
|
||
// leaving the scaling to the caller. This allows the caller to perform
|
||
// additional compression ratio scaling if necessary.
|
||
uncompressedSum := float64(keySum + valSum)
|
||
compressionRatio = float64(fileSum) / uncompressedSum
|
||
avgValueLogicalSize = (float64(valSum) / float64(entryCount))
|
||
return avgValueLogicalSize, compressionRatio, nil
|
||
}
|
||
|
||
func (d *DB) estimateReclaimedSizeBeneath(
|
||
v *version, level int, start, end []byte, hintType deleteCompactionHintType,
|
||
) (estimate uint64, hintSeqNum uint64, err error) {
|
||
// Find all files in lower levels that overlap with the deleted range
|
||
// [start, end).
|
||
//
|
||
// An overlapping file might be completely contained by the range
|
||
// tombstone, in which case we can count the entire file size in
|
||
// our estimate without doing any additional I/O.
|
||
//
|
||
// Otherwise, estimating the range for the file requires
|
||
// additional I/O to read the file's index blocks.
|
||
hintSeqNum = math.MaxUint64
|
||
for l := level + 1; l < numLevels; l++ {
|
||
overlaps := v.Overlaps(l, d.cmp, start, end, true /* exclusiveEnd */)
|
||
iter := overlaps.Iter()
|
||
for file := iter.First(); file != nil; file = iter.Next() {
|
||
startCmp := d.cmp(start, file.Smallest.UserKey)
|
||
endCmp := d.cmp(file.Largest.UserKey, end)
|
||
if startCmp <= 0 && (endCmp < 0 || endCmp == 0 && file.Largest.IsExclusiveSentinel()) {
|
||
// The range fully contains the file, so skip looking it up in table
|
||
// cache/looking at its indexes and add the full file size. Whether the
|
||
// disk estimate and hint seqnums are updated depends on a) the type of
|
||
// hint that requested the estimate and b) the keys contained in this
|
||
// current file.
|
||
var updateEstimates, updateHints bool
|
||
switch hintType {
|
||
case deleteCompactionHintTypePointKeyOnly:
|
||
// The range deletion byte estimates should only be updated if this
|
||
// table contains point keys. This ends up being an overestimate in
|
||
// the case that table also has range keys, but such keys are expected
|
||
// to contribute a negligible amount of the table's overall size,
|
||
// relative to point keys.
|
||
if file.HasPointKeys {
|
||
updateEstimates = true
|
||
}
|
||
// As the initiating span contained only range dels, hints can only be
|
||
// updated if this table does _not_ contain range keys.
|
||
if !file.HasRangeKeys {
|
||
updateHints = true
|
||
}
|
||
case deleteCompactionHintTypeRangeKeyOnly:
|
||
// The initiating span contained only range key dels. The estimates
|
||
// apply only to point keys, and are therefore not updated.
|
||
updateEstimates = false
|
||
// As the initiating span contained only range key dels, hints can
|
||
// only be updated if this table does _not_ contain point keys.
|
||
if !file.HasPointKeys {
|
||
updateHints = true
|
||
}
|
||
case deleteCompactionHintTypePointAndRangeKey:
|
||
// Always update the estimates and hints, as this hint type can drop a
|
||
// file, irrespective of the mixture of keys. Similar to above, the
|
||
// range del bytes estimates is an overestimate.
|
||
updateEstimates, updateHints = true, true
|
||
default:
|
||
panic(fmt.Sprintf("pebble: unknown hint type %s", hintType))
|
||
}
|
||
if updateEstimates {
|
||
estimate += file.Size
|
||
}
|
||
if updateHints && hintSeqNum > file.SmallestSeqNum {
|
||
hintSeqNum = file.SmallestSeqNum
|
||
}
|
||
} else if d.cmp(file.Smallest.UserKey, end) <= 0 && d.cmp(start, file.Largest.UserKey) <= 0 {
|
||
// Partial overlap.
|
||
if hintType == deleteCompactionHintTypeRangeKeyOnly {
|
||
// If the hint that generated this overlap contains only range keys,
|
||
// there is no need to calculate disk usage, as the reclaimable space
|
||
// is expected to be minimal relative to point keys.
|
||
continue
|
||
}
|
||
var size uint64
|
||
var err error
|
||
if file.Virtual {
|
||
err = d.tableCache.withVirtualReader(
|
||
file.VirtualMeta(), func(r sstable.VirtualReader) (err error) {
|
||
size, err = r.EstimateDiskUsage(start, end)
|
||
return err
|
||
})
|
||
} else {
|
||
err = d.tableCache.withReader(
|
||
file.PhysicalMeta(), func(r *sstable.Reader) (err error) {
|
||
size, err = r.EstimateDiskUsage(start, end)
|
||
return err
|
||
})
|
||
}
|
||
|
||
if err != nil {
|
||
return 0, hintSeqNum, err
|
||
}
|
||
estimate += size
|
||
}
|
||
}
|
||
}
|
||
return estimate, hintSeqNum, nil
|
||
}
|
||
|
||
func maybeSetStatsFromProperties(meta physicalMeta, props *sstable.Properties) bool {
|
||
// If a table contains range deletions or range key deletions, we defer the
|
||
// stats collection. There are two main reasons for this:
|
||
//
|
||
// 1. Estimating the potential for reclaimed space due to a range deletion
|
||
// tombstone requires scanning the LSM - a potentially expensive operation
|
||
// that should be deferred.
|
||
// 2. Range deletions and / or range key deletions present an opportunity to
|
||
// compute "deletion hints", which also requires a scan of the LSM to
|
||
// compute tables that would be eligible for deletion.
|
||
//
|
||
// These two tasks are deferred to the table stats collector goroutine.
|
||
if props.NumRangeDeletions != 0 || props.NumRangeKeyDels != 0 {
|
||
return false
|
||
}
|
||
|
||
// If a table is more than 10% point deletions without user-provided size
|
||
// estimates, don't calculate the PointDeletionsBytesEstimate statistic
|
||
// using our limited knowledge. The table stats collector can populate the
|
||
// stats and calculate an average of value size of all the tables beneath
|
||
// the table in the LSM, which will be more accurate.
|
||
if unsizedDels := (props.NumDeletions - props.NumSizedDeletions); unsizedDels > props.NumEntries/10 {
|
||
return false
|
||
}
|
||
|
||
var pointEstimate uint64
|
||
if props.NumEntries > 0 {
|
||
// Use the file's own average key and value sizes as an estimate. This
|
||
// doesn't require any additional IO and since the number of point
|
||
// deletions in the file is low, the error introduced by this crude
|
||
// estimate is expected to be small.
|
||
commonProps := &props.CommonProperties
|
||
avgValSize, compressionRatio := estimatePhysicalSizes(meta.Size, commonProps)
|
||
pointEstimate = pointDeletionsBytesEstimate(meta.Size, commonProps, avgValSize, compressionRatio)
|
||
}
|
||
|
||
meta.Stats.NumEntries = props.NumEntries
|
||
meta.Stats.NumDeletions = props.NumDeletions
|
||
meta.Stats.NumRangeKeySets = props.NumRangeKeySets
|
||
meta.Stats.PointDeletionsBytesEstimate = pointEstimate
|
||
meta.Stats.RangeDeletionsBytesEstimate = 0
|
||
meta.Stats.ValueBlocksSize = props.ValueBlocksSize
|
||
meta.StatsMarkValid()
|
||
return true
|
||
}
|
||
|
||
func pointDeletionsBytesEstimate(
|
||
fileSize uint64, props *sstable.CommonProperties, avgValLogicalSize, compressionRatio float64,
|
||
) (estimate uint64) {
|
||
if props.NumEntries == 0 {
|
||
return 0
|
||
}
|
||
numPointDels := props.NumPointDeletions()
|
||
if numPointDels == 0 {
|
||
return 0
|
||
}
|
||
// Estimate the potential space to reclaim using the table's own properties.
|
||
// There may or may not be keys covered by any individual point tombstone.
|
||
// If not, compacting the point tombstone into L6 will at least allow us to
|
||
// drop the point deletion key and will reclaim the tombstone's key bytes.
|
||
// If there are covered key(s), we also get to drop key and value bytes for
|
||
// each covered key.
|
||
//
|
||
// Some point tombstones (DELSIZEDs) carry a user-provided estimate of the
|
||
// uncompressed size of entries that will be elided by fully compacting the
|
||
// tombstone. For these tombstones, there's no guesswork—we use the
|
||
// RawPointTombstoneValueSizeHint property which is the sum of all these
|
||
// tombstones' encoded values.
|
||
//
|
||
// For un-sized point tombstones (DELs), we estimate assuming that each
|
||
// point tombstone on average covers 1 key and using average value sizes.
|
||
// This is almost certainly an overestimate, but that's probably okay
|
||
// because point tombstones can slow range iterations even when they don't
|
||
// cover a key.
|
||
//
|
||
// TODO(jackson): This logic doesn't directly incorporate fixed per-key
|
||
// overhead (8-byte trailer, plus at least 1 byte encoding the length of the
|
||
// key and 1 byte encoding the length of the value). This overhead is
|
||
// indirectly incorporated through the compression ratios, but that results
|
||
// in the overhead being smeared per key-byte and value-byte, rather than
|
||
// per-entry. This per-key fixed overhead can be nontrivial, especially for
|
||
// dense swaths of point tombstones. Give some thought as to whether we
|
||
// should directly include fixed per-key overhead in the calculations.
|
||
|
||
// Below, we calculate the tombstone contributions and the shadowed keys'
|
||
// contributions separately.
|
||
var tombstonesLogicalSize float64
|
||
var shadowedLogicalSize float64
|
||
|
||
// 1. Calculate the contribution of the tombstone keys themselves.
|
||
if props.RawPointTombstoneKeySize > 0 {
|
||
tombstonesLogicalSize += float64(props.RawPointTombstoneKeySize)
|
||
} else {
|
||
// This sstable predates the existence of the RawPointTombstoneKeySize
|
||
// property. We can use the average key size within the file itself and
|
||
// the count of point deletions to estimate the size.
|
||
tombstonesLogicalSize += float64(numPointDels * props.RawKeySize / props.NumEntries)
|
||
}
|
||
|
||
// 2. Calculate the contribution of the keys shadowed by tombstones.
|
||
//
|
||
// 2a. First account for keys shadowed by DELSIZED tombstones. THE DELSIZED
|
||
// tombstones encode the size of both the key and value of the shadowed KV
|
||
// entries. These sizes are aggregated into a sstable property.
|
||
shadowedLogicalSize += float64(props.RawPointTombstoneValueSize)
|
||
|
||
// 2b. Calculate the contribution of the KV entries shadowed by ordinary DEL
|
||
// keys.
|
||
numUnsizedDels := numPointDels - props.NumSizedDeletions
|
||
{
|
||
// The shadowed keys have the same exact user keys as the tombstones
|
||
// themselves, so we can use the `tombstonesLogicalSize` we computed
|
||
// earlier as an estimate. There's a complication that
|
||
// `tombstonesLogicalSize` may include DELSIZED keys we already
|
||
// accounted for.
|
||
shadowedLogicalSize += float64(tombstonesLogicalSize) / float64(numPointDels) * float64(numUnsizedDels)
|
||
|
||
// Calculate the contribution of the deleted values. The caller has
|
||
// already computed an average logical size (possibly computed across
|
||
// many sstables).
|
||
shadowedLogicalSize += float64(numUnsizedDels) * avgValLogicalSize
|
||
}
|
||
|
||
// Scale both tombstone and shadowed totals by logical:physical ratios to
|
||
// account for compression, metadata overhead, etc.
|
||
//
|
||
// Physical FileSize
|
||
// ----------- = -----------------------
|
||
// Logical RawKeySize+RawValueSize
|
||
//
|
||
return uint64((tombstonesLogicalSize + shadowedLogicalSize) * compressionRatio)
|
||
}
|
||
|
||
func estimatePhysicalSizes(
|
||
fileSize uint64, props *sstable.CommonProperties,
|
||
) (avgValLogicalSize, compressionRatio float64) {
|
||
// RawKeySize and RawValueSize are uncompressed totals. Scale according to
|
||
// the data size to account for compression, index blocks and metadata
|
||
// overhead. Eg:
|
||
//
|
||
// Compression rate × Average uncompressed value size
|
||
//
|
||
// ↓
|
||
//
|
||
// FileSize RawValSize
|
||
// ----------------------- × ----------
|
||
// RawKeySize+RawValueSize NumEntries
|
||
//
|
||
uncompressedSum := props.RawKeySize + props.RawValueSize
|
||
compressionRatio = float64(fileSize) / float64(uncompressedSum)
|
||
avgValLogicalSize = (float64(props.RawValueSize) / float64(props.NumEntries))
|
||
return avgValLogicalSize, compressionRatio
|
||
}
|
||
|
||
// newCombinedDeletionKeyspanIter returns a keyspan.FragmentIterator that
|
||
// returns "ranged deletion" spans for a single table, providing a combined view
|
||
// of both range deletion and range key deletion spans. The
|
||
// tableRangedDeletionIter is intended for use in the specific case of computing
|
||
// the statistics and deleteCompactionHints for a single table.
|
||
//
|
||
// As an example, consider the following set of spans from the range deletion
|
||
// and range key blocks of a table:
|
||
//
|
||
// |---------| |---------| |-------| RANGEKEYDELs
|
||
// |-----------|-------------| |-----| RANGEDELs
|
||
// __________________________________________________________
|
||
// a b c d e f g h i j k l m n o p q r s t u v w x y z
|
||
//
|
||
// The tableRangedDeletionIter produces the following set of output spans, where
|
||
// '1' indicates a span containing only range deletions, '2' is a span
|
||
// containing only range key deletions, and '3' is a span containing a mixture
|
||
// of both range deletions and range key deletions.
|
||
//
|
||
// 1 3 1 3 2 1 3 2
|
||
// |-----|---------|-----|---|-----| |---|-|-----|
|
||
// __________________________________________________________
|
||
// a b c d e f g h i j k l m n o p q r s t u v w x y z
|
||
//
|
||
// Algorithm.
|
||
//
|
||
// The iterator first defragments the range deletion and range key blocks
|
||
// separately. During this defragmentation, the range key block is also filtered
|
||
// so that keys other than range key deletes are ignored. The range delete and
|
||
// range key delete keyspaces are then merged.
|
||
//
|
||
// Note that the only fragmentation introduced by merging is from where a range
|
||
// del span overlaps with a range key del span. Within the bounds of any overlap
|
||
// there is guaranteed to be no further fragmentation, as the constituent spans
|
||
// have already been defragmented. To the left and right of any overlap, the
|
||
// same reasoning applies. For example,
|
||
//
|
||
// |--------| |-------| RANGEKEYDEL
|
||
// |---------------------------| RANGEDEL
|
||
// |----1---|----3---|----1----|---2---| Merged, fragmented spans.
|
||
// __________________________________________________________
|
||
// a b c d e f g h i j k l m n o p q r s t u v w x y z
|
||
//
|
||
// Any fragmented abutting spans produced by the merging iter will be of
|
||
// differing types (i.e. a transition from a span with homogenous key kinds to a
|
||
// heterogeneous span, or a transition from a span with exclusively range dels
|
||
// to a span with exclusively range key dels). Therefore, further
|
||
// defragmentation is not required.
|
||
//
|
||
// Each span returned by the tableRangeDeletionIter will have at most four keys,
|
||
// corresponding to the largest and smallest sequence numbers encountered across
|
||
// the range deletes and range keys deletes that comprised the merged spans.
|
||
func newCombinedDeletionKeyspanIter(
|
||
comparer *base.Comparer, cr sstable.CommonReader, m *fileMetadata,
|
||
) (keyspan.FragmentIterator, error) {
|
||
// The range del iter and range key iter are each wrapped in their own
|
||
// defragmenting iter. For each iter, abutting spans can always be merged.
|
||
var equal = keyspan.DefragmentMethodFunc(func(_ base.Equal, a, b *keyspan.Span) bool { return true })
|
||
// Reduce keys by maintaining a slice of at most length two, corresponding to
|
||
// the largest and smallest keys in the defragmented span. This maintains the
|
||
// contract that the emitted slice is sorted by (SeqNum, Kind) descending.
|
||
reducer := func(current, incoming []keyspan.Key) []keyspan.Key {
|
||
if len(current) == 0 && len(incoming) == 0 {
|
||
// While this should never occur in practice, a defensive return is used
|
||
// here to preserve correctness.
|
||
return current
|
||
}
|
||
var largest, smallest keyspan.Key
|
||
var set bool
|
||
for _, keys := range [2][]keyspan.Key{current, incoming} {
|
||
if len(keys) == 0 {
|
||
continue
|
||
}
|
||
first, last := keys[0], keys[len(keys)-1]
|
||
if !set {
|
||
largest, smallest = first, last
|
||
set = true
|
||
continue
|
||
}
|
||
if first.Trailer > largest.Trailer {
|
||
largest = first
|
||
}
|
||
if last.Trailer < smallest.Trailer {
|
||
smallest = last
|
||
}
|
||
}
|
||
if largest.Equal(comparer.Equal, smallest) {
|
||
current = append(current[:0], largest)
|
||
} else {
|
||
current = append(current[:0], largest, smallest)
|
||
}
|
||
return current
|
||
}
|
||
|
||
// The separate iters for the range dels and range keys are wrapped in a
|
||
// merging iter to join the keyspaces into a single keyspace. The separate
|
||
// iters are only added if the particular key kind is present.
|
||
mIter := &keyspan.MergingIter{}
|
||
var transform = keyspan.TransformerFunc(func(cmp base.Compare, in keyspan.Span, out *keyspan.Span) error {
|
||
if in.KeysOrder != keyspan.ByTrailerDesc {
|
||
panic("pebble: combined deletion iter encountered keys in non-trailer descending order")
|
||
}
|
||
out.Start, out.End = in.Start, in.End
|
||
out.Keys = append(out.Keys[:0], in.Keys...)
|
||
out.KeysOrder = keyspan.ByTrailerDesc
|
||
// NB: The order of by-trailer descending may have been violated,
|
||
// because we've layered rangekey and rangedel iterators from the same
|
||
// sstable into the same keyspan.MergingIter. The MergingIter will
|
||
// return the keys in the order that the child iterators were provided.
|
||
// Sort the keys to ensure they're sorted by trailer descending.
|
||
keyspan.SortKeysByTrailer(&out.Keys)
|
||
return nil
|
||
})
|
||
mIter.Init(comparer.Compare, transform, new(keyspan.MergingBuffers))
|
||
|
||
iter, err := cr.NewRawRangeDelIter()
|
||
if err != nil {
|
||
return nil, err
|
||
}
|
||
if iter != nil {
|
||
dIter := &keyspan.DefragmentingIter{}
|
||
dIter.Init(comparer, iter, equal, reducer, new(keyspan.DefragmentingBuffers))
|
||
iter = dIter
|
||
// Truncate tombstones to the containing file's bounds if necessary.
|
||
// See docs/range_deletions.md for why this is necessary.
|
||
iter = keyspan.Truncate(
|
||
comparer.Compare, iter, m.Smallest.UserKey, m.Largest.UserKey,
|
||
nil, nil, false, /* panicOnUpperTruncate */
|
||
)
|
||
mIter.AddLevel(iter)
|
||
}
|
||
|
||
iter, err = cr.NewRawRangeKeyIter()
|
||
if err != nil {
|
||
return nil, err
|
||
}
|
||
if iter != nil {
|
||
// Wrap the range key iterator in a filter that elides keys other than range
|
||
// key deletions.
|
||
iter = keyspan.Filter(iter, func(in *keyspan.Span, out *keyspan.Span) (keep bool) {
|
||
out.Start, out.End = in.Start, in.End
|
||
out.Keys = out.Keys[:0]
|
||
for _, k := range in.Keys {
|
||
if k.Kind() != base.InternalKeyKindRangeKeyDelete {
|
||
continue
|
||
}
|
||
out.Keys = append(out.Keys, k)
|
||
}
|
||
return len(out.Keys) > 0
|
||
}, comparer.Compare)
|
||
dIter := &keyspan.DefragmentingIter{}
|
||
dIter.Init(comparer, iter, equal, reducer, new(keyspan.DefragmentingBuffers))
|
||
iter = dIter
|
||
mIter.AddLevel(iter)
|
||
}
|
||
|
||
return mIter, nil
|
||
}
|
||
|
||
// rangeKeySetsAnnotator implements manifest.Annotator, annotating B-Tree nodes
|
||
// with the sum of the files' counts of range key fragments. Its annotation type
|
||
// is a *uint64. The count of range key sets may change once a table's stats are
|
||
// loaded asynchronously, so its values are marked as cacheable only if a file's
|
||
// stats have been loaded.
|
||
type rangeKeySetsAnnotator struct{}
|
||
|
||
var _ manifest.Annotator = rangeKeySetsAnnotator{}
|
||
|
||
func (a rangeKeySetsAnnotator) Zero(dst interface{}) interface{} {
|
||
if dst == nil {
|
||
return new(uint64)
|
||
}
|
||
v := dst.(*uint64)
|
||
*v = 0
|
||
return v
|
||
}
|
||
|
||
func (a rangeKeySetsAnnotator) Accumulate(
|
||
f *fileMetadata, dst interface{},
|
||
) (v interface{}, cacheOK bool) {
|
||
vptr := dst.(*uint64)
|
||
*vptr = *vptr + f.Stats.NumRangeKeySets
|
||
return vptr, f.StatsValid()
|
||
}
|
||
|
||
func (a rangeKeySetsAnnotator) Merge(src interface{}, dst interface{}) interface{} {
|
||
srcV := src.(*uint64)
|
||
dstV := dst.(*uint64)
|
||
*dstV = *dstV + *srcV
|
||
return dstV
|
||
}
|
||
|
||
// countRangeKeySetFragments counts the number of RANGEKEYSET keys across all
|
||
// files of the LSM. It only counts keys in files for which table stats have
|
||
// been loaded. It uses a b-tree annotator to cache intermediate values between
|
||
// calculations when possible.
|
||
func countRangeKeySetFragments(v *version) (count uint64) {
|
||
for l := 0; l < numLevels; l++ {
|
||
if v.RangeKeyLevels[l].Empty() {
|
||
continue
|
||
}
|
||
count += *v.RangeKeyLevels[l].Annotation(rangeKeySetsAnnotator{}).(*uint64)
|
||
}
|
||
return count
|
||
}
|
||
|
||
// tombstonesAnnotator implements manifest.Annotator, annotating B-Tree nodes
|
||
// with the sum of the files' counts of tombstones (DEL, SINGLEDEL and RANGEDELk
|
||
// eys). Its annotation type is a *uint64. The count of tombstones may change
|
||
// once a table's stats are loaded asynchronously, so its values are marked as
|
||
// cacheable only if a file's stats have been loaded.
|
||
type tombstonesAnnotator struct{}
|
||
|
||
var _ manifest.Annotator = tombstonesAnnotator{}
|
||
|
||
func (a tombstonesAnnotator) Zero(dst interface{}) interface{} {
|
||
if dst == nil {
|
||
return new(uint64)
|
||
}
|
||
v := dst.(*uint64)
|
||
*v = 0
|
||
return v
|
||
}
|
||
|
||
func (a tombstonesAnnotator) Accumulate(
|
||
f *fileMetadata, dst interface{},
|
||
) (v interface{}, cacheOK bool) {
|
||
vptr := dst.(*uint64)
|
||
*vptr = *vptr + f.Stats.NumDeletions
|
||
return vptr, f.StatsValid()
|
||
}
|
||
|
||
func (a tombstonesAnnotator) Merge(src interface{}, dst interface{}) interface{} {
|
||
srcV := src.(*uint64)
|
||
dstV := dst.(*uint64)
|
||
*dstV = *dstV + *srcV
|
||
return dstV
|
||
}
|
||
|
||
// countTombstones counts the number of tombstone (DEL, SINGLEDEL and RANGEDEL)
|
||
// internal keys across all files of the LSM. It only counts keys in files for
|
||
// which table stats have been loaded. It uses a b-tree annotator to cache
|
||
// intermediate values between calculations when possible.
|
||
func countTombstones(v *version) (count uint64) {
|
||
for l := 0; l < numLevels; l++ {
|
||
if v.Levels[l].Empty() {
|
||
continue
|
||
}
|
||
count += *v.Levels[l].Annotation(tombstonesAnnotator{}).(*uint64)
|
||
}
|
||
return count
|
||
}
|
||
|
||
// valueBlocksSizeAnnotator implements manifest.Annotator, annotating B-Tree
|
||
// nodes with the sum of the files' Properties.ValueBlocksSize. Its annotation
|
||
// type is a *uint64. The value block size may change once a table's stats are
|
||
// loaded asynchronously, so its values are marked as cacheable only if a
|
||
// file's stats have been loaded.
|
||
type valueBlocksSizeAnnotator struct{}
|
||
|
||
var _ manifest.Annotator = valueBlocksSizeAnnotator{}
|
||
|
||
func (a valueBlocksSizeAnnotator) Zero(dst interface{}) interface{} {
|
||
if dst == nil {
|
||
return new(uint64)
|
||
}
|
||
v := dst.(*uint64)
|
||
*v = 0
|
||
return v
|
||
}
|
||
|
||
func (a valueBlocksSizeAnnotator) Accumulate(
|
||
f *fileMetadata, dst interface{},
|
||
) (v interface{}, cacheOK bool) {
|
||
vptr := dst.(*uint64)
|
||
*vptr = *vptr + f.Stats.ValueBlocksSize
|
||
return vptr, f.StatsValid()
|
||
}
|
||
|
||
func (a valueBlocksSizeAnnotator) Merge(src interface{}, dst interface{}) interface{} {
|
||
srcV := src.(*uint64)
|
||
dstV := dst.(*uint64)
|
||
*dstV = *dstV + *srcV
|
||
return dstV
|
||
}
|
||
|
||
// valueBlocksSizeForLevel returns the Properties.ValueBlocksSize across all
|
||
// files for a level of the LSM. It only includes the size for files for which
|
||
// table stats have been loaded. It uses a b-tree annotator to cache
|
||
// intermediate values between calculations when possible. It must not be
|
||
// called concurrently.
|
||
//
|
||
// REQUIRES: 0 <= level <= numLevels.
|
||
func valueBlocksSizeForLevel(v *version, level int) (count uint64) {
|
||
if v.Levels[level].Empty() {
|
||
return 0
|
||
}
|
||
return *v.Levels[level].Annotation(valueBlocksSizeAnnotator{}).(*uint64)
|
||
}
|