mirror of
https://source.quilibrium.com/quilibrium/ceremonyclient.git
synced 2024-12-25 08:05:17 +00:00
391 lines
8.8 KiB
Go
391 lines
8.8 KiB
Go
package metamorphic
|
|
|
|
import (
|
|
"bytes"
|
|
"fmt"
|
|
"io"
|
|
"strings"
|
|
"testing"
|
|
|
|
"github.com/cockroachdb/datadriven"
|
|
"github.com/cockroachdb/pebble/internal/randvar"
|
|
"github.com/stretchr/testify/require"
|
|
)
|
|
|
|
func TestObjKey(t *testing.T) {
|
|
testCases := []struct {
|
|
key objKey
|
|
want string
|
|
}{
|
|
{
|
|
key: makeObjKey(makeObjID(dbTag, 1), []byte("foo")),
|
|
want: "db1:foo",
|
|
},
|
|
{
|
|
key: makeObjKey(makeObjID(batchTag, 1), []byte("bar")),
|
|
want: "batch1:bar",
|
|
},
|
|
}
|
|
|
|
for _, tc := range testCases {
|
|
t.Run("", func(t *testing.T) {
|
|
require.Equal(t, tc.want, tc.key.String())
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestGlobalStateIndicatesEligibleForSingleDelete(t *testing.T) {
|
|
key := makeObjKey(makeObjID(dbTag, 1), []byte("foo"))
|
|
testCases := []struct {
|
|
meta keyMeta
|
|
want bool
|
|
}{
|
|
{
|
|
meta: keyMeta{
|
|
objKey: key,
|
|
},
|
|
want: false,
|
|
},
|
|
{
|
|
meta: keyMeta{
|
|
objKey: key,
|
|
sets: 1,
|
|
},
|
|
want: true,
|
|
},
|
|
{
|
|
meta: keyMeta{
|
|
objKey: key,
|
|
sets: 2,
|
|
},
|
|
want: false,
|
|
},
|
|
{
|
|
meta: keyMeta{
|
|
objKey: key,
|
|
sets: 1,
|
|
merges: 1,
|
|
},
|
|
want: false,
|
|
},
|
|
{
|
|
meta: keyMeta{
|
|
objKey: key,
|
|
sets: 1,
|
|
dels: 1,
|
|
},
|
|
want: false,
|
|
},
|
|
{
|
|
meta: keyMeta{
|
|
objKey: key,
|
|
sets: 1,
|
|
singleDel: true,
|
|
},
|
|
want: false,
|
|
},
|
|
}
|
|
|
|
for _, tc := range testCases {
|
|
k := newKeyManager(1)
|
|
t.Run("", func(t *testing.T) {
|
|
k.globalKeysMap[string(key.key)] = &tc.meta
|
|
require.Equal(t, tc.want, k.globalStateIndicatesEligibleForSingleDelete(key.key))
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestKeyMeta_MergeInto(t *testing.T) {
|
|
testCases := []struct {
|
|
existing keyMeta
|
|
toMerge keyMeta
|
|
expected keyMeta
|
|
}{
|
|
{
|
|
existing: keyMeta{
|
|
sets: 1,
|
|
merges: 0,
|
|
singleDel: false,
|
|
},
|
|
toMerge: keyMeta{
|
|
sets: 0,
|
|
merges: 0,
|
|
singleDel: true,
|
|
},
|
|
expected: keyMeta{
|
|
sets: 1,
|
|
merges: 0,
|
|
singleDel: true,
|
|
updateOps: []keyUpdate{
|
|
{deleted: true, metaTimestamp: 0},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
existing: keyMeta{
|
|
sets: 3,
|
|
merges: 1,
|
|
dels: 7,
|
|
},
|
|
toMerge: keyMeta{
|
|
sets: 4,
|
|
merges: 2,
|
|
dels: 8,
|
|
del: true,
|
|
},
|
|
expected: keyMeta{
|
|
sets: 7,
|
|
merges: 3,
|
|
dels: 15,
|
|
del: true,
|
|
updateOps: []keyUpdate{
|
|
{deleted: true, metaTimestamp: 1},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
existing: keyMeta{
|
|
sets: 3,
|
|
merges: 1,
|
|
dels: 7,
|
|
del: true,
|
|
},
|
|
toMerge: keyMeta{
|
|
sets: 1,
|
|
merges: 0,
|
|
dels: 8,
|
|
del: false,
|
|
},
|
|
expected: keyMeta{
|
|
sets: 4,
|
|
merges: 1,
|
|
dels: 15,
|
|
del: false,
|
|
updateOps: []keyUpdate{
|
|
{deleted: false, metaTimestamp: 2},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
keyManager := newKeyManager(1 /* numInstances */)
|
|
for _, tc := range testCases {
|
|
t.Run("", func(t *testing.T) {
|
|
tc.toMerge.mergeInto(keyManager, &tc.existing)
|
|
require.Equal(t, tc.expected, tc.existing)
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestKeyManager_AddKey(t *testing.T) {
|
|
m := newKeyManager(1 /* numInstances */)
|
|
require.Empty(t, m.globalKeys)
|
|
|
|
k1 := []byte("foo")
|
|
require.True(t, m.addNewKey(k1))
|
|
require.Len(t, m.globalKeys, 1)
|
|
require.Len(t, m.globalKeyPrefixes, 1)
|
|
require.Contains(t, m.globalKeys, k1)
|
|
require.Contains(t, m.globalKeyPrefixes, k1)
|
|
require.False(t, m.addNewKey(k1))
|
|
require.True(t, m.prefixExists([]byte("foo")))
|
|
require.False(t, m.prefixExists([]byte("bar")))
|
|
|
|
k2 := []byte("bar")
|
|
require.True(t, m.addNewKey(k2))
|
|
require.Len(t, m.globalKeys, 2)
|
|
require.Len(t, m.globalKeyPrefixes, 2)
|
|
require.Contains(t, m.globalKeys, k2)
|
|
require.Contains(t, m.globalKeyPrefixes, k2)
|
|
require.True(t, m.prefixExists([]byte("bar")))
|
|
k3 := []byte("bax@4")
|
|
require.True(t, m.addNewKey(k3))
|
|
require.Len(t, m.globalKeys, 3)
|
|
require.Len(t, m.globalKeyPrefixes, 3)
|
|
require.Contains(t, m.globalKeys, k3)
|
|
require.Contains(t, m.globalKeyPrefixes, []byte("bax"))
|
|
require.True(t, m.prefixExists([]byte("bax")))
|
|
k4 := []byte("foo@6")
|
|
require.True(t, m.addNewKey(k4))
|
|
require.Len(t, m.globalKeys, 4)
|
|
require.Len(t, m.globalKeyPrefixes, 3)
|
|
require.Contains(t, m.globalKeys, k4)
|
|
require.True(t, m.prefixExists([]byte("foo")))
|
|
|
|
require.Equal(t, [][]byte{
|
|
[]byte("bar"), []byte("bax"), []byte("foo"),
|
|
}, m.prefixes())
|
|
}
|
|
|
|
func TestKeyManager_GetOrInit(t *testing.T) {
|
|
id := makeObjID(batchTag, 1)
|
|
key := []byte("foo")
|
|
o := makeObjKey(id, key)
|
|
|
|
m := newKeyManager(1 /* numInstances */)
|
|
require.NotContains(t, m.byObjKey, o.String())
|
|
require.NotContains(t, m.byObj, id)
|
|
require.Contains(t, m.byObj, makeObjID(dbTag, 1)) // Always contains the DB key.
|
|
|
|
meta1 := m.getOrInit(id, key)
|
|
require.Contains(t, m.byObjKey, o.String())
|
|
require.Contains(t, m.byObj, id)
|
|
|
|
// Idempotent.
|
|
meta2 := m.getOrInit(id, key)
|
|
require.Equal(t, meta1, meta2)
|
|
}
|
|
|
|
func TestKeyManager_Contains(t *testing.T) {
|
|
id := makeObjID(dbTag, 1)
|
|
key := []byte("foo")
|
|
|
|
m := newKeyManager(1 /* numInstances */)
|
|
require.False(t, m.contains(id, key))
|
|
|
|
m.getOrInit(id, key)
|
|
require.True(t, m.contains(id, key))
|
|
}
|
|
|
|
func TestKeyManager_MergeInto(t *testing.T) {
|
|
fromID := makeObjID(batchTag, 1)
|
|
toID := makeObjID(dbTag, 1)
|
|
|
|
m := newKeyManager(1 /* numInstances */)
|
|
|
|
// Two keys in "from".
|
|
a := m.getOrInit(fromID, []byte("foo"))
|
|
a.sets = 1
|
|
b := m.getOrInit(fromID, []byte("bar"))
|
|
b.merges = 2
|
|
|
|
// One key in "to", with same value as a key in "from", that will be merged.
|
|
m.getOrInit(toID, []byte("foo"))
|
|
|
|
// Before, there are two sets.
|
|
require.Len(t, m.byObj[fromID], 2)
|
|
require.Len(t, m.byObj[toID], 1)
|
|
|
|
m.mergeKeysInto(fromID, toID)
|
|
|
|
// Keys in "from" sets are moved to "to" set.
|
|
require.Len(t, m.byObj[toID], 2)
|
|
|
|
// Key "foo" was merged into "to".
|
|
foo := m.getOrInit(toID, []byte("foo"))
|
|
require.Equal(t, 1, foo.sets) // value was merged.
|
|
|
|
// Key "bar" was merged into "to".
|
|
bar := m.getOrInit(toID, []byte("bar"))
|
|
require.Equal(t, 2, bar.merges) // value was unchanged.
|
|
|
|
// Keys in "from" sets are removed from maps.
|
|
require.NotContains(t, m.byObjKey, makeObjKey(fromID, a.key))
|
|
require.NotContains(t, m.byObjKey, makeObjKey(fromID, b.key))
|
|
require.NotContains(t, m.byObj, fromID)
|
|
}
|
|
|
|
func mustParseObjID(s string) objID {
|
|
id, err := parseObjID(s)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
return id
|
|
}
|
|
|
|
func printKeys(w io.Writer, keys [][]byte) {
|
|
if len(keys) == 0 {
|
|
fmt.Fprintln(w, "(none)")
|
|
return
|
|
}
|
|
for i, key := range keys {
|
|
if i > 0 {
|
|
fmt.Fprint(w, ", ")
|
|
}
|
|
fmt.Fprintf(w, "%q", key)
|
|
}
|
|
fmt.Fprintln(w)
|
|
}
|
|
|
|
func TestKeyManager(t *testing.T) {
|
|
var buf bytes.Buffer
|
|
km := newKeyManager(1 /* numInstances */)
|
|
datadriven.RunTest(t, "testdata/key_manager", func(t *testing.T, td *datadriven.TestData) string {
|
|
switch td.Cmd {
|
|
case "reset":
|
|
km = newKeyManager(1 /* numInstances */)
|
|
return ""
|
|
case "run":
|
|
buf.Reset()
|
|
for _, line := range strings.Split(td.Input, "\n") {
|
|
fields := strings.Fields(line)
|
|
switch fields[0] {
|
|
case "add-new-key":
|
|
if km.addNewKey([]byte(fields[1])) {
|
|
fmt.Fprintf(&buf, "%q is new\n", fields[1])
|
|
} else {
|
|
fmt.Fprintf(&buf, "%q already tracked\n", fields[1])
|
|
}
|
|
case "keys":
|
|
case "read-keys":
|
|
fmt.Fprintf(&buf, "read keys: ")
|
|
printKeys(&buf, km.eligibleReadKeys())
|
|
case "write-keys":
|
|
fmt.Fprintf(&buf, "write keys: ")
|
|
printKeys(&buf, km.eligibleWriteKeys())
|
|
case "singledel-keys":
|
|
fmt.Fprintf(&buf, "singledel keys: ")
|
|
printKeys(&buf, km.eligibleSingleDeleteKeys(
|
|
mustParseObjID(fields[1]), mustParseObjID(fields[2])))
|
|
case "op":
|
|
ops, err := parse([]byte(strings.TrimPrefix(line, "op")), parserOpts{
|
|
allowUndefinedObjs: true,
|
|
})
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
} else if len(ops) != 1 {
|
|
t.Fatalf("expected 1 op but found %d", len(ops))
|
|
}
|
|
km.update(ops[0])
|
|
fmt.Fprintf(&buf, "[%s]\n", ops[0])
|
|
default:
|
|
return fmt.Sprintf("unrecognized subcommand %q", fields[0])
|
|
}
|
|
}
|
|
return buf.String()
|
|
default:
|
|
return fmt.Sprintf("unrecognized command %q", td.Cmd)
|
|
}
|
|
})
|
|
}
|
|
|
|
func TestOpWrittenKeys(t *testing.T) {
|
|
for name, info := range methods {
|
|
t.Run(name, func(t *testing.T) {
|
|
// Any operations that exist in methods but are not handled in
|
|
// opWrittenKeys will result in a panic, failing the subtest.
|
|
opWrittenKeys(info.constructor())
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestLoadPrecedingKeys(t *testing.T) {
|
|
rng := randvar.NewRand()
|
|
cfg := defaultConfig()
|
|
km := newKeyManager(1 /* numInstances */)
|
|
ops := generate(rng, 1000, cfg, km)
|
|
|
|
cfg2 := defaultConfig()
|
|
km2 := newKeyManager(1 /* numInstances */)
|
|
loadPrecedingKeys(t, ops, &cfg2, km2)
|
|
|
|
// NB: We can't assert equality, because the original run may not have
|
|
// ever used the max of the distribution.
|
|
require.Greater(t, cfg2.writeSuffixDist.Max(), uint64(1))
|
|
|
|
// NB: We can't assert equality, because the original run may have generated
|
|
// keys that it didn't end up using in operations.
|
|
require.Subset(t, km.globalKeys, km2.globalKeys)
|
|
require.Subset(t, km.globalKeyPrefixes, km2.globalKeyPrefixes)
|
|
}
|