mirror of
https://source.quilibrium.com/quilibrium/ceremonyclient.git
synced 2024-12-26 16:45:18 +00:00
325 lines
7.7 KiB
Go
325 lines
7.7 KiB
Go
|
package data
|
||
|
|
||
|
import (
|
||
|
"bytes"
|
||
|
"context"
|
||
|
"time"
|
||
|
|
||
|
"golang.org/x/crypto/sha3"
|
||
|
"source.quilibrium.com/quilibrium/monorepo/node/config"
|
||
|
|
||
|
"github.com/pkg/errors"
|
||
|
"go.uber.org/zap"
|
||
|
"google.golang.org/grpc"
|
||
|
"google.golang.org/protobuf/proto"
|
||
|
"source.quilibrium.com/quilibrium/monorepo/node/execution/intrinsics/token/application"
|
||
|
"source.quilibrium.com/quilibrium/monorepo/node/p2p"
|
||
|
"source.quilibrium.com/quilibrium/monorepo/node/protobufs"
|
||
|
)
|
||
|
|
||
|
func (e *DataClockConsensusEngine) prove(
|
||
|
previousFrame *protobufs.ClockFrame,
|
||
|
) (*protobufs.ClockFrame, error) {
|
||
|
e.stagedTransactionsMx.Lock()
|
||
|
executionOutput := &protobufs.IntrinsicExecutionOutput{}
|
||
|
app, err := application.MaterializeApplicationFromFrame(
|
||
|
previousFrame,
|
||
|
e.frameProverTries,
|
||
|
e.coinStore,
|
||
|
e.logger,
|
||
|
)
|
||
|
if err != nil {
|
||
|
e.stagedTransactions = &protobufs.TokenRequests{}
|
||
|
e.stagedTransactionsMx.Unlock()
|
||
|
return nil, errors.Wrap(err, "prove")
|
||
|
}
|
||
|
|
||
|
if e.stagedTransactions == nil {
|
||
|
e.stagedTransactions = &protobufs.TokenRequests{}
|
||
|
}
|
||
|
|
||
|
e.logger.Info(
|
||
|
"proving new frame",
|
||
|
zap.Int("transactions", len(e.stagedTransactions.Requests)),
|
||
|
)
|
||
|
|
||
|
var validTransactions *protobufs.TokenRequests
|
||
|
app, validTransactions, _, err = app.ApplyTransitions(
|
||
|
previousFrame.FrameNumber,
|
||
|
e.stagedTransactions,
|
||
|
true,
|
||
|
)
|
||
|
if err != nil {
|
||
|
e.stagedTransactions = &protobufs.TokenRequests{}
|
||
|
e.stagedTransactionsMx.Unlock()
|
||
|
return nil, errors.Wrap(err, "prove")
|
||
|
}
|
||
|
|
||
|
defer e.stagedTransactionsMx.Unlock()
|
||
|
|
||
|
outputState, err := app.MaterializeStateFromApplication()
|
||
|
if err != nil {
|
||
|
return nil, errors.Wrap(err, "prove")
|
||
|
}
|
||
|
|
||
|
executionOutput.Address = application.TOKEN_ADDRESS
|
||
|
executionOutput.Output, err = proto.Marshal(outputState)
|
||
|
if err != nil {
|
||
|
return nil, errors.Wrap(err, "prove")
|
||
|
}
|
||
|
|
||
|
executionOutput.Proof, err = proto.Marshal(validTransactions)
|
||
|
if err != nil {
|
||
|
return nil, errors.Wrap(err, "prove")
|
||
|
}
|
||
|
|
||
|
data, err := proto.Marshal(executionOutput)
|
||
|
if err != nil {
|
||
|
return nil, errors.Wrap(err, "prove")
|
||
|
}
|
||
|
|
||
|
e.logger.Debug("encoded execution output")
|
||
|
digest := sha3.NewShake256()
|
||
|
_, err = digest.Write(data)
|
||
|
if err != nil {
|
||
|
e.logger.Error(
|
||
|
"error writing digest",
|
||
|
zap.Error(err),
|
||
|
)
|
||
|
return nil, errors.Wrap(err, "prove")
|
||
|
}
|
||
|
|
||
|
expand := make([]byte, 1024)
|
||
|
_, err = digest.Read(expand)
|
||
|
if err != nil {
|
||
|
e.logger.Error(
|
||
|
"error expanding digest",
|
||
|
zap.Error(err),
|
||
|
)
|
||
|
return nil, errors.Wrap(err, "prove")
|
||
|
}
|
||
|
|
||
|
commitment, err := e.inclusionProver.CommitRaw(
|
||
|
expand,
|
||
|
16,
|
||
|
)
|
||
|
if err != nil {
|
||
|
return nil, errors.Wrap(err, "prove")
|
||
|
}
|
||
|
|
||
|
e.logger.Debug("creating kzg proof")
|
||
|
proof, err := e.inclusionProver.ProveRaw(
|
||
|
expand,
|
||
|
int(expand[0]%16),
|
||
|
16,
|
||
|
)
|
||
|
if err != nil {
|
||
|
return nil, errors.Wrap(err, "prove")
|
||
|
}
|
||
|
|
||
|
e.logger.Debug("finalizing execution proof")
|
||
|
|
||
|
frame, err := e.frameProver.ProveDataClockFrame(
|
||
|
previousFrame,
|
||
|
[][]byte{proof},
|
||
|
[]*protobufs.InclusionAggregateProof{
|
||
|
{
|
||
|
Filter: e.filter,
|
||
|
FrameNumber: previousFrame.FrameNumber + 1,
|
||
|
InclusionCommitments: []*protobufs.InclusionCommitment{
|
||
|
{
|
||
|
Filter: e.filter,
|
||
|
FrameNumber: previousFrame.FrameNumber + 1,
|
||
|
TypeUrl: protobufs.IntrinsicExecutionOutputType,
|
||
|
Commitment: commitment,
|
||
|
Data: data,
|
||
|
Position: 0,
|
||
|
},
|
||
|
},
|
||
|
Proof: proof,
|
||
|
},
|
||
|
},
|
||
|
e.provingKey,
|
||
|
time.Now().UnixMilli(),
|
||
|
e.difficulty,
|
||
|
)
|
||
|
if err != nil {
|
||
|
return nil, errors.Wrap(err, "prove")
|
||
|
}
|
||
|
e.logger.Info(
|
||
|
"returning new proven frame",
|
||
|
zap.Uint64("frame_number", frame.FrameNumber),
|
||
|
zap.Int("proof_count", len(frame.AggregateProofs)),
|
||
|
zap.Int("commitment_count", len(frame.Input[516:])/74),
|
||
|
)
|
||
|
return frame, nil
|
||
|
}
|
||
|
|
||
|
func (e *DataClockConsensusEngine) GetMostAheadPeer(
|
||
|
frameNumber uint64,
|
||
|
) (
|
||
|
[]byte,
|
||
|
uint64,
|
||
|
error,
|
||
|
) {
|
||
|
e.logger.Info(
|
||
|
"checking peer list",
|
||
|
zap.Int("peers", len(e.peerMap)),
|
||
|
zap.Int("uncooperative_peers", len(e.uncooperativePeersMap)),
|
||
|
zap.Uint64("current_head_frame", frameNumber),
|
||
|
)
|
||
|
|
||
|
max := frameNumber
|
||
|
var peer []byte = nil
|
||
|
e.peerMapMx.RLock()
|
||
|
for _, v := range e.peerMap {
|
||
|
e.logger.Debug(
|
||
|
"checking peer info",
|
||
|
zap.Binary("peer_id", v.peerId),
|
||
|
zap.Uint64("max_frame_number", v.maxFrame),
|
||
|
zap.Int64("timestamp", v.timestamp),
|
||
|
zap.Binary("version", v.version),
|
||
|
)
|
||
|
_, ok := e.uncooperativePeersMap[string(v.peerId)]
|
||
|
if v.maxFrame > max &&
|
||
|
v.timestamp > config.GetMinimumVersionCutoff().UnixMilli() &&
|
||
|
bytes.Compare(v.version, config.GetMinimumVersion()) >= 0 && !ok {
|
||
|
peer = v.peerId
|
||
|
max = v.maxFrame
|
||
|
}
|
||
|
}
|
||
|
e.peerMapMx.RUnlock()
|
||
|
|
||
|
if peer == nil {
|
||
|
return nil, 0, p2p.ErrNoPeersAvailable
|
||
|
}
|
||
|
|
||
|
return peer, max, nil
|
||
|
}
|
||
|
|
||
|
func (e *DataClockConsensusEngine) sync(
|
||
|
currentLatest *protobufs.ClockFrame,
|
||
|
maxFrame uint64,
|
||
|
peerId []byte,
|
||
|
) (*protobufs.ClockFrame, error) {
|
||
|
latest := currentLatest
|
||
|
e.logger.Info("polling peer for new frames", zap.Binary("peer_id", peerId))
|
||
|
cc, err := e.pubSub.GetDirectChannel(peerId, "")
|
||
|
if err != nil {
|
||
|
e.logger.Debug(
|
||
|
"could not establish direct channel",
|
||
|
zap.Error(err),
|
||
|
)
|
||
|
e.peerMapMx.Lock()
|
||
|
if _, ok := e.peerMap[string(peerId)]; ok {
|
||
|
e.uncooperativePeersMap[string(peerId)] = e.peerMap[string(peerId)]
|
||
|
e.uncooperativePeersMap[string(peerId)].timestamp = time.Now().UnixMilli()
|
||
|
delete(e.peerMap, string(peerId))
|
||
|
}
|
||
|
e.peerMapMx.Unlock()
|
||
|
return latest, errors.Wrap(err, "sync")
|
||
|
}
|
||
|
|
||
|
client := protobufs.NewDataServiceClient(cc)
|
||
|
|
||
|
response, err := client.GetDataFrame(
|
||
|
context.TODO(),
|
||
|
&protobufs.GetDataFrameRequest{
|
||
|
FrameNumber: 0,
|
||
|
},
|
||
|
grpc.MaxCallRecvMsgSize(600*1024*1024),
|
||
|
)
|
||
|
if err != nil {
|
||
|
e.logger.Debug(
|
||
|
"could not get frame",
|
||
|
zap.Error(err),
|
||
|
)
|
||
|
e.peerMapMx.Lock()
|
||
|
if _, ok := e.peerMap[string(peerId)]; ok {
|
||
|
e.uncooperativePeersMap[string(peerId)] = e.peerMap[string(peerId)]
|
||
|
e.uncooperativePeersMap[string(peerId)].timestamp = time.Now().UnixMilli()
|
||
|
delete(e.peerMap, string(peerId))
|
||
|
}
|
||
|
e.peerMapMx.Unlock()
|
||
|
if err := cc.Close(); err != nil {
|
||
|
e.logger.Error("error while closing connection", zap.Error(err))
|
||
|
}
|
||
|
return latest, errors.Wrap(err, "sync")
|
||
|
}
|
||
|
|
||
|
if response == nil {
|
||
|
e.logger.Debug("received no response from peer")
|
||
|
if err := cc.Close(); err != nil {
|
||
|
e.logger.Error("error while closing connection", zap.Error(err))
|
||
|
}
|
||
|
return latest, nil
|
||
|
}
|
||
|
|
||
|
e.logger.Info(
|
||
|
"received new leading frame",
|
||
|
zap.Uint64("frame_number", response.ClockFrame.FrameNumber),
|
||
|
)
|
||
|
if err := cc.Close(); err != nil {
|
||
|
e.logger.Error("error while closing connection", zap.Error(err))
|
||
|
}
|
||
|
|
||
|
e.dataTimeReel.Insert(response.ClockFrame, false)
|
||
|
|
||
|
return response.ClockFrame, nil
|
||
|
}
|
||
|
|
||
|
func (e *DataClockConsensusEngine) collect(
|
||
|
currentFramePublished *protobufs.ClockFrame,
|
||
|
) (*protobufs.ClockFrame, error) {
|
||
|
e.logger.Info("collecting vdf proofs")
|
||
|
|
||
|
latest := currentFramePublished
|
||
|
|
||
|
for {
|
||
|
peerId, maxFrame, err := e.GetMostAheadPeer(latest.FrameNumber)
|
||
|
if maxFrame > latest.FrameNumber {
|
||
|
e.syncingStatus = SyncStatusSynchronizing
|
||
|
if err != nil {
|
||
|
e.logger.Info("no peers available for sync, waiting")
|
||
|
time.Sleep(5 * time.Second)
|
||
|
} else if maxFrame > latest.FrameNumber {
|
||
|
masterHead, err := e.masterTimeReel.Head()
|
||
|
if err != nil {
|
||
|
panic(err)
|
||
|
}
|
||
|
|
||
|
if masterHead.FrameNumber < maxFrame {
|
||
|
e.logger.Info(
|
||
|
"master frame synchronization needed to continue, waiting",
|
||
|
zap.Uint64("master_frame_head", masterHead.FrameNumber),
|
||
|
zap.Uint64("max_data_frame_target", maxFrame),
|
||
|
)
|
||
|
|
||
|
time.Sleep(30 * time.Second)
|
||
|
continue
|
||
|
}
|
||
|
|
||
|
latest, err = e.sync(latest, maxFrame, peerId)
|
||
|
if err == nil {
|
||
|
break
|
||
|
}
|
||
|
}
|
||
|
} else {
|
||
|
break
|
||
|
}
|
||
|
}
|
||
|
|
||
|
e.syncingStatus = SyncStatusNotSyncing
|
||
|
|
||
|
if latest.FrameNumber < currentFramePublished.FrameNumber {
|
||
|
latest = currentFramePublished
|
||
|
}
|
||
|
|
||
|
e.logger.Info(
|
||
|
"returning leader frame",
|
||
|
zap.Uint64("frame_number", latest.FrameNumber),
|
||
|
)
|
||
|
|
||
|
return latest, nil
|
||
|
}
|