mirror of
https://source.quilibrium.com/quilibrium/ceremonyclient.git
synced 2024-11-20 15:15:18 +00:00
v1.4.18-patch-2 (#230)
* feat: IPC for wesolowski * update self peer info * remove digests and signatures * add new binaries and digests * Signatory #13 added * Signatory #4 added (#231) * added sig.6 files (#232) * Signatory #9 added (#233) * Added signatories #1, #2, #3, #5, #8, #12, #14, #15, #16, #17 * remove binaries, release ready --------- Co-authored-by: 0xOzgur <29779769+0xOzgur@users.noreply.github.com> Co-authored-by: Demipoet <161999657+demipoet@users.noreply.github.com> Co-authored-by: Freekers <1370857+Freekers@users.noreply.github.com>
This commit is contained in:
parent
13bac91367
commit
58456c1057
@ -89,7 +89,10 @@ func GenerateVDFWithStopChan(seed []byte, iterations, int_size_bits uint32, stop
|
|||||||
func GenerateVDFIteration(seed, x_blob []byte, iterations, int_size_bits uint32) ([]byte, []byte) {
|
func GenerateVDFIteration(seed, x_blob []byte, iterations, int_size_bits uint32) ([]byte, []byte) {
|
||||||
int_size := (int_size_bits + 16) >> 4
|
int_size := (int_size_bits + 16) >> 4
|
||||||
D := iqc.CreateDiscriminant(seed, int_size_bits)
|
D := iqc.CreateDiscriminant(seed, int_size_bits)
|
||||||
x, _ := iqc.NewClassGroupFromBytesDiscriminant(x_blob[:(2*int_size)], D)
|
x, ok := iqc.NewClassGroupFromBytesDiscriminant(x_blob[:(2*int_size)], D)
|
||||||
|
if !ok {
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
|
||||||
y, proof := calculateVDF(D, x, iterations, int_size_bits, nil)
|
y, proof := calculateVDF(D, x, iterations, int_size_bits, nil)
|
||||||
|
|
||||||
@ -105,8 +108,15 @@ func VerifyVDF(seed, proof_blob []byte, iterations, int_size_bits uint32) bool {
|
|||||||
|
|
||||||
D := iqc.CreateDiscriminant(seed, int_size_bits)
|
D := iqc.CreateDiscriminant(seed, int_size_bits)
|
||||||
x := iqc.NewClassGroupFromAbDiscriminant(big.NewInt(2), big.NewInt(1), D)
|
x := iqc.NewClassGroupFromAbDiscriminant(big.NewInt(2), big.NewInt(1), D)
|
||||||
y, _ := iqc.NewClassGroupFromBytesDiscriminant(proof_blob[:(2*int_size)], D)
|
y, ok := iqc.NewClassGroupFromBytesDiscriminant(proof_blob[:(2*int_size)], D)
|
||||||
proof, _ := iqc.NewClassGroupFromBytesDiscriminant(proof_blob[2*int_size:], D)
|
if !ok {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
proof, ok := iqc.NewClassGroupFromBytesDiscriminant(proof_blob[2*int_size:], D)
|
||||||
|
if !ok {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
return verifyProof(x, y, proof, iterations)
|
return verifyProof(x, y, proof, iterations)
|
||||||
}
|
}
|
||||||
@ -114,9 +124,20 @@ func VerifyVDF(seed, proof_blob []byte, iterations, int_size_bits uint32) bool {
|
|||||||
func VerifyVDFIteration(seed, x_blob, proof_blob []byte, iterations, int_size_bits uint32) bool {
|
func VerifyVDFIteration(seed, x_blob, proof_blob []byte, iterations, int_size_bits uint32) bool {
|
||||||
int_size := (int_size_bits + 16) >> 4
|
int_size := (int_size_bits + 16) >> 4
|
||||||
D := iqc.CreateDiscriminant(seed, int_size_bits)
|
D := iqc.CreateDiscriminant(seed, int_size_bits)
|
||||||
x, _ := iqc.NewClassGroupFromBytesDiscriminant(x_blob[:(2*int_size)], D)
|
x, ok := iqc.NewClassGroupFromBytesDiscriminant(x_blob[:(2*int_size)], D)
|
||||||
y, _ := iqc.NewClassGroupFromBytesDiscriminant(proof_blob[:(2*int_size)], D)
|
if !ok {
|
||||||
proof, _ := iqc.NewClassGroupFromBytesDiscriminant(proof_blob[2*int_size:], D)
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
y, ok := iqc.NewClassGroupFromBytesDiscriminant(proof_blob[:(2*int_size)], D)
|
||||||
|
if !ok {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
proof, ok := iqc.NewClassGroupFromBytesDiscriminant(proof_blob[2*int_size:], D)
|
||||||
|
if !ok {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
return verifyProof(x, y, proof, iterations)
|
return verifyProof(x, y, proof, iterations)
|
||||||
}
|
}
|
||||||
|
@ -8,6 +8,14 @@ type EngineConfig struct {
|
|||||||
PendingCommitWorkers int64 `yaml:"pendingCommitWorkers"`
|
PendingCommitWorkers int64 `yaml:"pendingCommitWorkers"`
|
||||||
MinimumPeersRequired int `yaml:"minimumPeersRequired"`
|
MinimumPeersRequired int `yaml:"minimumPeersRequired"`
|
||||||
StatsMultiaddr string `yaml:"statsMultiaddr"`
|
StatsMultiaddr string `yaml:"statsMultiaddr"`
|
||||||
|
// Sets the fmt.Sprintf format string to use as the listen multiaddrs for
|
||||||
|
// data worker processes
|
||||||
|
DataWorkerBaseListenMultiaddr string `yaml:"dataWorkerBaseListenMultiaddr"`
|
||||||
|
// Sets the starting port number to use as the listen port for data worker
|
||||||
|
// processes, incrementing by 1 until n-1, n = cores. (Example: a 4 core
|
||||||
|
// system, base listen port of 40000 will listen on 40000, 40001, 40002)
|
||||||
|
DataWorkerBaseListenPort uint16 `yaml:"dataWorkerBaseListenPort"`
|
||||||
|
DataWorkerMemoryLimit int64 `yaml:"dataWorkerMemoryLimit"`
|
||||||
|
|
||||||
// Values used only for testing – do not override these in production, your
|
// Values used only for testing – do not override these in production, your
|
||||||
// node will get kicked out
|
// node will get kicked out
|
||||||
|
@ -29,5 +29,5 @@ func FormatVersion(version []byte) string {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func GetPatchNumber() byte {
|
func GetPatchNumber() byte {
|
||||||
return 0x01
|
return 0x02
|
||||||
}
|
}
|
||||||
|
@ -116,15 +116,19 @@ func (e *MasterClockConsensusEngine) handleSelfTestReport(
|
|||||||
peerID []byte,
|
peerID []byte,
|
||||||
any *anypb.Any,
|
any *anypb.Any,
|
||||||
) error {
|
) error {
|
||||||
if bytes.Equal(peerID, e.pubSub.GetPeerID()) {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
report := &protobufs.SelfTestReport{}
|
report := &protobufs.SelfTestReport{}
|
||||||
if err := any.UnmarshalTo(report); err != nil {
|
if err := any.UnmarshalTo(report); err != nil {
|
||||||
return errors.Wrap(err, "handle self test report")
|
return errors.Wrap(err, "handle self test report")
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if bytes.Equal(peerID, e.pubSub.GetPeerID()) {
|
||||||
|
info := e.peerInfoManager.GetPeerInfo(peerID)
|
||||||
|
info.LastSeen = time.Now().UnixMilli()
|
||||||
|
info.DifficultyMetric = report.DifficultyMetric
|
||||||
|
info.MasterHeadFrame = report.MasterHeadFrame
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
// minimum proof size is one timestamp, one vdf proof, must match one fewer
|
// minimum proof size is one timestamp, one vdf proof, must match one fewer
|
||||||
// than core count
|
// than core count
|
||||||
if len(report.Proof) < 516+8 ||
|
if len(report.Proof) < 516+8 ||
|
||||||
|
@ -7,6 +7,7 @@ import (
|
|||||||
"crypto/rand"
|
"crypto/rand"
|
||||||
"encoding/binary"
|
"encoding/binary"
|
||||||
"encoding/hex"
|
"encoding/hex"
|
||||||
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
"math/big"
|
"math/big"
|
||||||
"sync"
|
"sync"
|
||||||
@ -15,9 +16,12 @@ import (
|
|||||||
"github.com/iden3/go-iden3-crypto/poseidon"
|
"github.com/iden3/go-iden3-crypto/poseidon"
|
||||||
"github.com/libp2p/go-libp2p/core/peer"
|
"github.com/libp2p/go-libp2p/core/peer"
|
||||||
"github.com/mr-tron/base58"
|
"github.com/mr-tron/base58"
|
||||||
|
"github.com/multiformats/go-multiaddr"
|
||||||
|
mn "github.com/multiformats/go-multiaddr/net"
|
||||||
"github.com/pkg/errors"
|
"github.com/pkg/errors"
|
||||||
"go.uber.org/zap"
|
"go.uber.org/zap"
|
||||||
"google.golang.org/grpc"
|
"google.golang.org/grpc"
|
||||||
|
"google.golang.org/grpc/credentials/insecure"
|
||||||
"source.quilibrium.com/quilibrium/monorepo/node/config"
|
"source.quilibrium.com/quilibrium/monorepo/node/config"
|
||||||
"source.quilibrium.com/quilibrium/monorepo/node/consensus"
|
"source.quilibrium.com/quilibrium/monorepo/node/consensus"
|
||||||
qtime "source.quilibrium.com/quilibrium/monorepo/node/consensus/time"
|
qtime "source.quilibrium.com/quilibrium/monorepo/node/consensus/time"
|
||||||
@ -68,6 +72,7 @@ type MasterClockConsensusEngine struct {
|
|||||||
verifyTestCh chan verifyChallenge
|
verifyTestCh chan verifyChallenge
|
||||||
currentReceivingSyncPeers int
|
currentReceivingSyncPeers int
|
||||||
currentReceivingSyncPeersMx sync.Mutex
|
currentReceivingSyncPeersMx sync.Mutex
|
||||||
|
engineConfig *config.EngineConfig
|
||||||
}
|
}
|
||||||
|
|
||||||
var _ consensus.ConsensusEngine = (*MasterClockConsensusEngine)(nil)
|
var _ consensus.ConsensusEngine = (*MasterClockConsensusEngine)(nil)
|
||||||
@ -131,6 +136,7 @@ func NewMasterClockConsensusEngine(
|
|||||||
frameValidationCh: make(chan *protobufs.ClockFrame),
|
frameValidationCh: make(chan *protobufs.ClockFrame),
|
||||||
bandwidthTestCh: make(chan []byte),
|
bandwidthTestCh: make(chan []byte),
|
||||||
verifyTestCh: make(chan verifyChallenge),
|
verifyTestCh: make(chan verifyChallenge),
|
||||||
|
engineConfig: engineConfig,
|
||||||
}
|
}
|
||||||
|
|
||||||
e.addPeerManifestReport(e.pubSub.GetPeerID(), report)
|
e.addPeerManifestReport(e.pubSub.GetPeerID(), report)
|
||||||
@ -241,6 +247,16 @@ func (e *MasterClockConsensusEngine) Start() <-chan error {
|
|||||||
time.Sleep(30 * time.Second)
|
time.Sleep(30 * time.Second)
|
||||||
difficultyMetric := int64(100000)
|
difficultyMetric := int64(100000)
|
||||||
skew := (difficultyMetric * 12) / 10
|
skew := (difficultyMetric * 12) / 10
|
||||||
|
parallelism := e.report.Cores - 1
|
||||||
|
|
||||||
|
if parallelism < 3 {
|
||||||
|
panic("invalid system configuration, minimum system configuration must be four cores")
|
||||||
|
}
|
||||||
|
|
||||||
|
clients, err := e.createParallelDataClients(int(parallelism))
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
|
||||||
for {
|
for {
|
||||||
head, err := e.masterTimeReel.Head()
|
head, err := e.masterTimeReel.Head()
|
||||||
@ -250,7 +266,6 @@ func (e *MasterClockConsensusEngine) Start() <-chan error {
|
|||||||
|
|
||||||
e.report.MasterHeadFrame = head.FrameNumber
|
e.report.MasterHeadFrame = head.FrameNumber
|
||||||
e.report.DifficultyMetric = difficultyMetric
|
e.report.DifficultyMetric = difficultyMetric
|
||||||
parallelism := e.report.Cores - 1
|
|
||||||
|
|
||||||
challenge := binary.BigEndian.AppendUint64(
|
challenge := binary.BigEndian.AppendUint64(
|
||||||
[]byte{},
|
[]byte{},
|
||||||
@ -258,15 +273,42 @@ func (e *MasterClockConsensusEngine) Start() <-chan error {
|
|||||||
)
|
)
|
||||||
challenge = append(challenge, e.pubSub.GetPeerID()...)
|
challenge = append(challenge, e.pubSub.GetPeerID()...)
|
||||||
|
|
||||||
ts, proofs, nextDifficultyMetric, err :=
|
proofs := make([][]byte, parallelism)
|
||||||
e.frameProver.CalculateChallengeProof(
|
nextMetrics := make([]int64, parallelism)
|
||||||
challenge,
|
|
||||||
parallelism,
|
wg := sync.WaitGroup{}
|
||||||
skew,
|
wg.Add(int(parallelism))
|
||||||
)
|
|
||||||
if err != nil {
|
ts := time.Now().UnixMilli()
|
||||||
panic(err)
|
for i := uint32(0); i < parallelism; i++ {
|
||||||
|
i := i
|
||||||
|
go func() {
|
||||||
|
resp, err :=
|
||||||
|
clients[i].CalculateChallengeProof(
|
||||||
|
context.Background(),
|
||||||
|
&protobufs.ChallengeProofRequest{
|
||||||
|
Challenge: challenge,
|
||||||
|
Core: i,
|
||||||
|
Skew: skew,
|
||||||
|
NowMs: ts,
|
||||||
|
},
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
proofs[i], nextMetrics[i] = resp.Output, resp.NextSkew
|
||||||
|
wg.Done()
|
||||||
|
}()
|
||||||
}
|
}
|
||||||
|
wg.Wait()
|
||||||
|
nextDifficultySum := uint64(0)
|
||||||
|
for i := 0; i < int(parallelism); i++ {
|
||||||
|
nextDifficultySum += uint64(nextMetrics[i])
|
||||||
|
}
|
||||||
|
|
||||||
|
nextDifficultyMetric := int64(nextDifficultySum / uint64(parallelism))
|
||||||
|
|
||||||
e.logger.Info(
|
e.logger.Info(
|
||||||
"recalibrating difficulty metric",
|
"recalibrating difficulty metric",
|
||||||
zap.Int64("previous_difficulty_metric", difficultyMetric),
|
zap.Int64("previous_difficulty_metric", difficultyMetric),
|
||||||
@ -336,6 +378,64 @@ func (e *MasterClockConsensusEngine) Start() <-chan error {
|
|||||||
return errChan
|
return errChan
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (e *MasterClockConsensusEngine) createParallelDataClients(
|
||||||
|
paralellism int,
|
||||||
|
) ([]protobufs.DataIPCServiceClient, error) {
|
||||||
|
e.logger.Info(
|
||||||
|
"connecting to data worker processes",
|
||||||
|
zap.Int("parallelism", paralellism),
|
||||||
|
)
|
||||||
|
|
||||||
|
if e.engineConfig.DataWorkerBaseListenMultiaddr == "" {
|
||||||
|
e.engineConfig.DataWorkerBaseListenMultiaddr = "/ip4/127.0.0.1/tcp/%d"
|
||||||
|
}
|
||||||
|
|
||||||
|
if e.engineConfig.DataWorkerBaseListenPort == 0 {
|
||||||
|
e.engineConfig.DataWorkerBaseListenPort = 40000
|
||||||
|
}
|
||||||
|
|
||||||
|
clients := make([]protobufs.DataIPCServiceClient, paralellism)
|
||||||
|
|
||||||
|
for i := 0; i < paralellism; i++ {
|
||||||
|
ma, err := multiaddr.NewMultiaddr(
|
||||||
|
fmt.Sprintf(
|
||||||
|
e.engineConfig.DataWorkerBaseListenMultiaddr,
|
||||||
|
int(e.engineConfig.DataWorkerBaseListenPort)+i,
|
||||||
|
),
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
_, addr, err := mn.DialArgs(ma)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
conn, err := grpc.Dial(
|
||||||
|
addr,
|
||||||
|
grpc.WithTransportCredentials(
|
||||||
|
insecure.NewCredentials(),
|
||||||
|
),
|
||||||
|
grpc.WithDefaultCallOptions(
|
||||||
|
grpc.MaxCallSendMsgSize(10*1024*1024),
|
||||||
|
grpc.MaxCallRecvMsgSize(10*1024*1024),
|
||||||
|
),
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
clients[i] = protobufs.NewDataIPCServiceClient(conn)
|
||||||
|
}
|
||||||
|
|
||||||
|
e.logger.Info(
|
||||||
|
"connected to data worker processes",
|
||||||
|
zap.Int("parallelism", paralellism),
|
||||||
|
)
|
||||||
|
return clients, nil
|
||||||
|
}
|
||||||
|
|
||||||
func (e *MasterClockConsensusEngine) PerformValidation(
|
func (e *MasterClockConsensusEngine) PerformValidation(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
msg *protobufs.ValidationMessage,
|
msg *protobufs.ValidationMessage,
|
||||||
|
@ -53,9 +53,10 @@ type FrameProver interface {
|
|||||||
) bool
|
) bool
|
||||||
CalculateChallengeProof(
|
CalculateChallengeProof(
|
||||||
challenge []byte,
|
challenge []byte,
|
||||||
parallelism uint32,
|
core uint32,
|
||||||
skew int64,
|
skew int64,
|
||||||
) (int64, [][]byte, int64, error)
|
nowMs int64,
|
||||||
|
) ([]byte, int64, error)
|
||||||
VerifyChallengeProof(
|
VerifyChallengeProof(
|
||||||
challenge []byte,
|
challenge []byte,
|
||||||
timestamp int64,
|
timestamp int64,
|
||||||
|
@ -6,7 +6,6 @@ import (
|
|||||||
"crypto/rand"
|
"crypto/rand"
|
||||||
"encoding/binary"
|
"encoding/binary"
|
||||||
"math/big"
|
"math/big"
|
||||||
"sync"
|
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/cloudflare/circl/sign/ed448"
|
"github.com/cloudflare/circl/sign/ed448"
|
||||||
@ -601,43 +600,31 @@ func (w *WesolowskiFrameProver) VerifyWeakRecursiveProof(
|
|||||||
|
|
||||||
func (w *WesolowskiFrameProver) CalculateChallengeProof(
|
func (w *WesolowskiFrameProver) CalculateChallengeProof(
|
||||||
challenge []byte,
|
challenge []byte,
|
||||||
parallelism uint32,
|
core uint32,
|
||||||
skew int64,
|
skew int64,
|
||||||
) (int64, [][]byte, int64, error) {
|
nowMs int64,
|
||||||
now := time.Now()
|
) ([]byte, int64, error) {
|
||||||
nowMs := now.UnixMilli()
|
|
||||||
input := binary.BigEndian.AppendUint64([]byte{}, uint64(nowMs))
|
input := binary.BigEndian.AppendUint64([]byte{}, uint64(nowMs))
|
||||||
input = append(input, challenge...)
|
input = append(input, challenge...)
|
||||||
outputs := make([][]byte, parallelism)
|
|
||||||
|
|
||||||
wg := sync.WaitGroup{}
|
|
||||||
wg.Add(int(parallelism))
|
|
||||||
|
|
||||||
// 4.5 minutes = 270 seconds, one increment should be ten seconds
|
// 4.5 minutes = 270 seconds, one increment should be ten seconds
|
||||||
proofDuration := 270 * 1000
|
proofDuration := 270 * 1000
|
||||||
calibratedDifficulty := (int64(proofDuration) * 10000) / skew
|
calibratedDifficulty := (int64(proofDuration) * 10000) / skew
|
||||||
for i := uint32(0); i < parallelism; i++ {
|
instanceInput := binary.BigEndian.AppendUint32([]byte{}, core)
|
||||||
i := i
|
instanceInput = append(instanceInput, input...)
|
||||||
go func() {
|
b := sha3.Sum256(instanceInput)
|
||||||
instanceInput := binary.BigEndian.AppendUint32([]byte{}, i)
|
v := vdf.New(uint32(calibratedDifficulty), b)
|
||||||
instanceInput = append(instanceInput, input...)
|
|
||||||
b := sha3.Sum256(instanceInput)
|
|
||||||
v := vdf.New(uint32(calibratedDifficulty), b)
|
|
||||||
|
|
||||||
v.Execute()
|
v.Execute()
|
||||||
o := v.GetOutput()
|
o := v.GetOutput()
|
||||||
|
|
||||||
outputs[i] = make([]byte, 516)
|
output := make([]byte, 516)
|
||||||
copy(outputs[i][:], o[:])
|
copy(output[:], o[:])
|
||||||
wg.Done()
|
now := time.UnixMilli(nowMs)
|
||||||
}()
|
|
||||||
}
|
|
||||||
|
|
||||||
wg.Wait()
|
|
||||||
after := time.Since(now)
|
after := time.Since(now)
|
||||||
nextSkew := (skew * after.Milliseconds()) / int64(proofDuration)
|
nextSkew := (skew * after.Milliseconds()) / int64(proofDuration)
|
||||||
|
|
||||||
return nowMs, outputs, nextSkew, nil
|
return output, nextSkew, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (w *WesolowskiFrameProver) VerifyChallengeProof(
|
func (w *WesolowskiFrameProver) VerifyChallengeProof(
|
||||||
|
@ -30,10 +30,12 @@ func TestMasterProve(t *testing.T) {
|
|||||||
func TestChallengeProof(t *testing.T) {
|
func TestChallengeProof(t *testing.T) {
|
||||||
l, _ := zap.NewProduction()
|
l, _ := zap.NewProduction()
|
||||||
w := crypto.NewWesolowskiFrameProver(l)
|
w := crypto.NewWesolowskiFrameProver(l)
|
||||||
now, proofs, nextSkew, err := w.CalculateChallengeProof([]byte{0x01, 0x02, 0x03}, 3, 120000)
|
now := time.Now().UnixMilli()
|
||||||
|
proofs, nextSkew, err := w.CalculateChallengeProof([]byte{0x01, 0x02, 0x03}, 0, 120000, now)
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
assert.True(t, w.VerifyChallengeProof([]byte{0x01, 0x02, 0x03}, now, 100000, proofs))
|
assert.True(t, w.VerifyChallengeProof([]byte{0x01, 0x02, 0x03}, now, 100000, [][]byte{proofs}))
|
||||||
now, proofs, _, err = w.CalculateChallengeProof([]byte{0x01, 0x02, 0x03}, 3, nextSkew*12/10)
|
now = time.Now().UnixMilli()
|
||||||
|
proofs, _, err = w.CalculateChallengeProof([]byte{0x01, 0x02, 0x03}, 0, nextSkew*12/10, now)
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
assert.True(t, w.VerifyChallengeProof([]byte{0x01, 0x02, 0x03}, now, nextSkew, proofs))
|
assert.True(t, w.VerifyChallengeProof([]byte{0x01, 0x02, 0x03}, now, nextSkew, [][]byte{proofs}))
|
||||||
}
|
}
|
||||||
|
111
node/main.go
111
node/main.go
@ -12,9 +12,11 @@ import (
|
|||||||
"io/fs"
|
"io/fs"
|
||||||
"log"
|
"log"
|
||||||
"os"
|
"os"
|
||||||
|
"os/exec"
|
||||||
"os/signal"
|
"os/signal"
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
"runtime"
|
"runtime"
|
||||||
|
rdebug "runtime/debug"
|
||||||
"runtime/pprof"
|
"runtime/pprof"
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
@ -103,6 +105,16 @@ var (
|
|||||||
true,
|
true,
|
||||||
"enables or disables signature validation (default true)",
|
"enables or disables signature validation (default true)",
|
||||||
)
|
)
|
||||||
|
core = flag.Int(
|
||||||
|
"core",
|
||||||
|
0,
|
||||||
|
"specifies the core of the process (defaults to zero, the initial launcher)",
|
||||||
|
)
|
||||||
|
parentProcess = flag.Int(
|
||||||
|
"parent-process",
|
||||||
|
0,
|
||||||
|
"specifies the parent process pid for a data worker",
|
||||||
|
)
|
||||||
)
|
)
|
||||||
|
|
||||||
var signatories = []string{
|
var signatories = []string{
|
||||||
@ -264,7 +276,7 @@ func main() {
|
|||||||
done := make(chan os.Signal, 1)
|
done := make(chan os.Signal, 1)
|
||||||
signal.Notify(done, syscall.SIGINT, syscall.SIGTERM)
|
signal.Notify(done, syscall.SIGINT, syscall.SIGTERM)
|
||||||
|
|
||||||
if !*dbConsole {
|
if !*dbConsole && *core == 0 {
|
||||||
printLogo()
|
printLogo()
|
||||||
printVersion()
|
printVersion()
|
||||||
fmt.Println(" ")
|
fmt.Println(" ")
|
||||||
@ -322,7 +334,58 @@ func main() {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if *core != 0 {
|
||||||
|
runtime.GOMAXPROCS(1)
|
||||||
|
rdebug.SetGCPercent(9999)
|
||||||
|
|
||||||
|
if nodeConfig.Engine.DataWorkerMemoryLimit == 0 {
|
||||||
|
nodeConfig.Engine.DataWorkerMemoryLimit = 1792 * 1024 * 1024 // 1.75GiB
|
||||||
|
}
|
||||||
|
|
||||||
|
rdebug.SetMemoryLimit(nodeConfig.Engine.DataWorkerMemoryLimit)
|
||||||
|
|
||||||
|
if nodeConfig.Engine.DataWorkerBaseListenMultiaddr == "" {
|
||||||
|
nodeConfig.Engine.DataWorkerBaseListenMultiaddr = "/ip4/127.0.0.1/tcp/%d"
|
||||||
|
}
|
||||||
|
|
||||||
|
if nodeConfig.Engine.DataWorkerBaseListenPort == 0 {
|
||||||
|
nodeConfig.Engine.DataWorkerBaseListenPort = 40000
|
||||||
|
}
|
||||||
|
|
||||||
|
if *parentProcess == 0 {
|
||||||
|
panic("parent process pid not specified")
|
||||||
|
}
|
||||||
|
|
||||||
|
l, err := zap.NewProduction()
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
rpcMultiaddr := fmt.Sprintf(
|
||||||
|
nodeConfig.Engine.DataWorkerBaseListenMultiaddr,
|
||||||
|
int(nodeConfig.Engine.DataWorkerBaseListenPort)+*core-1,
|
||||||
|
)
|
||||||
|
srv, err := rpc.NewDataWorkerIPCServer(
|
||||||
|
rpcMultiaddr,
|
||||||
|
l,
|
||||||
|
uint32(*core)-1,
|
||||||
|
qcrypto.NewWesolowskiFrameProver(l),
|
||||||
|
*parentProcess,
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
err = srv.Start()
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
fmt.Println("Loading ceremony state and starting node...")
|
fmt.Println("Loading ceremony state and starting node...")
|
||||||
|
go spawnDataWorkers()
|
||||||
|
|
||||||
kzg.Init()
|
kzg.Init()
|
||||||
|
|
||||||
report := RunSelfTestIfNeeded(*configDirectory, nodeConfig)
|
report := RunSelfTestIfNeeded(*configDirectory, nodeConfig)
|
||||||
@ -365,9 +428,55 @@ func main() {
|
|||||||
node.Start()
|
node.Start()
|
||||||
|
|
||||||
<-done
|
<-done
|
||||||
|
stopDataWorkers()
|
||||||
node.Stop()
|
node.Stop()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
var dataWorkers []*exec.Cmd
|
||||||
|
|
||||||
|
func spawnDataWorkers() {
|
||||||
|
process, err := os.Executable()
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
cores := runtime.GOMAXPROCS(0)
|
||||||
|
dataWorkers = make([]*exec.Cmd, cores-1)
|
||||||
|
fmt.Printf("Spawning %d data workers...\n", cores-1)
|
||||||
|
|
||||||
|
for i := 1; i <= cores-1; i++ {
|
||||||
|
i := i
|
||||||
|
go func() {
|
||||||
|
args := []string{
|
||||||
|
fmt.Sprintf("--core=%d", i),
|
||||||
|
fmt.Sprintf("--parent-process=%d", os.Getpid()),
|
||||||
|
}
|
||||||
|
args = append(args, os.Args[1:]...)
|
||||||
|
cmd := exec.Command(process, args...)
|
||||||
|
cmd.Stdout = os.Stdout
|
||||||
|
cmd.Stderr = os.Stderr
|
||||||
|
err := cmd.Start()
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
dataWorkers[i-1] = cmd
|
||||||
|
}()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func stopDataWorkers() {
|
||||||
|
for i := 0; i < len(dataWorkers); i++ {
|
||||||
|
err := dataWorkers[i].Process.Signal(os.Kill)
|
||||||
|
if err != nil {
|
||||||
|
fmt.Printf(
|
||||||
|
"fatal: unable to kill worker with pid %d, please kill this process!\n",
|
||||||
|
dataWorkers[i].Process.Pid,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// Reintroduce at a later date
|
// Reintroduce at a later date
|
||||||
func RunCompaction(clockStore *store.PebbleClockStore) {
|
func RunCompaction(clockStore *store.PebbleClockStore) {
|
||||||
intrinsicFilter := append(
|
intrinsicFilter := append(
|
||||||
|
@ -1 +1 @@
|
|||||||
SHA3-256(node-1.4.18-darwin-arm64)= dc14a02268d88540bb364259775743c536d7541011bf26d4630f7fed425b5986
|
SHA3-256(node-1.4.18-darwin-arm64)= 0f666d6c3814f96bebce455034fe6a44b64f223b4e313ac506f42fa2371ac30d
|
||||||
|
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
BIN
node/node-1.4.18-darwin-arm64.dgst.sig.15
Normal file
BIN
node/node-1.4.18-darwin-arm64.dgst.sig.15
Normal file
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
BIN
node/node-1.4.18-darwin-arm64.dgst.sig.5
Normal file
BIN
node/node-1.4.18-darwin-arm64.dgst.sig.5
Normal file
Binary file not shown.
BIN
node/node-1.4.18-darwin-arm64.dgst.sig.6
Normal file
BIN
node/node-1.4.18-darwin-arm64.dgst.sig.6
Normal file
Binary file not shown.
Binary file not shown.
BIN
node/node-1.4.18-darwin-arm64.dgst.sig.9
Normal file
BIN
node/node-1.4.18-darwin-arm64.dgst.sig.9
Normal file
Binary file not shown.
@ -1 +1 @@
|
|||||||
SHA3-256(node-1.4.18-linux-amd64)= e41bf8538990e201637521b0eb278a1aebfc46e5d8de102f4870de30616b44e6
|
SHA3-256(node-1.4.18-linux-amd64)= f2da38a5c8d4257767f4a9659ca9c05d4cf312fd7294e6a18847c8923f6815a0
|
||||||
|
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
BIN
node/node-1.4.18-linux-amd64.dgst.sig.15
Normal file
BIN
node/node-1.4.18-linux-amd64.dgst.sig.15
Normal file
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
BIN
node/node-1.4.18-linux-amd64.dgst.sig.5
Normal file
BIN
node/node-1.4.18-linux-amd64.dgst.sig.5
Normal file
Binary file not shown.
BIN
node/node-1.4.18-linux-amd64.dgst.sig.6
Normal file
BIN
node/node-1.4.18-linux-amd64.dgst.sig.6
Normal file
Binary file not shown.
Binary file not shown.
BIN
node/node-1.4.18-linux-amd64.dgst.sig.9
Normal file
BIN
node/node-1.4.18-linux-amd64.dgst.sig.9
Normal file
Binary file not shown.
@ -1 +1 @@
|
|||||||
SHA3-256(node-1.4.18-linux-arm64)= de488e85acfd5ced235c8b7994d37cb153251826a52d1b0e964034b80d65b478
|
SHA3-256(node-1.4.18-linux-arm64)= e12e14a18a491921fcdd4ffe6e9792f1e195abf4950fe56316219f920867c45a
|
||||||
|
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
BIN
node/node-1.4.18-linux-arm64.dgst.sig.15
Normal file
BIN
node/node-1.4.18-linux-arm64.dgst.sig.15
Normal file
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
BIN
node/node-1.4.18-linux-arm64.dgst.sig.5
Normal file
BIN
node/node-1.4.18-linux-arm64.dgst.sig.5
Normal file
Binary file not shown.
BIN
node/node-1.4.18-linux-arm64.dgst.sig.6
Normal file
BIN
node/node-1.4.18-linux-arm64.dgst.sig.6
Normal file
Binary file not shown.
Binary file not shown.
BIN
node/node-1.4.18-linux-arm64.dgst.sig.9
Normal file
BIN
node/node-1.4.18-linux-arm64.dgst.sig.9
Normal file
Binary file not shown.
258
node/protobufs/data.pb.go
Normal file
258
node/protobufs/data.pb.go
Normal file
@ -0,0 +1,258 @@
|
|||||||
|
// Code generated by protoc-gen-go. DO NOT EDIT.
|
||||||
|
// versions:
|
||||||
|
// protoc-gen-go v1.30.0
|
||||||
|
// protoc v3.21.12
|
||||||
|
// source: data.proto
|
||||||
|
|
||||||
|
package protobufs
|
||||||
|
|
||||||
|
import (
|
||||||
|
protoreflect "google.golang.org/protobuf/reflect/protoreflect"
|
||||||
|
protoimpl "google.golang.org/protobuf/runtime/protoimpl"
|
||||||
|
reflect "reflect"
|
||||||
|
sync "sync"
|
||||||
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
// Verify that this generated code is sufficiently up-to-date.
|
||||||
|
_ = protoimpl.EnforceVersion(20 - protoimpl.MinVersion)
|
||||||
|
// Verify that runtime/protoimpl is sufficiently up-to-date.
|
||||||
|
_ = protoimpl.EnforceVersion(protoimpl.MaxVersion - 20)
|
||||||
|
)
|
||||||
|
|
||||||
|
type ChallengeProofRequest struct {
|
||||||
|
state protoimpl.MessageState
|
||||||
|
sizeCache protoimpl.SizeCache
|
||||||
|
unknownFields protoimpl.UnknownFields
|
||||||
|
|
||||||
|
Challenge []byte `protobuf:"bytes,1,opt,name=challenge,proto3" json:"challenge,omitempty"`
|
||||||
|
Core uint32 `protobuf:"varint,2,opt,name=core,proto3" json:"core,omitempty"`
|
||||||
|
Skew int64 `protobuf:"varint,3,opt,name=skew,proto3" json:"skew,omitempty"`
|
||||||
|
NowMs int64 `protobuf:"varint,4,opt,name=now_ms,json=nowMs,proto3" json:"now_ms,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
func (x *ChallengeProofRequest) Reset() {
|
||||||
|
*x = ChallengeProofRequest{}
|
||||||
|
if protoimpl.UnsafeEnabled {
|
||||||
|
mi := &file_data_proto_msgTypes[0]
|
||||||
|
ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x))
|
||||||
|
ms.StoreMessageInfo(mi)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (x *ChallengeProofRequest) String() string {
|
||||||
|
return protoimpl.X.MessageStringOf(x)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (*ChallengeProofRequest) ProtoMessage() {}
|
||||||
|
|
||||||
|
func (x *ChallengeProofRequest) ProtoReflect() protoreflect.Message {
|
||||||
|
mi := &file_data_proto_msgTypes[0]
|
||||||
|
if protoimpl.UnsafeEnabled && x != nil {
|
||||||
|
ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x))
|
||||||
|
if ms.LoadMessageInfo() == nil {
|
||||||
|
ms.StoreMessageInfo(mi)
|
||||||
|
}
|
||||||
|
return ms
|
||||||
|
}
|
||||||
|
return mi.MessageOf(x)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Deprecated: Use ChallengeProofRequest.ProtoReflect.Descriptor instead.
|
||||||
|
func (*ChallengeProofRequest) Descriptor() ([]byte, []int) {
|
||||||
|
return file_data_proto_rawDescGZIP(), []int{0}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (x *ChallengeProofRequest) GetChallenge() []byte {
|
||||||
|
if x != nil {
|
||||||
|
return x.Challenge
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (x *ChallengeProofRequest) GetCore() uint32 {
|
||||||
|
if x != nil {
|
||||||
|
return x.Core
|
||||||
|
}
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
|
||||||
|
func (x *ChallengeProofRequest) GetSkew() int64 {
|
||||||
|
if x != nil {
|
||||||
|
return x.Skew
|
||||||
|
}
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
|
||||||
|
func (x *ChallengeProofRequest) GetNowMs() int64 {
|
||||||
|
if x != nil {
|
||||||
|
return x.NowMs
|
||||||
|
}
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
|
||||||
|
type ChallengeProofResponse struct {
|
||||||
|
state protoimpl.MessageState
|
||||||
|
sizeCache protoimpl.SizeCache
|
||||||
|
unknownFields protoimpl.UnknownFields
|
||||||
|
|
||||||
|
Output []byte `protobuf:"bytes,1,opt,name=output,proto3" json:"output,omitempty"`
|
||||||
|
NextSkew int64 `protobuf:"varint,2,opt,name=next_skew,json=nextSkew,proto3" json:"next_skew,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
func (x *ChallengeProofResponse) Reset() {
|
||||||
|
*x = ChallengeProofResponse{}
|
||||||
|
if protoimpl.UnsafeEnabled {
|
||||||
|
mi := &file_data_proto_msgTypes[1]
|
||||||
|
ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x))
|
||||||
|
ms.StoreMessageInfo(mi)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (x *ChallengeProofResponse) String() string {
|
||||||
|
return protoimpl.X.MessageStringOf(x)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (*ChallengeProofResponse) ProtoMessage() {}
|
||||||
|
|
||||||
|
func (x *ChallengeProofResponse) ProtoReflect() protoreflect.Message {
|
||||||
|
mi := &file_data_proto_msgTypes[1]
|
||||||
|
if protoimpl.UnsafeEnabled && x != nil {
|
||||||
|
ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x))
|
||||||
|
if ms.LoadMessageInfo() == nil {
|
||||||
|
ms.StoreMessageInfo(mi)
|
||||||
|
}
|
||||||
|
return ms
|
||||||
|
}
|
||||||
|
return mi.MessageOf(x)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Deprecated: Use ChallengeProofResponse.ProtoReflect.Descriptor instead.
|
||||||
|
func (*ChallengeProofResponse) Descriptor() ([]byte, []int) {
|
||||||
|
return file_data_proto_rawDescGZIP(), []int{1}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (x *ChallengeProofResponse) GetOutput() []byte {
|
||||||
|
if x != nil {
|
||||||
|
return x.Output
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (x *ChallengeProofResponse) GetNextSkew() int64 {
|
||||||
|
if x != nil {
|
||||||
|
return x.NextSkew
|
||||||
|
}
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
|
||||||
|
var File_data_proto protoreflect.FileDescriptor
|
||||||
|
|
||||||
|
var file_data_proto_rawDesc = []byte{
|
||||||
|
0x0a, 0x0a, 0x64, 0x61, 0x74, 0x61, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x12, 0x17, 0x71, 0x75,
|
||||||
|
0x69, 0x6c, 0x69, 0x62, 0x72, 0x69, 0x75, 0x6d, 0x2e, 0x6e, 0x6f, 0x64, 0x65, 0x2e, 0x64, 0x61,
|
||||||
|
0x74, 0x61, 0x2e, 0x70, 0x62, 0x22, 0x74, 0x0a, 0x15, 0x43, 0x68, 0x61, 0x6c, 0x6c, 0x65, 0x6e,
|
||||||
|
0x67, 0x65, 0x50, 0x72, 0x6f, 0x6f, 0x66, 0x52, 0x65, 0x71, 0x75, 0x65, 0x73, 0x74, 0x12, 0x1c,
|
||||||
|
0x0a, 0x09, 0x63, 0x68, 0x61, 0x6c, 0x6c, 0x65, 0x6e, 0x67, 0x65, 0x18, 0x01, 0x20, 0x01, 0x28,
|
||||||
|
0x0c, 0x52, 0x09, 0x63, 0x68, 0x61, 0x6c, 0x6c, 0x65, 0x6e, 0x67, 0x65, 0x12, 0x12, 0x0a, 0x04,
|
||||||
|
0x63, 0x6f, 0x72, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x0d, 0x52, 0x04, 0x63, 0x6f, 0x72, 0x65,
|
||||||
|
0x12, 0x12, 0x0a, 0x04, 0x73, 0x6b, 0x65, 0x77, 0x18, 0x03, 0x20, 0x01, 0x28, 0x03, 0x52, 0x04,
|
||||||
|
0x73, 0x6b, 0x65, 0x77, 0x12, 0x15, 0x0a, 0x06, 0x6e, 0x6f, 0x77, 0x5f, 0x6d, 0x73, 0x18, 0x04,
|
||||||
|
0x20, 0x01, 0x28, 0x03, 0x52, 0x05, 0x6e, 0x6f, 0x77, 0x4d, 0x73, 0x22, 0x4d, 0x0a, 0x16, 0x43,
|
||||||
|
0x68, 0x61, 0x6c, 0x6c, 0x65, 0x6e, 0x67, 0x65, 0x50, 0x72, 0x6f, 0x6f, 0x66, 0x52, 0x65, 0x73,
|
||||||
|
0x70, 0x6f, 0x6e, 0x73, 0x65, 0x12, 0x16, 0x0a, 0x06, 0x6f, 0x75, 0x74, 0x70, 0x75, 0x74, 0x18,
|
||||||
|
0x01, 0x20, 0x01, 0x28, 0x0c, 0x52, 0x06, 0x6f, 0x75, 0x74, 0x70, 0x75, 0x74, 0x12, 0x1b, 0x0a,
|
||||||
|
0x09, 0x6e, 0x65, 0x78, 0x74, 0x5f, 0x73, 0x6b, 0x65, 0x77, 0x18, 0x02, 0x20, 0x01, 0x28, 0x03,
|
||||||
|
0x52, 0x08, 0x6e, 0x65, 0x78, 0x74, 0x53, 0x6b, 0x65, 0x77, 0x32, 0x8c, 0x01, 0x0a, 0x0e, 0x44,
|
||||||
|
0x61, 0x74, 0x61, 0x49, 0x50, 0x43, 0x53, 0x65, 0x72, 0x76, 0x69, 0x63, 0x65, 0x12, 0x7a, 0x0a,
|
||||||
|
0x17, 0x43, 0x61, 0x6c, 0x63, 0x75, 0x6c, 0x61, 0x74, 0x65, 0x43, 0x68, 0x61, 0x6c, 0x6c, 0x65,
|
||||||
|
0x6e, 0x67, 0x65, 0x50, 0x72, 0x6f, 0x6f, 0x66, 0x12, 0x2e, 0x2e, 0x71, 0x75, 0x69, 0x6c, 0x69,
|
||||||
|
0x62, 0x72, 0x69, 0x75, 0x6d, 0x2e, 0x6e, 0x6f, 0x64, 0x65, 0x2e, 0x64, 0x61, 0x74, 0x61, 0x2e,
|
||||||
|
0x70, 0x62, 0x2e, 0x43, 0x68, 0x61, 0x6c, 0x6c, 0x65, 0x6e, 0x67, 0x65, 0x50, 0x72, 0x6f, 0x6f,
|
||||||
|
0x66, 0x52, 0x65, 0x71, 0x75, 0x65, 0x73, 0x74, 0x1a, 0x2f, 0x2e, 0x71, 0x75, 0x69, 0x6c, 0x69,
|
||||||
|
0x62, 0x72, 0x69, 0x75, 0x6d, 0x2e, 0x6e, 0x6f, 0x64, 0x65, 0x2e, 0x64, 0x61, 0x74, 0x61, 0x2e,
|
||||||
|
0x70, 0x62, 0x2e, 0x43, 0x68, 0x61, 0x6c, 0x6c, 0x65, 0x6e, 0x67, 0x65, 0x50, 0x72, 0x6f, 0x6f,
|
||||||
|
0x66, 0x52, 0x65, 0x73, 0x70, 0x6f, 0x6e, 0x73, 0x65, 0x42, 0x3a, 0x5a, 0x38, 0x73, 0x6f, 0x75,
|
||||||
|
0x72, 0x63, 0x65, 0x2e, 0x71, 0x75, 0x69, 0x6c, 0x69, 0x62, 0x72, 0x69, 0x75, 0x6d, 0x2e, 0x63,
|
||||||
|
0x6f, 0x6d, 0x2f, 0x71, 0x75, 0x69, 0x6c, 0x69, 0x62, 0x72, 0x69, 0x75, 0x6d, 0x2f, 0x6d, 0x6f,
|
||||||
|
0x6e, 0x6f, 0x72, 0x65, 0x70, 0x6f, 0x2f, 0x6e, 0x6f, 0x64, 0x65, 0x2f, 0x70, 0x72, 0x6f, 0x74,
|
||||||
|
0x6f, 0x62, 0x75, 0x66, 0x73, 0x62, 0x06, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x33,
|
||||||
|
}
|
||||||
|
|
||||||
|
var (
|
||||||
|
file_data_proto_rawDescOnce sync.Once
|
||||||
|
file_data_proto_rawDescData = file_data_proto_rawDesc
|
||||||
|
)
|
||||||
|
|
||||||
|
func file_data_proto_rawDescGZIP() []byte {
|
||||||
|
file_data_proto_rawDescOnce.Do(func() {
|
||||||
|
file_data_proto_rawDescData = protoimpl.X.CompressGZIP(file_data_proto_rawDescData)
|
||||||
|
})
|
||||||
|
return file_data_proto_rawDescData
|
||||||
|
}
|
||||||
|
|
||||||
|
var file_data_proto_msgTypes = make([]protoimpl.MessageInfo, 2)
|
||||||
|
var file_data_proto_goTypes = []interface{}{
|
||||||
|
(*ChallengeProofRequest)(nil), // 0: quilibrium.node.data.pb.ChallengeProofRequest
|
||||||
|
(*ChallengeProofResponse)(nil), // 1: quilibrium.node.data.pb.ChallengeProofResponse
|
||||||
|
}
|
||||||
|
var file_data_proto_depIdxs = []int32{
|
||||||
|
0, // 0: quilibrium.node.data.pb.DataIPCService.CalculateChallengeProof:input_type -> quilibrium.node.data.pb.ChallengeProofRequest
|
||||||
|
1, // 1: quilibrium.node.data.pb.DataIPCService.CalculateChallengeProof:output_type -> quilibrium.node.data.pb.ChallengeProofResponse
|
||||||
|
1, // [1:2] is the sub-list for method output_type
|
||||||
|
0, // [0:1] is the sub-list for method input_type
|
||||||
|
0, // [0:0] is the sub-list for extension type_name
|
||||||
|
0, // [0:0] is the sub-list for extension extendee
|
||||||
|
0, // [0:0] is the sub-list for field type_name
|
||||||
|
}
|
||||||
|
|
||||||
|
func init() { file_data_proto_init() }
|
||||||
|
func file_data_proto_init() {
|
||||||
|
if File_data_proto != nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if !protoimpl.UnsafeEnabled {
|
||||||
|
file_data_proto_msgTypes[0].Exporter = func(v interface{}, i int) interface{} {
|
||||||
|
switch v := v.(*ChallengeProofRequest); i {
|
||||||
|
case 0:
|
||||||
|
return &v.state
|
||||||
|
case 1:
|
||||||
|
return &v.sizeCache
|
||||||
|
case 2:
|
||||||
|
return &v.unknownFields
|
||||||
|
default:
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
file_data_proto_msgTypes[1].Exporter = func(v interface{}, i int) interface{} {
|
||||||
|
switch v := v.(*ChallengeProofResponse); i {
|
||||||
|
case 0:
|
||||||
|
return &v.state
|
||||||
|
case 1:
|
||||||
|
return &v.sizeCache
|
||||||
|
case 2:
|
||||||
|
return &v.unknownFields
|
||||||
|
default:
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
type x struct{}
|
||||||
|
out := protoimpl.TypeBuilder{
|
||||||
|
File: protoimpl.DescBuilder{
|
||||||
|
GoPackagePath: reflect.TypeOf(x{}).PkgPath(),
|
||||||
|
RawDescriptor: file_data_proto_rawDesc,
|
||||||
|
NumEnums: 0,
|
||||||
|
NumMessages: 2,
|
||||||
|
NumExtensions: 0,
|
||||||
|
NumServices: 1,
|
||||||
|
},
|
||||||
|
GoTypes: file_data_proto_goTypes,
|
||||||
|
DependencyIndexes: file_data_proto_depIdxs,
|
||||||
|
MessageInfos: file_data_proto_msgTypes,
|
||||||
|
}.Build()
|
||||||
|
File_data_proto = out.File
|
||||||
|
file_data_proto_rawDesc = nil
|
||||||
|
file_data_proto_goTypes = nil
|
||||||
|
file_data_proto_depIdxs = nil
|
||||||
|
}
|
171
node/protobufs/data.pb.gw.go
Normal file
171
node/protobufs/data.pb.gw.go
Normal file
@ -0,0 +1,171 @@
|
|||||||
|
// Code generated by protoc-gen-grpc-gateway. DO NOT EDIT.
|
||||||
|
// source: data.proto
|
||||||
|
|
||||||
|
/*
|
||||||
|
Package protobufs is a reverse proxy.
|
||||||
|
|
||||||
|
It translates gRPC into RESTful JSON APIs.
|
||||||
|
*/
|
||||||
|
package protobufs
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"io"
|
||||||
|
"net/http"
|
||||||
|
|
||||||
|
"github.com/grpc-ecosystem/grpc-gateway/v2/runtime"
|
||||||
|
"github.com/grpc-ecosystem/grpc-gateway/v2/utilities"
|
||||||
|
"google.golang.org/grpc"
|
||||||
|
"google.golang.org/grpc/codes"
|
||||||
|
"google.golang.org/grpc/grpclog"
|
||||||
|
"google.golang.org/grpc/metadata"
|
||||||
|
"google.golang.org/grpc/status"
|
||||||
|
"google.golang.org/protobuf/proto"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Suppress "imported and not used" errors
|
||||||
|
var _ codes.Code
|
||||||
|
var _ io.Reader
|
||||||
|
var _ status.Status
|
||||||
|
var _ = runtime.String
|
||||||
|
var _ = utilities.NewDoubleArray
|
||||||
|
var _ = metadata.Join
|
||||||
|
|
||||||
|
func request_DataIPCService_CalculateChallengeProof_0(ctx context.Context, marshaler runtime.Marshaler, client DataIPCServiceClient, req *http.Request, pathParams map[string]string) (proto.Message, runtime.ServerMetadata, error) {
|
||||||
|
var protoReq ChallengeProofRequest
|
||||||
|
var metadata runtime.ServerMetadata
|
||||||
|
|
||||||
|
newReader, berr := utilities.IOReaderFactory(req.Body)
|
||||||
|
if berr != nil {
|
||||||
|
return nil, metadata, status.Errorf(codes.InvalidArgument, "%v", berr)
|
||||||
|
}
|
||||||
|
if err := marshaler.NewDecoder(newReader()).Decode(&protoReq); err != nil && err != io.EOF {
|
||||||
|
return nil, metadata, status.Errorf(codes.InvalidArgument, "%v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
msg, err := client.CalculateChallengeProof(ctx, &protoReq, grpc.Header(&metadata.HeaderMD), grpc.Trailer(&metadata.TrailerMD))
|
||||||
|
return msg, metadata, err
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
func local_request_DataIPCService_CalculateChallengeProof_0(ctx context.Context, marshaler runtime.Marshaler, server DataIPCServiceServer, req *http.Request, pathParams map[string]string) (proto.Message, runtime.ServerMetadata, error) {
|
||||||
|
var protoReq ChallengeProofRequest
|
||||||
|
var metadata runtime.ServerMetadata
|
||||||
|
|
||||||
|
newReader, berr := utilities.IOReaderFactory(req.Body)
|
||||||
|
if berr != nil {
|
||||||
|
return nil, metadata, status.Errorf(codes.InvalidArgument, "%v", berr)
|
||||||
|
}
|
||||||
|
if err := marshaler.NewDecoder(newReader()).Decode(&protoReq); err != nil && err != io.EOF {
|
||||||
|
return nil, metadata, status.Errorf(codes.InvalidArgument, "%v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
msg, err := server.CalculateChallengeProof(ctx, &protoReq)
|
||||||
|
return msg, metadata, err
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
// RegisterDataIPCServiceHandlerServer registers the http handlers for service DataIPCService to "mux".
|
||||||
|
// UnaryRPC :call DataIPCServiceServer directly.
|
||||||
|
// StreamingRPC :currently unsupported pending https://github.com/grpc/grpc-go/issues/906.
|
||||||
|
// Note that using this registration option will cause many gRPC library features to stop working. Consider using RegisterDataIPCServiceHandlerFromEndpoint instead.
|
||||||
|
func RegisterDataIPCServiceHandlerServer(ctx context.Context, mux *runtime.ServeMux, server DataIPCServiceServer) error {
|
||||||
|
|
||||||
|
mux.Handle("POST", pattern_DataIPCService_CalculateChallengeProof_0, func(w http.ResponseWriter, req *http.Request, pathParams map[string]string) {
|
||||||
|
ctx, cancel := context.WithCancel(req.Context())
|
||||||
|
defer cancel()
|
||||||
|
var stream runtime.ServerTransportStream
|
||||||
|
ctx = grpc.NewContextWithServerTransportStream(ctx, &stream)
|
||||||
|
inboundMarshaler, outboundMarshaler := runtime.MarshalerForRequest(mux, req)
|
||||||
|
var err error
|
||||||
|
var annotatedContext context.Context
|
||||||
|
annotatedContext, err = runtime.AnnotateIncomingContext(ctx, mux, req, "/quilibrium.node.data.pb.DataIPCService/CalculateChallengeProof", runtime.WithHTTPPathPattern("/quilibrium.node.data.pb.DataIPCService/CalculateChallengeProof"))
|
||||||
|
if err != nil {
|
||||||
|
runtime.HTTPError(ctx, mux, outboundMarshaler, w, req, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
resp, md, err := local_request_DataIPCService_CalculateChallengeProof_0(annotatedContext, inboundMarshaler, server, req, pathParams)
|
||||||
|
md.HeaderMD, md.TrailerMD = metadata.Join(md.HeaderMD, stream.Header()), metadata.Join(md.TrailerMD, stream.Trailer())
|
||||||
|
annotatedContext = runtime.NewServerMetadataContext(annotatedContext, md)
|
||||||
|
if err != nil {
|
||||||
|
runtime.HTTPError(annotatedContext, mux, outboundMarshaler, w, req, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
forward_DataIPCService_CalculateChallengeProof_0(annotatedContext, mux, outboundMarshaler, w, req, resp, mux.GetForwardResponseOptions()...)
|
||||||
|
|
||||||
|
})
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// RegisterDataIPCServiceHandlerFromEndpoint is same as RegisterDataIPCServiceHandler but
|
||||||
|
// automatically dials to "endpoint" and closes the connection when "ctx" gets done.
|
||||||
|
func RegisterDataIPCServiceHandlerFromEndpoint(ctx context.Context, mux *runtime.ServeMux, endpoint string, opts []grpc.DialOption) (err error) {
|
||||||
|
conn, err := grpc.DialContext(ctx, endpoint, opts...)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
defer func() {
|
||||||
|
if err != nil {
|
||||||
|
if cerr := conn.Close(); cerr != nil {
|
||||||
|
grpclog.Infof("Failed to close conn to %s: %v", endpoint, cerr)
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
go func() {
|
||||||
|
<-ctx.Done()
|
||||||
|
if cerr := conn.Close(); cerr != nil {
|
||||||
|
grpclog.Infof("Failed to close conn to %s: %v", endpoint, cerr)
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
}()
|
||||||
|
|
||||||
|
return RegisterDataIPCServiceHandler(ctx, mux, conn)
|
||||||
|
}
|
||||||
|
|
||||||
|
// RegisterDataIPCServiceHandler registers the http handlers for service DataIPCService to "mux".
|
||||||
|
// The handlers forward requests to the grpc endpoint over "conn".
|
||||||
|
func RegisterDataIPCServiceHandler(ctx context.Context, mux *runtime.ServeMux, conn *grpc.ClientConn) error {
|
||||||
|
return RegisterDataIPCServiceHandlerClient(ctx, mux, NewDataIPCServiceClient(conn))
|
||||||
|
}
|
||||||
|
|
||||||
|
// RegisterDataIPCServiceHandlerClient registers the http handlers for service DataIPCService
|
||||||
|
// to "mux". The handlers forward requests to the grpc endpoint over the given implementation of "DataIPCServiceClient".
|
||||||
|
// Note: the gRPC framework executes interceptors within the gRPC handler. If the passed in "DataIPCServiceClient"
|
||||||
|
// doesn't go through the normal gRPC flow (creating a gRPC client etc.) then it will be up to the passed in
|
||||||
|
// "DataIPCServiceClient" to call the correct interceptors.
|
||||||
|
func RegisterDataIPCServiceHandlerClient(ctx context.Context, mux *runtime.ServeMux, client DataIPCServiceClient) error {
|
||||||
|
|
||||||
|
mux.Handle("POST", pattern_DataIPCService_CalculateChallengeProof_0, func(w http.ResponseWriter, req *http.Request, pathParams map[string]string) {
|
||||||
|
ctx, cancel := context.WithCancel(req.Context())
|
||||||
|
defer cancel()
|
||||||
|
inboundMarshaler, outboundMarshaler := runtime.MarshalerForRequest(mux, req)
|
||||||
|
var err error
|
||||||
|
var annotatedContext context.Context
|
||||||
|
annotatedContext, err = runtime.AnnotateContext(ctx, mux, req, "/quilibrium.node.data.pb.DataIPCService/CalculateChallengeProof", runtime.WithHTTPPathPattern("/quilibrium.node.data.pb.DataIPCService/CalculateChallengeProof"))
|
||||||
|
if err != nil {
|
||||||
|
runtime.HTTPError(ctx, mux, outboundMarshaler, w, req, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
resp, md, err := request_DataIPCService_CalculateChallengeProof_0(annotatedContext, inboundMarshaler, client, req, pathParams)
|
||||||
|
annotatedContext = runtime.NewServerMetadataContext(annotatedContext, md)
|
||||||
|
if err != nil {
|
||||||
|
runtime.HTTPError(annotatedContext, mux, outboundMarshaler, w, req, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
forward_DataIPCService_CalculateChallengeProof_0(annotatedContext, mux, outboundMarshaler, w, req, resp, mux.GetForwardResponseOptions()...)
|
||||||
|
|
||||||
|
})
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
var (
|
||||||
|
pattern_DataIPCService_CalculateChallengeProof_0 = runtime.MustPattern(runtime.NewPattern(1, []int{2, 0, 2, 1}, []string{"quilibrium.node.data.pb.DataIPCService", "CalculateChallengeProof"}, ""))
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
forward_DataIPCService_CalculateChallengeProof_0 = runtime.ForwardResponseMessage
|
||||||
|
)
|
21
node/protobufs/data.proto
Normal file
21
node/protobufs/data.proto
Normal file
@ -0,0 +1,21 @@
|
|||||||
|
syntax = "proto3";
|
||||||
|
|
||||||
|
package quilibrium.node.data.pb;
|
||||||
|
|
||||||
|
option go_package = "source.quilibrium.com/quilibrium/monorepo/node/protobufs";
|
||||||
|
|
||||||
|
message ChallengeProofRequest {
|
||||||
|
bytes challenge = 1;
|
||||||
|
uint32 core = 2;
|
||||||
|
int64 skew = 3;
|
||||||
|
int64 now_ms = 4;
|
||||||
|
}
|
||||||
|
|
||||||
|
message ChallengeProofResponse {
|
||||||
|
bytes output = 1;
|
||||||
|
int64 next_skew = 2;
|
||||||
|
}
|
||||||
|
|
||||||
|
service DataIPCService {
|
||||||
|
rpc CalculateChallengeProof(ChallengeProofRequest) returns (ChallengeProofResponse);
|
||||||
|
}
|
109
node/protobufs/data_grpc.pb.go
Normal file
109
node/protobufs/data_grpc.pb.go
Normal file
@ -0,0 +1,109 @@
|
|||||||
|
// Code generated by protoc-gen-go-grpc. DO NOT EDIT.
|
||||||
|
// versions:
|
||||||
|
// - protoc-gen-go-grpc v1.3.0
|
||||||
|
// - protoc v3.21.12
|
||||||
|
// source: data.proto
|
||||||
|
|
||||||
|
package protobufs
|
||||||
|
|
||||||
|
import (
|
||||||
|
context "context"
|
||||||
|
grpc "google.golang.org/grpc"
|
||||||
|
codes "google.golang.org/grpc/codes"
|
||||||
|
status "google.golang.org/grpc/status"
|
||||||
|
)
|
||||||
|
|
||||||
|
// This is a compile-time assertion to ensure that this generated file
|
||||||
|
// is compatible with the grpc package it is being compiled against.
|
||||||
|
// Requires gRPC-Go v1.32.0 or later.
|
||||||
|
const _ = grpc.SupportPackageIsVersion7
|
||||||
|
|
||||||
|
const (
|
||||||
|
DataIPCService_CalculateChallengeProof_FullMethodName = "/quilibrium.node.data.pb.DataIPCService/CalculateChallengeProof"
|
||||||
|
)
|
||||||
|
|
||||||
|
// DataIPCServiceClient is the client API for DataIPCService service.
|
||||||
|
//
|
||||||
|
// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://pkg.go.dev/google.golang.org/grpc/?tab=doc#ClientConn.NewStream.
|
||||||
|
type DataIPCServiceClient interface {
|
||||||
|
CalculateChallengeProof(ctx context.Context, in *ChallengeProofRequest, opts ...grpc.CallOption) (*ChallengeProofResponse, error)
|
||||||
|
}
|
||||||
|
|
||||||
|
type dataIPCServiceClient struct {
|
||||||
|
cc grpc.ClientConnInterface
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewDataIPCServiceClient(cc grpc.ClientConnInterface) DataIPCServiceClient {
|
||||||
|
return &dataIPCServiceClient{cc}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *dataIPCServiceClient) CalculateChallengeProof(ctx context.Context, in *ChallengeProofRequest, opts ...grpc.CallOption) (*ChallengeProofResponse, error) {
|
||||||
|
out := new(ChallengeProofResponse)
|
||||||
|
err := c.cc.Invoke(ctx, DataIPCService_CalculateChallengeProof_FullMethodName, in, out, opts...)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return out, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// DataIPCServiceServer is the server API for DataIPCService service.
|
||||||
|
// All implementations must embed UnimplementedDataIPCServiceServer
|
||||||
|
// for forward compatibility
|
||||||
|
type DataIPCServiceServer interface {
|
||||||
|
CalculateChallengeProof(context.Context, *ChallengeProofRequest) (*ChallengeProofResponse, error)
|
||||||
|
mustEmbedUnimplementedDataIPCServiceServer()
|
||||||
|
}
|
||||||
|
|
||||||
|
// UnimplementedDataIPCServiceServer must be embedded to have forward compatible implementations.
|
||||||
|
type UnimplementedDataIPCServiceServer struct {
|
||||||
|
}
|
||||||
|
|
||||||
|
func (UnimplementedDataIPCServiceServer) CalculateChallengeProof(context.Context, *ChallengeProofRequest) (*ChallengeProofResponse, error) {
|
||||||
|
return nil, status.Errorf(codes.Unimplemented, "method CalculateChallengeProof not implemented")
|
||||||
|
}
|
||||||
|
func (UnimplementedDataIPCServiceServer) mustEmbedUnimplementedDataIPCServiceServer() {}
|
||||||
|
|
||||||
|
// UnsafeDataIPCServiceServer may be embedded to opt out of forward compatibility for this service.
|
||||||
|
// Use of this interface is not recommended, as added methods to DataIPCServiceServer will
|
||||||
|
// result in compilation errors.
|
||||||
|
type UnsafeDataIPCServiceServer interface {
|
||||||
|
mustEmbedUnimplementedDataIPCServiceServer()
|
||||||
|
}
|
||||||
|
|
||||||
|
func RegisterDataIPCServiceServer(s grpc.ServiceRegistrar, srv DataIPCServiceServer) {
|
||||||
|
s.RegisterService(&DataIPCService_ServiceDesc, srv)
|
||||||
|
}
|
||||||
|
|
||||||
|
func _DataIPCService_CalculateChallengeProof_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) {
|
||||||
|
in := new(ChallengeProofRequest)
|
||||||
|
if err := dec(in); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
if interceptor == nil {
|
||||||
|
return srv.(DataIPCServiceServer).CalculateChallengeProof(ctx, in)
|
||||||
|
}
|
||||||
|
info := &grpc.UnaryServerInfo{
|
||||||
|
Server: srv,
|
||||||
|
FullMethod: DataIPCService_CalculateChallengeProof_FullMethodName,
|
||||||
|
}
|
||||||
|
handler := func(ctx context.Context, req interface{}) (interface{}, error) {
|
||||||
|
return srv.(DataIPCServiceServer).CalculateChallengeProof(ctx, req.(*ChallengeProofRequest))
|
||||||
|
}
|
||||||
|
return interceptor(ctx, in, info, handler)
|
||||||
|
}
|
||||||
|
|
||||||
|
// DataIPCService_ServiceDesc is the grpc.ServiceDesc for DataIPCService service.
|
||||||
|
// It's only intended for direct use with grpc.RegisterService,
|
||||||
|
// and not to be introspected or modified (even as a copy)
|
||||||
|
var DataIPCService_ServiceDesc = grpc.ServiceDesc{
|
||||||
|
ServiceName: "quilibrium.node.data.pb.DataIPCService",
|
||||||
|
HandlerType: (*DataIPCServiceServer)(nil),
|
||||||
|
Methods: []grpc.MethodDesc{
|
||||||
|
{
|
||||||
|
MethodName: "CalculateChallengeProof",
|
||||||
|
Handler: _DataIPCService_CalculateChallengeProof_Handler,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
Streams: []grpc.StreamDesc{},
|
||||||
|
Metadata: "data.proto",
|
||||||
|
}
|
119
node/rpc/data_worker_ipc_server.go
Normal file
119
node/rpc/data_worker_ipc_server.go
Normal file
@ -0,0 +1,119 @@
|
|||||||
|
package rpc
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"os"
|
||||||
|
"runtime"
|
||||||
|
"syscall"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"source.quilibrium.com/quilibrium/monorepo/node/crypto"
|
||||||
|
|
||||||
|
"github.com/multiformats/go-multiaddr"
|
||||||
|
mn "github.com/multiformats/go-multiaddr/net"
|
||||||
|
"github.com/pkg/errors"
|
||||||
|
"go.uber.org/zap"
|
||||||
|
"google.golang.org/grpc"
|
||||||
|
"google.golang.org/grpc/reflection"
|
||||||
|
"source.quilibrium.com/quilibrium/monorepo/node/protobufs"
|
||||||
|
)
|
||||||
|
|
||||||
|
type DataWorkerIPCServer struct {
|
||||||
|
protobufs.UnimplementedDataIPCServiceServer
|
||||||
|
listenAddrGRPC string
|
||||||
|
logger *zap.Logger
|
||||||
|
coreId uint32
|
||||||
|
prover crypto.FrameProver
|
||||||
|
parentProcessId int
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetFrameInfo implements protobufs.NodeServiceServer.
|
||||||
|
func (r *DataWorkerIPCServer) CalculateChallengeProof(
|
||||||
|
ctx context.Context,
|
||||||
|
req *protobufs.ChallengeProofRequest,
|
||||||
|
) (*protobufs.ChallengeProofResponse, error) {
|
||||||
|
if r.coreId != req.Core {
|
||||||
|
return nil, errors.Wrap(
|
||||||
|
errors.New("invalid core id"),
|
||||||
|
"calculate challenge proof",
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
proof, nextSkew, err := r.prover.CalculateChallengeProof(
|
||||||
|
req.Challenge,
|
||||||
|
uint32(r.coreId),
|
||||||
|
req.Skew,
|
||||||
|
req.NowMs,
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
return nil, errors.Wrap(err, "calculate challenge proof")
|
||||||
|
}
|
||||||
|
|
||||||
|
return &protobufs.ChallengeProofResponse{
|
||||||
|
Output: proof,
|
||||||
|
NextSkew: nextSkew,
|
||||||
|
}, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewDataWorkerIPCServer(
|
||||||
|
listenAddrGRPC string,
|
||||||
|
logger *zap.Logger,
|
||||||
|
coreId uint32,
|
||||||
|
prover crypto.FrameProver,
|
||||||
|
parentProcessId int,
|
||||||
|
) (*DataWorkerIPCServer, error) {
|
||||||
|
return &DataWorkerIPCServer{
|
||||||
|
listenAddrGRPC: listenAddrGRPC,
|
||||||
|
logger: logger,
|
||||||
|
coreId: coreId,
|
||||||
|
prover: prover,
|
||||||
|
parentProcessId: parentProcessId,
|
||||||
|
}, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (r *DataWorkerIPCServer) Start() error {
|
||||||
|
s := grpc.NewServer(
|
||||||
|
grpc.MaxRecvMsgSize(10*1024*1024),
|
||||||
|
grpc.MaxSendMsgSize(10*1024*1024),
|
||||||
|
)
|
||||||
|
protobufs.RegisterDataIPCServiceServer(s, r)
|
||||||
|
reflection.Register(s)
|
||||||
|
|
||||||
|
mg, err := multiaddr.NewMultiaddr(r.listenAddrGRPC)
|
||||||
|
if err != nil {
|
||||||
|
return errors.Wrap(err, "start")
|
||||||
|
}
|
||||||
|
|
||||||
|
lis, err := mn.Listen(mg)
|
||||||
|
if err != nil {
|
||||||
|
return errors.Wrap(err, "start")
|
||||||
|
}
|
||||||
|
|
||||||
|
go r.monitorParent()
|
||||||
|
|
||||||
|
if err := s.Serve(mn.NetListener(lis)); err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (r *DataWorkerIPCServer) monitorParent() {
|
||||||
|
for {
|
||||||
|
time.Sleep(1 * time.Second)
|
||||||
|
proc, err := os.FindProcess(r.parentProcessId)
|
||||||
|
if err != nil {
|
||||||
|
r.logger.Error("parent process not found, terminating")
|
||||||
|
os.Exit(1)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Windows returns an error if the process is dead, nobody else does
|
||||||
|
if runtime.GOOS != "windows" {
|
||||||
|
err := proc.Signal(syscall.Signal(0))
|
||||||
|
if err != nil {
|
||||||
|
r.logger.Error("parent process not found, terminating")
|
||||||
|
os.Exit(1)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
Loading…
Reference in New Issue
Block a user