mirror of
https://github.com/0glabs/0g-storage-node.git
synced 2024-12-27 16:55:18 +00:00
c2c6e2d5fb
* Implement Pruner. * Put pruner in a crate. * Fix clippy. * Add rpc zgs_getShardConfig. * Fix. * Increase wait time. * Add pruner_test and use max_num_chunks instead of size_limit. * Store shard config for peers and choose sync peers accordingly. * Add test and fix sync. * Fix clippy and test. * Fix some ut. * Add AnnounceShardConfig gossip and fix tests. * Add sharded tx finalize check in LogManager. * Try, * Rename. * Longer timeout for mine_test. * Save test logs.
115 lines
4.6 KiB
Python
Executable File
115 lines
4.6 KiB
Python
Executable File
#!/usr/bin/env python3
|
|
|
|
import random
|
|
import time
|
|
|
|
from test_framework.test_framework import TestFramework
|
|
from utility.submission import create_submission
|
|
from utility.submission import submit_data, data_to_segments
|
|
from utility.utils import (
|
|
assert_equal,
|
|
wait_until,
|
|
)
|
|
|
|
class SyncTest(TestFramework):
|
|
def setup_params(self):
|
|
self.num_blockchain_nodes = 2
|
|
self.num_nodes = 2
|
|
self.__deployed_contracts = 0
|
|
|
|
def run_test(self):
|
|
# By default, auto_sync_enabled and sync_file_on_announcement_enabled are both false,
|
|
# and file or chunks sync should be triggered by rpc.
|
|
self.__test_sync_file_by_rpc()
|
|
self.__test_sync_chunks_by_rpc()
|
|
|
|
def __test_sync_file_by_rpc(self):
|
|
self.log.info("Begin to test file sync by rpc")
|
|
|
|
client1 = self.nodes[0]
|
|
client2 = self.nodes[1]
|
|
|
|
# Create submission
|
|
chunk_data = random.randbytes(256 * 1024)
|
|
data_root = self.__create_submission(chunk_data)
|
|
|
|
# Ensure log entry sync from blockchain node
|
|
wait_until(lambda: client1.zgs_get_file_info(data_root) is not None)
|
|
assert_equal(client1.zgs_get_file_info(data_root)["finalized"], False)
|
|
|
|
# Upload file to storage node
|
|
segments = submit_data(client1, chunk_data)
|
|
self.log.info("segments: %s", [(s["root"], s["index"], s["proof"]) for s in segments])
|
|
wait_until(lambda: client1.zgs_get_file_info(data_root)["finalized"])
|
|
|
|
# File should not be auto sync on node 2
|
|
wait_until(lambda: client2.zgs_get_file_info(data_root) is not None)
|
|
time.sleep(3)
|
|
assert_equal(client2.zgs_get_file_info(data_root)["finalized"], False)
|
|
|
|
# Trigger file sync by rpc
|
|
assert(client2.admin_start_sync_file(0) is None)
|
|
wait_until(lambda: client2.sync_status_is_completed_or_unknown(0))
|
|
wait_until(lambda: client2.zgs_get_file_info(data_root)["finalized"])
|
|
|
|
# Validate data
|
|
assert_equal(
|
|
client2.zgs_download_segment(data_root, 0, 1024),
|
|
client1.zgs_download_segment(data_root, 0, 1024),
|
|
)
|
|
|
|
def __test_sync_chunks_by_rpc(self):
|
|
self.log.info("Begin to test chunks sync by rpc")
|
|
|
|
client1 = self.nodes[0]
|
|
client2 = self.nodes[1]
|
|
|
|
# Prepare 3 segments to upload
|
|
chunk_data = random.randbytes(256 * 1024 * 3)
|
|
data_root = self.__create_submission(chunk_data)
|
|
|
|
# Ensure log entry sync from blockchain node
|
|
wait_until(lambda: client1.zgs_get_file_info(data_root) is not None)
|
|
assert_equal(client1.zgs_get_file_info(data_root)["finalized"], False)
|
|
|
|
# Upload only 2nd segment to storage node
|
|
segments = data_to_segments(chunk_data)
|
|
self.log.info("segments: %s", [(s["root"], s["index"], s["proof"]) for s in segments])
|
|
assert(client1.zgs_upload_segment(segments[1]) is None)
|
|
|
|
# segment 0 is not able to download
|
|
assert(client1.zgs_download_segment_decoded(data_root, 0, 1024) is None)
|
|
# segment 1 is available to download
|
|
assert_equal(client1.zgs_download_segment_decoded(data_root, 1024, 2048), chunk_data[1024*256:2048*256])
|
|
# segment 2 is not able to download
|
|
assert(client1.zgs_download_segment_decoded(data_root, 2048, 3072) is None)
|
|
|
|
# Segment 1 should not be able to download on node 2
|
|
wait_until(lambda: client2.zgs_get_file_info(data_root) is not None)
|
|
assert_equal(client2.zgs_get_file_info(data_root)["finalized"], False)
|
|
assert(client2.zgs_download_segment_decoded(data_root, 1024, 2048) is None)
|
|
|
|
# Restart node 1 to check if the proof nodes are persisted.
|
|
self.stop_storage_node(0)
|
|
self.start_storage_node(0)
|
|
self.nodes[0].wait_for_rpc_connection()
|
|
|
|
# Trigger chunks sync by rpc
|
|
assert(client2.admin_start_sync_chunks(1, 1024, 2048) is None)
|
|
wait_until(lambda: client2.sync_status_is_completed_or_unknown(1))
|
|
wait_until(lambda: client2.zgs_download_segment_decoded(data_root, 1024, 2048) is not None)
|
|
|
|
# Validate data
|
|
assert_equal(client2.zgs_download_segment_decoded(data_root, 1024, 2048), chunk_data[1024*256:2048*256])
|
|
|
|
def __create_submission(self, chunk_data: bytes) -> str:
|
|
submissions, data_root = create_submission(chunk_data)
|
|
self.contract.submit(submissions)
|
|
self.__deployed_contracts += 1
|
|
wait_until(lambda: self.contract.num_submissions() == self.__deployed_contracts)
|
|
self.log.info("Submission created, data root: %s, submissions(%s) = %s", data_root, len(submissions), submissions)
|
|
return data_root
|
|
|
|
if __name__ == "__main__":
|
|
SyncTest().main()
|