mirror of
https://github.com/0glabs/0g-storage-node.git
synced 2024-12-26 08:15:17 +00:00
4eb2a50b0e
* Use inner lock in storage.
* Remove mut.
* Remove async lock for storage.
* Fix tests and warnings.
* Use spawn_blocking for storage task.
* Fix clippy.
* Finalize the new tx at last.
* Revert "Finalize the new tx at last."
This reverts commit b56ad5582d
.
* Wait for old same-root txs to finalize.
* Use async storage in miner.
* Update rust version to 1.79.0.
* Use Vec to avoid stack overflow.
* Fix unused warning.
* Fix clippy.
* Fix test warning.
* Fix test.
* fmt.
* Use async storage in pruner.
* nit.
146 lines
4.4 KiB
Rust
146 lines
4.4 KiB
Rust
use rand::random;
|
|
use shared_types::{compute_padded_chunk_size, ChunkArray, Transaction, CHUNK_SIZE};
|
|
use std::{cmp, sync::Arc};
|
|
use storage::{
|
|
log_store::{
|
|
log_manager::{sub_merkle_tree, tx_subtree_root_list_padded, LogConfig, PORA_CHUNK_SIZE},
|
|
LogStoreChunkWrite, LogStoreWrite,
|
|
},
|
|
LogManager,
|
|
};
|
|
|
|
/// Creates stores for local node and peers with initialized transaction of specified chunk count.
|
|
/// The first store is for local node, and data not stored. The second store is for peers, and all
|
|
/// transactions are finalized for file sync.
|
|
#[allow(clippy::type_complexity)]
|
|
pub fn create_2_store(
|
|
chunk_count: Vec<usize>,
|
|
) -> (
|
|
Arc<LogManager>,
|
|
Arc<LogManager>,
|
|
Vec<Transaction>,
|
|
Vec<Vec<u8>>,
|
|
) {
|
|
let config = LogConfig::default();
|
|
let mut store = LogManager::memorydb(config.clone()).unwrap();
|
|
let mut peer_store = LogManager::memorydb(config).unwrap();
|
|
|
|
let mut offset = 1;
|
|
let mut txs = vec![];
|
|
let mut data = vec![];
|
|
|
|
for (tx_seq, chunks) in chunk_count.iter().enumerate() {
|
|
let ret = generate_data(*chunks, &mut store, &mut peer_store, tx_seq as u64, offset);
|
|
txs.push(ret.0);
|
|
data.push(ret.1);
|
|
offset = ret.2;
|
|
}
|
|
|
|
(Arc::new(store), Arc::new(peer_store), txs, data)
|
|
}
|
|
|
|
fn generate_data(
|
|
chunk_count: usize,
|
|
store: &mut LogManager,
|
|
peer_store: &mut LogManager,
|
|
seq: u64,
|
|
offset: u64,
|
|
) -> (Transaction, Vec<u8>, u64) {
|
|
let data_size = CHUNK_SIZE * chunk_count;
|
|
let mut data = vec![0u8; data_size];
|
|
|
|
for i in 0..chunk_count {
|
|
data[i * CHUNK_SIZE] = random();
|
|
}
|
|
|
|
let merkel_nodes = tx_subtree_root_list_padded(&data);
|
|
let first_tree_size = 1 << (merkel_nodes[0].0 - 1);
|
|
let start_offset = if offset % first_tree_size == 0 {
|
|
offset
|
|
} else {
|
|
(offset / first_tree_size + 1) * first_tree_size
|
|
};
|
|
|
|
let merkle = sub_merkle_tree(&data).unwrap();
|
|
let tx = Transaction {
|
|
stream_ids: vec![],
|
|
size: data_size as u64,
|
|
data_merkle_root: merkle.root().into(),
|
|
seq,
|
|
data: vec![],
|
|
start_entry_index: start_offset,
|
|
merkle_nodes: merkel_nodes,
|
|
};
|
|
store.put_tx(tx.clone()).unwrap();
|
|
peer_store.put_tx(tx.clone()).unwrap();
|
|
for start_index in (0..chunk_count).step_by(PORA_CHUNK_SIZE) {
|
|
let end = cmp::min((start_index + PORA_CHUNK_SIZE) * CHUNK_SIZE, data.len());
|
|
let chunk_array = ChunkArray {
|
|
data: data[start_index * CHUNK_SIZE..end].to_vec(),
|
|
start_index: start_index as u64,
|
|
};
|
|
peer_store.put_chunks(tx.seq, chunk_array.clone()).unwrap();
|
|
}
|
|
peer_store.finalize_tx(tx.seq).unwrap();
|
|
|
|
let (padded_chunk_count, _) = compute_padded_chunk_size(data_size);
|
|
(tx, data, start_offset + padded_chunk_count as u64)
|
|
}
|
|
|
|
#[cfg(test)]
|
|
pub mod tests {
|
|
use file_location_cache::{test_util::AnnounceFileBuilder, FileLocationCache};
|
|
use libp2p::PeerId;
|
|
use shared_types::TxID;
|
|
use std::sync::Arc;
|
|
use storage::config::ShardConfig;
|
|
use storage::{
|
|
log_store::{log_manager::LogConfig, Store as LogStore},
|
|
LogManager,
|
|
};
|
|
use storage_async::Store;
|
|
use task_executor::test_utils::TestRuntime;
|
|
|
|
pub struct TestStoreRuntime {
|
|
pub runtime: TestRuntime,
|
|
pub store: Store,
|
|
}
|
|
|
|
impl Default for TestStoreRuntime {
|
|
fn default() -> Self {
|
|
let store = Arc::new(Self::new_store());
|
|
Self::new(store)
|
|
}
|
|
}
|
|
|
|
impl TestStoreRuntime {
|
|
pub fn new_store() -> impl LogStore {
|
|
LogManager::memorydb(LogConfig::default()).unwrap()
|
|
}
|
|
|
|
pub fn new(store: Arc<dyn LogStore>) -> TestStoreRuntime {
|
|
let runtime = TestRuntime::default();
|
|
let executor = runtime.task_executor.clone();
|
|
Self {
|
|
runtime,
|
|
store: Store::new(store, executor),
|
|
}
|
|
}
|
|
}
|
|
|
|
pub fn create_file_location_cache(peer_id: PeerId, txs: Vec<TxID>) -> Arc<FileLocationCache> {
|
|
let cache = FileLocationCache::default();
|
|
|
|
for tx_id in txs {
|
|
let announcement = AnnounceFileBuilder::default()
|
|
.with_tx_id(tx_id)
|
|
.with_peer_id(peer_id)
|
|
.build();
|
|
cache.insert(announcement);
|
|
}
|
|
cache.insert_peer_config(peer_id, ShardConfig::default());
|
|
|
|
Arc::new(cache)
|
|
}
|
|
}
|