mirror of
https://github.com/0glabs/0g-storage-node.git
synced 2024-11-20 15:05:19 +00:00
feat: upload_segments (#21)
* feat: upload_segments * feat: max request body
This commit is contained in:
parent
8f328a9eab
commit
01c2dd1135
@ -213,6 +213,10 @@ impl MemoryChunkPool {
|
||||
|
||||
/// Updates the cached file info when log entry retrieved from blockchain.
|
||||
pub async fn update_file_info(&self, tx: &Transaction) -> Result<bool> {
|
||||
info!(
|
||||
"start to flush cached segments to log store. data root: {}, tx_seq:{}",
|
||||
tx.data_merkle_root, tx.seq
|
||||
);
|
||||
let maybe_file = self
|
||||
.inner
|
||||
.lock()
|
||||
@ -225,7 +229,7 @@ impl MemoryChunkPool {
|
||||
self.write_chunks(
|
||||
SegmentInfo {
|
||||
root: tx.data_merkle_root,
|
||||
seg_data: seg.data.clone(),
|
||||
seg_data: seg.data,
|
||||
seg_proof: proof,
|
||||
seg_index,
|
||||
chunks_per_segment: file.chunks_per_segment,
|
||||
@ -236,6 +240,10 @@ impl MemoryChunkPool {
|
||||
.await?
|
||||
}
|
||||
}
|
||||
info!(
|
||||
"cached segments flushed to log store. data root: {}, tx_seq:{}",
|
||||
tx.data_merkle_root, tx.seq
|
||||
);
|
||||
Ok(true)
|
||||
}
|
||||
|
||||
|
@ -6,5 +6,6 @@ pub struct Config {
|
||||
pub listen_address: SocketAddr,
|
||||
pub listen_address_admin: Option<SocketAddr>,
|
||||
pub chunks_per_segment: usize,
|
||||
pub max_request_body_size: u32,
|
||||
pub max_cache_file_size: usize,
|
||||
}
|
||||
|
@ -74,6 +74,10 @@ pub async fn run_server(
|
||||
Ok(handles)
|
||||
}
|
||||
|
||||
fn server_builder(ctx: Context) -> HttpServerBuilder {
|
||||
HttpServerBuilder::default().max_request_body_size(ctx.config.max_request_body_size)
|
||||
}
|
||||
|
||||
/// Run a single RPC server for all namespace RPCs.
|
||||
async fn run_server_all(ctx: Context) -> Result<HttpServerHandle, Box<dyn Error>> {
|
||||
// public rpc
|
||||
@ -89,7 +93,7 @@ async fn run_server_all(ctx: Context) -> Result<HttpServerHandle, Box<dyn Error>
|
||||
zgs.merge(mine)?;
|
||||
}
|
||||
|
||||
Ok(HttpServerBuilder::default()
|
||||
Ok(server_builder(ctx.clone())
|
||||
.build(ctx.config.listen_address)
|
||||
.await?
|
||||
.start(zgs)?)
|
||||
@ -112,12 +116,12 @@ async fn run_server_public_private(
|
||||
admin.merge(mine)?;
|
||||
}
|
||||
|
||||
let handle_public = HttpServerBuilder::default()
|
||||
let handle_public = server_builder(ctx.clone())
|
||||
.build(ctx.config.listen_address)
|
||||
.await?
|
||||
.start(zgs)?;
|
||||
|
||||
let handle_private = HttpServerBuilder::default()
|
||||
let handle_private = server_builder(ctx.clone())
|
||||
.build(listen_addr_private)
|
||||
.await?
|
||||
.start(admin)?;
|
||||
|
@ -11,6 +11,9 @@ pub trait Rpc {
|
||||
#[method(name = "uploadSegment")]
|
||||
async fn upload_segment(&self, segment: SegmentWithProof) -> RpcResult<()>;
|
||||
|
||||
#[method(name = "uploadSegments")]
|
||||
async fn upload_segments(&self, segments: Vec<SegmentWithProof>) -> RpcResult<()>;
|
||||
|
||||
#[method(name = "downloadSegment")]
|
||||
async fn download_segment(
|
||||
&self,
|
||||
|
@ -35,52 +35,14 @@ impl RpcServer for RpcServerImpl {
|
||||
|
||||
async fn upload_segment(&self, segment: SegmentWithProof) -> RpcResult<()> {
|
||||
debug!(root = %segment.root, index = %segment.index, "zgs_uploadSegment");
|
||||
self.put_segment(segment).await
|
||||
}
|
||||
|
||||
let _ = self.ctx.chunk_pool.validate_segment_size(&segment.data)?;
|
||||
|
||||
let maybe_tx = self
|
||||
.ctx
|
||||
.log_store
|
||||
.get_tx_by_data_root(&segment.root)
|
||||
.await?;
|
||||
let mut need_cache = false;
|
||||
|
||||
if self
|
||||
.ctx
|
||||
.chunk_pool
|
||||
.check_already_has_cache(&segment.root)
|
||||
.await
|
||||
{
|
||||
need_cache = true;
|
||||
async fn upload_segments(&self, segments: Vec<SegmentWithProof>) -> RpcResult<()> {
|
||||
debug!("zgs_uploadSegments()");
|
||||
for segment in segments.into_iter() {
|
||||
self.put_segment(segment).await?;
|
||||
}
|
||||
|
||||
if !need_cache {
|
||||
need_cache = self.check_need_cache(&maybe_tx, segment.file_size).await?;
|
||||
}
|
||||
|
||||
segment.validate(self.ctx.config.chunks_per_segment)?;
|
||||
|
||||
let seg_info = SegmentInfo {
|
||||
root: segment.root,
|
||||
seg_data: segment.data,
|
||||
seg_proof: segment.proof,
|
||||
seg_index: segment.index,
|
||||
chunks_per_segment: self.ctx.config.chunks_per_segment,
|
||||
};
|
||||
|
||||
if need_cache {
|
||||
self.ctx.chunk_pool.cache_chunks(seg_info).await?;
|
||||
} else {
|
||||
let file_id = FileID {
|
||||
root: seg_info.root,
|
||||
tx_id: maybe_tx.unwrap().id(),
|
||||
};
|
||||
self.ctx
|
||||
.chunk_pool
|
||||
.write_chunks(seg_info, file_id, segment.file_size)
|
||||
.await?;
|
||||
}
|
||||
|
||||
Ok(())
|
||||
}
|
||||
|
||||
@ -236,4 +198,54 @@ impl RpcServerImpl {
|
||||
uploaded_seg_num,
|
||||
})
|
||||
}
|
||||
|
||||
async fn put_segment(&self, segment: SegmentWithProof) -> RpcResult<()> {
|
||||
debug!(root = %segment.root, index = %segment.index, "putSegment");
|
||||
|
||||
self.ctx.chunk_pool.validate_segment_size(&segment.data)?;
|
||||
|
||||
let maybe_tx = self
|
||||
.ctx
|
||||
.log_store
|
||||
.get_tx_by_data_root(&segment.root)
|
||||
.await?;
|
||||
let mut need_cache = false;
|
||||
|
||||
if self
|
||||
.ctx
|
||||
.chunk_pool
|
||||
.check_already_has_cache(&segment.root)
|
||||
.await
|
||||
{
|
||||
need_cache = true;
|
||||
}
|
||||
|
||||
if !need_cache {
|
||||
need_cache = self.check_need_cache(&maybe_tx, segment.file_size).await?;
|
||||
}
|
||||
|
||||
segment.validate(self.ctx.config.chunks_per_segment)?;
|
||||
|
||||
let seg_info = SegmentInfo {
|
||||
root: segment.root,
|
||||
seg_data: segment.data,
|
||||
seg_proof: segment.proof,
|
||||
seg_index: segment.index,
|
||||
chunks_per_segment: self.ctx.config.chunks_per_segment,
|
||||
};
|
||||
|
||||
if need_cache {
|
||||
self.ctx.chunk_pool.cache_chunks(seg_info).await?;
|
||||
} else {
|
||||
let file_id = FileID {
|
||||
root: seg_info.root,
|
||||
tx_id: maybe_tx.unwrap().id(),
|
||||
};
|
||||
self.ctx
|
||||
.chunk_pool
|
||||
.write_chunks(seg_info, file_id, segment.file_size)
|
||||
.await?;
|
||||
}
|
||||
Ok(())
|
||||
}
|
||||
}
|
||||
|
@ -219,7 +219,7 @@ impl ClientBuilder {
|
||||
let async_store = require!("rpc", self, async_store).clone();
|
||||
let network_send = require!("rpc", self, network).send.clone();
|
||||
let mine_send = self.miner.as_ref().map(|x| x.send.clone());
|
||||
let synced_tx_recv = require!("sync", self, log_sync).send.subscribe();
|
||||
let synced_tx_recv = require!("rpc", self, log_sync).send.subscribe();
|
||||
|
||||
let (chunk_pool, chunk_pool_handler) =
|
||||
chunk_pool::unbounded(chunk_pool_config, async_store.clone(), network_send.clone());
|
||||
|
@ -77,6 +77,7 @@ impl ZgsConfig {
|
||||
enabled: self.rpc_enabled,
|
||||
listen_address,
|
||||
listen_address_admin,
|
||||
max_request_body_size: self.max_request_body_size,
|
||||
chunks_per_segment: self.rpc_chunks_per_segment,
|
||||
max_cache_file_size: self.rpc_max_cache_file_size,
|
||||
})
|
||||
|
@ -38,6 +38,7 @@ build_config! {
|
||||
(rpc_enabled, (bool), true)
|
||||
(rpc_listen_address, (String), "127.0.0.1:5678".to_string())
|
||||
(rpc_listen_address_admin, (String), "127.0.0.1:5679".to_string())
|
||||
(max_request_body_size, (u32), 100*1024*1024) // 100MB
|
||||
(rpc_chunks_per_segment, (usize), 1024)
|
||||
(rpc_max_cache_file_size, (usize), 10*1024*1024) //10MB
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user