mirror of
https://github.com/hl-archive-node/nanoreth.git
synced 2025-12-06 10:59:55 +00:00
225 lines
7.9 KiB
Rust
225 lines
7.9 KiB
Rust
use crate::{
|
|
dirs::{DataDirPath, MaybePlatformPath},
|
|
init::init_genesis,
|
|
node::events::{handle_events, NodeEvent},
|
|
version::SHORT_VERSION,
|
|
};
|
|
use clap::Parser;
|
|
use eyre::Context;
|
|
use futures::{Stream, StreamExt};
|
|
use reth_beacon_consensus::BeaconConsensus;
|
|
use reth_provider::{ProviderFactory, StageCheckpointReader};
|
|
|
|
use crate::args::{utils::genesis_value_parser, DatabaseArgs};
|
|
use reth_config::Config;
|
|
use reth_db::{database::Database, init_db};
|
|
use reth_downloaders::{
|
|
bodies::bodies::BodiesDownloaderBuilder,
|
|
headers::reverse_headers::ReverseHeadersDownloaderBuilder, test_utils::FileClient,
|
|
};
|
|
use reth_interfaces::consensus::Consensus;
|
|
use reth_primitives::{stage::StageId, ChainSpec, B256};
|
|
use reth_stages::{
|
|
prelude::*,
|
|
stages::{
|
|
ExecutionStage, ExecutionStageThresholds, HeaderSyncMode, SenderRecoveryStage,
|
|
TotalDifficultyStage,
|
|
},
|
|
};
|
|
use std::{path::PathBuf, sync::Arc};
|
|
use tokio::sync::watch;
|
|
use tracing::{debug, info};
|
|
|
|
/// Syncs RLP encoded blocks from a file.
|
|
#[derive(Debug, Parser)]
|
|
pub struct ImportCommand {
|
|
/// The path to the configuration file to use.
|
|
#[arg(long, value_name = "FILE", verbatim_doc_comment)]
|
|
config: Option<PathBuf>,
|
|
|
|
/// The path to the data dir for all reth files and subdirectories.
|
|
///
|
|
/// Defaults to the OS-specific data directory:
|
|
///
|
|
/// - Linux: `$XDG_DATA_HOME/reth/` or `$HOME/.local/share/reth/`
|
|
/// - Windows: `{FOLDERID_RoamingAppData}/reth/`
|
|
/// - macOS: `$HOME/Library/Application Support/reth/`
|
|
#[arg(long, value_name = "DATA_DIR", verbatim_doc_comment, default_value_t)]
|
|
datadir: MaybePlatformPath<DataDirPath>,
|
|
|
|
/// The chain this node is running.
|
|
///
|
|
/// Possible values are either a built-in chain or the path to a chain specification file.
|
|
///
|
|
/// Built-in chains:
|
|
/// - mainnet
|
|
/// - goerli
|
|
/// - sepolia
|
|
/// - holesky
|
|
#[arg(
|
|
long,
|
|
value_name = "CHAIN_OR_PATH",
|
|
verbatim_doc_comment,
|
|
default_value = "mainnet",
|
|
value_parser = genesis_value_parser
|
|
)]
|
|
chain: Arc<ChainSpec>,
|
|
|
|
#[clap(flatten)]
|
|
db: DatabaseArgs,
|
|
|
|
/// The path to a block file for import.
|
|
///
|
|
/// The online stages (headers and bodies) are replaced by a file import, after which the
|
|
/// remaining stages are executed.
|
|
#[arg(value_name = "IMPORT_PATH", verbatim_doc_comment)]
|
|
path: PathBuf,
|
|
}
|
|
|
|
impl ImportCommand {
|
|
/// Execute `import` command
|
|
pub async fn execute(self) -> eyre::Result<()> {
|
|
info!(target: "reth::cli", "reth {} starting", SHORT_VERSION);
|
|
|
|
// add network name to data dir
|
|
let data_dir = self.datadir.unwrap_or_chain_default(self.chain.chain);
|
|
let config_path = self.config.clone().unwrap_or(data_dir.config_path());
|
|
|
|
let config: Config = self.load_config(config_path.clone())?;
|
|
info!(target: "reth::cli", path = ?config_path, "Configuration loaded");
|
|
|
|
let db_path = data_dir.db_path();
|
|
|
|
info!(target: "reth::cli", path = ?db_path, "Opening database");
|
|
let db = Arc::new(init_db(db_path, self.db.log_level)?);
|
|
info!(target: "reth::cli", "Database opened");
|
|
|
|
debug!(target: "reth::cli", chain=%self.chain.chain, genesis=?self.chain.genesis_hash(), "Initializing genesis");
|
|
|
|
init_genesis(db.clone(), self.chain.clone())?;
|
|
|
|
let consensus = Arc::new(BeaconConsensus::new(self.chain.clone()));
|
|
info!(target: "reth::cli", "Consensus engine initialized");
|
|
|
|
// create a new FileClient
|
|
info!(target: "reth::cli", "Importing chain file");
|
|
let file_client = Arc::new(FileClient::new(&self.path).await?);
|
|
|
|
// override the tip
|
|
let tip = file_client.tip().expect("file client has no tip");
|
|
info!(target: "reth::cli", "Chain file imported");
|
|
|
|
let (mut pipeline, events) =
|
|
self.build_import_pipeline(config, Arc::clone(&db), &consensus, file_client).await?;
|
|
|
|
// override the tip
|
|
pipeline.set_tip(tip);
|
|
debug!(target: "reth::cli", ?tip, "Tip manually set");
|
|
|
|
let factory = ProviderFactory::new(&db, self.chain.clone());
|
|
let provider = factory.provider().map_err(PipelineError::Interface)?;
|
|
|
|
let latest_block_number =
|
|
provider.get_stage_checkpoint(StageId::Finish)?.map(|ch| ch.block_number);
|
|
tokio::spawn(handle_events(None, latest_block_number, events));
|
|
|
|
// Run pipeline
|
|
info!(target: "reth::cli", "Starting sync pipeline");
|
|
tokio::select! {
|
|
res = pipeline.run() => res?,
|
|
_ = tokio::signal::ctrl_c() => {},
|
|
};
|
|
|
|
info!(target: "reth::cli", "Finishing up");
|
|
Ok(())
|
|
}
|
|
|
|
async fn build_import_pipeline<DB, C>(
|
|
&self,
|
|
config: Config,
|
|
db: DB,
|
|
consensus: &Arc<C>,
|
|
file_client: Arc<FileClient>,
|
|
) -> eyre::Result<(Pipeline<DB>, impl Stream<Item = NodeEvent>)>
|
|
where
|
|
DB: Database + Clone + Unpin + 'static,
|
|
C: Consensus + 'static,
|
|
{
|
|
if !file_client.has_canonical_blocks() {
|
|
eyre::bail!("unable to import non canonical blocks");
|
|
}
|
|
|
|
let header_downloader = ReverseHeadersDownloaderBuilder::from(config.stages.headers)
|
|
.build(file_client.clone(), consensus.clone())
|
|
.into_task();
|
|
|
|
let body_downloader = BodiesDownloaderBuilder::from(config.stages.bodies)
|
|
.build(file_client.clone(), consensus.clone(), db.clone())
|
|
.into_task();
|
|
|
|
let (tip_tx, tip_rx) = watch::channel(B256::ZERO);
|
|
let factory = reth_revm::Factory::new(self.chain.clone());
|
|
|
|
let max_block = file_client.max_block().unwrap_or(0);
|
|
let mut pipeline = Pipeline::builder()
|
|
.with_tip_sender(tip_tx)
|
|
// we want to sync all blocks the file client provides or 0 if empty
|
|
.with_max_block(max_block)
|
|
.add_stages(
|
|
DefaultStages::new(
|
|
HeaderSyncMode::Tip(tip_rx),
|
|
consensus.clone(),
|
|
header_downloader,
|
|
body_downloader,
|
|
factory.clone(),
|
|
)
|
|
.set(
|
|
TotalDifficultyStage::new(consensus.clone())
|
|
.with_commit_threshold(config.stages.total_difficulty.commit_threshold),
|
|
)
|
|
.set(SenderRecoveryStage {
|
|
commit_threshold: config.stages.sender_recovery.commit_threshold,
|
|
})
|
|
.set(ExecutionStage::new(
|
|
factory,
|
|
ExecutionStageThresholds {
|
|
max_blocks: config.stages.execution.max_blocks,
|
|
max_changes: config.stages.execution.max_changes,
|
|
max_cumulative_gas: config.stages.execution.max_cumulative_gas,
|
|
},
|
|
config
|
|
.stages
|
|
.merkle
|
|
.clean_threshold
|
|
.max(config.stages.account_hashing.clean_threshold)
|
|
.max(config.stages.storage_hashing.clean_threshold),
|
|
config.prune.map(|prune| prune.segments).unwrap_or_default(),
|
|
)),
|
|
)
|
|
.build(db, self.chain.clone());
|
|
|
|
let events = pipeline.events().map(Into::into);
|
|
|
|
Ok((pipeline, events))
|
|
}
|
|
|
|
/// Loads the reth config
|
|
fn load_config(&self, config_path: PathBuf) -> eyre::Result<Config> {
|
|
confy::load_path::<Config>(config_path.clone())
|
|
.wrap_err_with(|| format!("Could not load config file {:?}", config_path))
|
|
}
|
|
}
|
|
|
|
#[cfg(test)]
|
|
mod tests {
|
|
use super::*;
|
|
|
|
#[test]
|
|
fn parse_common_import_command_chain_args() {
|
|
for chain in ["mainnet", "sepolia", "goerli"] {
|
|
let args: ImportCommand = ImportCommand::parse_from(["reth", "--chain", chain, "."]);
|
|
assert_eq!(args.chain.chain, chain.parse().unwrap());
|
|
}
|
|
}
|
|
}
|