Skip to main content

reth_cli_commands/
import_core.rs

1//! Core import functionality without CLI dependencies.
2
3use alloy_primitives::B256;
4use futures::StreamExt;
5use reth_config::Config;
6use reth_consensus::FullConsensus;
7use reth_db_api::{tables, transaction::DbTx};
8use reth_downloaders::{
9    bodies::bodies::BodiesDownloaderBuilder,
10    file_client::{ChunkedFileReader, FileClient, DEFAULT_BYTE_LEN_CHUNK_CHAIN_FILE},
11    headers::reverse_headers::ReverseHeadersDownloaderBuilder,
12};
13use reth_evm::ConfigureEvm;
14use reth_network_p2p::{
15    bodies::downloader::BodyDownloader,
16    headers::downloader::{HeaderDownloader, SyncTarget},
17};
18use reth_node_api::BlockTy;
19use reth_node_events::node::NodeEvent;
20use reth_provider::{
21    providers::ProviderNodeTypes, BlockNumReader, HeaderProvider, ProviderError, ProviderFactory,
22    StageCheckpointReader,
23};
24use reth_prune::PruneModes;
25use reth_stages::{prelude::*, ControlFlow, Pipeline, StageId, StageSet};
26use reth_static_file::StaticFileProducer;
27use std::{path::Path, sync::Arc};
28use tokio::sync::watch;
29use tracing::{debug, error, info, warn};
30
31/// Configuration for importing blocks from RLP files.
32#[derive(Debug, Clone, Default)]
33pub struct ImportConfig {
34    /// Disables stages that require state.
35    pub no_state: bool,
36    /// Chunk byte length to read from file.
37    pub chunk_len: Option<u64>,
38    /// If true, fail immediately when an invalid block is encountered.
39    /// By default (false), the import stops at the last valid block and exits successfully.
40    pub fail_on_invalid_block: bool,
41}
42
43/// Result of an import operation.
44#[derive(Debug)]
45pub struct ImportResult {
46    /// Total number of blocks decoded from the file.
47    pub total_decoded_blocks: usize,
48    /// Total number of transactions decoded from the file.
49    pub total_decoded_txns: usize,
50    /// Total number of blocks imported into the database.
51    pub total_imported_blocks: usize,
52    /// Total number of transactions imported into the database.
53    pub total_imported_txns: usize,
54    /// Whether the import was stopped due to an invalid block.
55    pub stopped_on_invalid_block: bool,
56    /// The block number that was invalid, if any.
57    pub bad_block: Option<u64>,
58    /// The last valid block number when stopped due to invalid block.
59    pub last_valid_block: Option<u64>,
60}
61
62impl ImportResult {
63    /// Returns true if all blocks and transactions were imported successfully.
64    pub fn is_complete(&self) -> bool {
65        self.total_decoded_blocks == self.total_imported_blocks &&
66            self.total_decoded_txns == self.total_imported_txns
67    }
68
69    /// Returns true if the import was successful, considering stop-on-invalid-block mode.
70    ///
71    /// In stop-on-invalid-block mode, a partial import is considered successful if we
72    /// stopped due to an invalid block (leaving the DB at the last valid block).
73    pub fn is_successful(&self) -> bool {
74        self.is_complete() || self.stopped_on_invalid_block
75    }
76}
77
78/// Imports blocks from an RLP-encoded file into the database.
79///
80/// This function reads RLP-encoded blocks from a file in chunks and imports them
81/// using the pipeline infrastructure. It's designed to be used both from the CLI
82/// and from test code.
83pub async fn import_blocks_from_file<N>(
84    path: &Path,
85    import_config: ImportConfig,
86    provider_factory: ProviderFactory<N>,
87    config: &Config,
88    executor: impl ConfigureEvm<Primitives = N::Primitives> + 'static,
89    consensus: Arc<impl FullConsensus<N::Primitives> + 'static>,
90) -> eyre::Result<ImportResult>
91where
92    N: ProviderNodeTypes,
93{
94    if import_config.no_state {
95        info!(target: "reth::import", "Disabled stages requiring state");
96    }
97
98    debug!(target: "reth::import",
99        chunk_byte_len=import_config.chunk_len.unwrap_or(DEFAULT_BYTE_LEN_CHUNK_CHAIN_FILE),
100        "Chunking chain import"
101    );
102
103    info!(target: "reth::import", "Consensus engine initialized");
104
105    // open file
106    let mut reader = ChunkedFileReader::new(path, import_config.chunk_len).await?;
107
108    let provider = provider_factory.provider()?;
109    let init_blocks = provider.tx_ref().entries::<tables::HeaderNumbers>()?;
110    let init_txns = provider.tx_ref().entries::<tables::TransactionHashNumbers>()?;
111    drop(provider);
112
113    let mut total_decoded_blocks = 0;
114    let mut total_decoded_txns = 0;
115
116    let mut sealed_header = provider_factory
117        .sealed_header(provider_factory.last_block_number()?)?
118        .expect("should have genesis");
119
120    let static_file_producer =
121        StaticFileProducer::new(provider_factory.clone(), PruneModes::default());
122
123    // Track if we stopped due to an invalid block
124    let mut stopped_on_invalid_block = false;
125    let mut bad_block_number: Option<u64> = None;
126    let mut last_valid_block_number: Option<u64> = None;
127
128    while let Some(file_client) =
129        reader.next_chunk::<BlockTy<N>>(consensus.clone(), Some(sealed_header)).await?
130    {
131        // create a new FileClient from chunk read from file
132        info!(target: "reth::import",
133            "Importing chain file chunk"
134        );
135
136        let tip = file_client.tip().ok_or(eyre::eyre!("file client has no tip"))?;
137        info!(target: "reth::import", "Chain file chunk read");
138
139        total_decoded_blocks += file_client.headers_len();
140        total_decoded_txns += file_client.total_transactions();
141
142        let (mut pipeline, events) = build_import_pipeline_impl(
143            config,
144            provider_factory.clone(),
145            &consensus,
146            Arc::new(file_client),
147            static_file_producer.clone(),
148            import_config.no_state,
149            executor.clone(),
150        )?;
151
152        // override the tip
153        pipeline.set_tip(tip);
154        debug!(target: "reth::import", ?tip, "Tip manually set");
155
156        let latest_block_number =
157            provider_factory.get_stage_checkpoint(StageId::Finish)?.map(|ch| ch.block_number);
158        tokio::spawn(reth_node_events::node::handle_events(None, latest_block_number, events));
159
160        // Run pipeline
161        info!(target: "reth::import", "Starting sync pipeline");
162        if import_config.fail_on_invalid_block {
163            // Original behavior: fail on unwind
164            tokio::select! {
165                res = pipeline.run() => res?,
166                _ = tokio::signal::ctrl_c() => {
167                    info!(target: "reth::import", "Import interrupted by user");
168                    break;
169                },
170            }
171        } else {
172            // Default behavior: Use run_loop() to handle unwinds gracefully
173            let result = tokio::select! {
174                res = pipeline.run_loop() => res,
175                _ = tokio::signal::ctrl_c() => {
176                    info!(target: "reth::import", "Import interrupted by user");
177                    break;
178                },
179            };
180
181            match result {
182                Ok(ControlFlow::Unwind { target, bad_block }) => {
183                    // An invalid block was encountered; stop at last valid block
184                    let bad = bad_block.block.number;
185                    warn!(
186                        target: "reth::import",
187                        bad_block = bad,
188                        last_valid_block = target,
189                        "Invalid block encountered during import; stopping at last valid block"
190                    );
191                    stopped_on_invalid_block = true;
192                    bad_block_number = Some(bad);
193                    last_valid_block_number = Some(target);
194                    break;
195                }
196                Ok(ControlFlow::Continue { block_number }) => {
197                    debug!(target: "reth::import", block_number, "Pipeline chunk completed");
198                }
199                Ok(ControlFlow::NoProgress { block_number }) => {
200                    debug!(target: "reth::import", ?block_number, "Pipeline made no progress");
201                }
202                Err(e) => {
203                    // Propagate other pipeline errors
204                    return Err(e.into());
205                }
206            }
207        }
208
209        sealed_header = provider_factory
210            .sealed_header(provider_factory.last_block_number()?)?
211            .expect("should have genesis");
212    }
213
214    let provider = provider_factory.provider()?;
215    let total_imported_blocks = provider.tx_ref().entries::<tables::HeaderNumbers>()? - init_blocks;
216    let total_imported_txns =
217        provider.tx_ref().entries::<tables::TransactionHashNumbers>()? - init_txns;
218
219    let result = ImportResult {
220        total_decoded_blocks,
221        total_decoded_txns,
222        total_imported_blocks,
223        total_imported_txns,
224        stopped_on_invalid_block,
225        bad_block: bad_block_number,
226        last_valid_block: last_valid_block_number,
227    };
228
229    if result.stopped_on_invalid_block {
230        info!(target: "reth::import",
231            total_imported_blocks,
232            total_imported_txns,
233            bad_block = ?result.bad_block,
234            last_valid_block = ?result.last_valid_block,
235            "Import stopped at last valid block due to invalid block"
236        );
237    } else if !result.is_complete() {
238        error!(target: "reth::import",
239            total_decoded_blocks,
240            total_imported_blocks,
241            total_decoded_txns,
242            total_imported_txns,
243            "Chain was partially imported"
244        );
245    } else {
246        info!(target: "reth::import",
247            total_imported_blocks,
248            total_imported_txns,
249            "Chain was fully imported"
250        );
251    }
252
253    Ok(result)
254}
255
256/// Builds import pipeline.
257///
258/// If configured to execute, all stages will run. Otherwise, only stages that don't require state
259/// will run.
260pub fn build_import_pipeline_impl<N, C, E>(
261    config: &Config,
262    provider_factory: ProviderFactory<N>,
263    consensus: &Arc<C>,
264    file_client: Arc<FileClient<BlockTy<N>>>,
265    static_file_producer: StaticFileProducer<ProviderFactory<N>>,
266    disable_exec: bool,
267    evm_config: E,
268) -> eyre::Result<(Pipeline<N>, impl futures::Stream<Item = NodeEvent<N::Primitives>> + use<N, C, E>)>
269where
270    N: ProviderNodeTypes,
271    C: FullConsensus<N::Primitives> + 'static,
272    E: ConfigureEvm<Primitives = N::Primitives> + 'static,
273{
274    if !file_client.has_canonical_blocks() {
275        eyre::bail!("unable to import non canonical blocks");
276    }
277
278    // Retrieve latest header found in the database.
279    let last_block_number = provider_factory.last_block_number()?;
280    let local_head = provider_factory
281        .sealed_header(last_block_number)?
282        .ok_or_else(|| ProviderError::HeaderNotFound(last_block_number.into()))?;
283
284    let mut header_downloader = ReverseHeadersDownloaderBuilder::new(config.stages.headers)
285        .build(file_client.clone(), consensus.clone())
286        .into_task();
287    // TODO: The pipeline should correctly configure the downloader on its own.
288    // Find the possibility to remove unnecessary pre-configuration.
289    header_downloader.update_local_head(local_head);
290    header_downloader.update_sync_target(SyncTarget::Tip(file_client.tip().unwrap()));
291
292    let mut body_downloader = BodiesDownloaderBuilder::new(config.stages.bodies)
293        .build(file_client.clone(), consensus.clone(), provider_factory.clone())
294        .into_task();
295    // TODO: The pipeline should correctly configure the downloader on its own.
296    // Find the possibility to remove unnecessary pre-configuration.
297    body_downloader
298        .set_download_range(file_client.min_block().unwrap()..=file_client.max_block().unwrap())
299        .expect("failed to set download range");
300
301    let (tip_tx, tip_rx) = watch::channel(B256::ZERO);
302
303    let max_block = file_client.max_block().unwrap_or(0);
304
305    let pipeline = Pipeline::builder()
306        .with_tip_sender(tip_tx)
307        // we want to sync all blocks the file client provides or 0 if empty
308        .with_max_block(max_block)
309        .with_fail_on_unwind(true)
310        .add_stages(
311            DefaultStages::new(
312                provider_factory.clone(),
313                tip_rx,
314                consensus.clone(),
315                header_downloader,
316                body_downloader,
317                evm_config,
318                config.stages.clone(),
319                PruneModes::default(),
320                None,
321            )
322            .builder()
323            .disable_all_if(&StageId::STATE_REQUIRED, || disable_exec),
324        )
325        .build(provider_factory, static_file_producer);
326
327    let events = pipeline.events().map(Into::into);
328
329    Ok((pipeline, events))
330}