From aef1bcc4359c2db80b116cf6198369122b1adebe Mon Sep 17 00:00:00 2001 From: Roman Krasiuk Date: Fri, 3 May 2024 12:18:41 +0200 Subject: [PATCH] chore: make clippy happy (#8068) --- bin/reth/src/commands/db/mod.rs | 2 +- bin/reth/src/utils.rs | 4 ++-- crates/blockchain-tree/src/blockchain_tree.rs | 4 ++-- crates/consensus/auto-seal/src/lib.rs | 2 +- crates/consensus/beacon/src/engine/hooks/controller.rs | 2 +- crates/consensus/beacon/src/engine/mod.rs | 4 ++-- crates/exex/src/manager.rs | 2 +- crates/metrics/src/common/mpsc.rs | 2 +- crates/net/discv4/src/lib.rs | 4 ++-- crates/net/discv4/src/test_utils.rs | 2 +- crates/net/discv5/src/lib.rs | 6 +++--- crates/net/dns/src/lib.rs | 4 ++-- crates/net/downloaders/src/bodies/bodies.rs | 2 +- crates/net/downloaders/src/file_client.rs | 2 +- crates/net/downloaders/src/headers/reverse_headers.rs | 4 ++-- crates/net/eth-wire/src/multiplex.rs | 2 +- crates/net/eth-wire/src/muxdemux.rs | 2 +- crates/net/network/src/budget.rs | 1 + crates/net/network/src/eth_requests.rs | 6 +++--- crates/net/network/src/fetch/mod.rs | 2 +- crates/net/network/src/manager.rs | 2 +- crates/net/network/src/state.rs | 2 +- crates/net/network/src/transactions/fetcher.rs | 6 +++--- crates/node/events/src/node.rs | 2 +- crates/rpc/rpc-builder/src/lib.rs | 6 +++--- crates/rpc/rpc/src/eth/api/sign.rs | 2 +- crates/stages-api/src/pipeline/mod.rs | 2 +- crates/stages/src/stages/merkle.rs | 2 +- crates/storage/libmdbx-rs/src/lib.rs | 2 +- crates/storage/nippy-jar/src/lib.rs | 10 +++++----- crates/storage/nippy-jar/src/writer.rs | 2 +- .../provider/src/providers/static_file/writer.rs | 2 +- crates/storage/provider/src/test_utils/events.rs | 4 ++-- crates/transaction-pool/src/pool/txpool.rs | 4 ++-- 34 files changed, 54 insertions(+), 53 deletions(-) diff --git a/bin/reth/src/commands/db/mod.rs b/bin/reth/src/commands/db/mod.rs index aeaf1d7e8..6eedabcc7 100644 --- a/bin/reth/src/commands/db/mod.rs +++ b/bin/reth/src/commands/db/mod.rs @@ -159,7 +159,7 @@ impl Command { let provider_factory = ProviderFactory::new(db, self.chain.clone(), static_files_path.clone())?; - let mut tool = DbTool::new(provider_factory, self.chain.clone())?; + let tool = DbTool::new(provider_factory, self.chain.clone())?; tool.drop(db_path, static_files_path)?; } Subcommands::Clear(command) => { diff --git a/bin/reth/src/utils.rs b/bin/reth/src/utils.rs index 5c56476a8..650fc9d70 100644 --- a/bin/reth/src/utils.rs +++ b/bin/reth/src/utils.rs @@ -132,7 +132,7 @@ impl DbTool { /// Drops the database and the static files at the given path. pub fn drop( - &mut self, + &self, db_path: impl AsRef, static_files_path: impl AsRef, ) -> Result<()> { @@ -149,7 +149,7 @@ impl DbTool { } /// Drops the provided table from the database. - pub fn drop_table(&mut self) -> Result<()> { + pub fn drop_table(&self) -> Result<()> { self.provider_factory.db_ref().update(|tx| tx.clear::())??; Ok(()) } diff --git a/crates/blockchain-tree/src/blockchain_tree.rs b/crates/blockchain-tree/src/blockchain_tree.rs index fc9e7685a..b2b30f132 100644 --- a/crates/blockchain-tree/src/blockchain_tree.rs +++ b/crates/blockchain-tree/src/blockchain_tree.rs @@ -1216,7 +1216,7 @@ where /// /// The block, `revert_until`, is __non-inclusive__, i.e. `revert_until` stays in the database. fn revert_canonical_from_database( - &mut self, + &self, revert_until: BlockNumber, ) -> Result, CanonicalError> { // read data that is needed for new sidechain @@ -1239,7 +1239,7 @@ where } } - fn update_reorg_metrics(&mut self, reorg_depth: f64) { + fn update_reorg_metrics(&self, reorg_depth: f64) { self.metrics.reorgs.increment(1); self.metrics.latest_reorg_depth.set(reorg_depth); } diff --git a/crates/consensus/auto-seal/src/lib.rs b/crates/consensus/auto-seal/src/lib.rs index 402a6c983..9f2f2c402 100644 --- a/crates/consensus/auto-seal/src/lib.rs +++ b/crates/consensus/auto-seal/src/lib.rs @@ -338,7 +338,7 @@ impl StorageInner { /// /// This returns the poststate from execution and post-block changes, as well as the gas used. pub(crate) fn execute( - &mut self, + &self, block: &BlockWithSenders, executor: &mut EVMProcessor<'_, EvmConfig>, ) -> Result<(BundleStateWithReceipts, u64), BlockExecutionError> diff --git a/crates/consensus/beacon/src/engine/hooks/controller.rs b/crates/consensus/beacon/src/engine/hooks/controller.rs index 48343d480..47085be00 100644 --- a/crates/consensus/beacon/src/engine/hooks/controller.rs +++ b/crates/consensus/beacon/src/engine/hooks/controller.rs @@ -124,7 +124,7 @@ impl EngineHooksController { } fn poll_next_hook_inner( - &mut self, + &self, cx: &mut Context<'_>, hook: &mut Box, args: EngineHookContext, diff --git a/crates/consensus/beacon/src/engine/mod.rs b/crates/consensus/beacon/src/engine/mod.rs index f3aa249fa..3e12c5f8e 100644 --- a/crates/consensus/beacon/src/engine/mod.rs +++ b/crates/consensus/beacon/src/engine/mod.rs @@ -446,7 +446,7 @@ where /// /// Returns `true` if the head needs to be updated. fn on_head_already_canonical( - &mut self, + &self, header: &SealedHeader, attrs: &mut Option, ) -> bool { @@ -804,7 +804,7 @@ where /// This also updates the safe and finalized blocks in the [CanonChainTracker], if they are /// consistent with the head block. fn ensure_consistent_forkchoice_state( - &mut self, + &self, state: ForkchoiceState, ) -> ProviderResult> { // Ensure that the finalized block, if not zero, is known and in the canonical chain diff --git a/crates/exex/src/manager.rs b/crates/exex/src/manager.rs index 1037395b8..1de8c102e 100644 --- a/crates/exex/src/manager.rs +++ b/crates/exex/src/manager.rs @@ -258,7 +258,7 @@ impl ExExManager { /// Updates the current buffer capacity and notifies all `is_ready` watchers of the manager's /// readiness to receive notifications. - fn update_capacity(&mut self) { + fn update_capacity(&self) { let capacity = self.max_capacity.saturating_sub(self.buffer.len()); self.current_capacity.store(capacity, Ordering::Relaxed); self.metrics.current_capacity.set(capacity as f64); diff --git a/crates/metrics/src/common/mpsc.rs b/crates/metrics/src/common/mpsc.rs index 3c35c745e..98c670ef7 100644 --- a/crates/metrics/src/common/mpsc.rs +++ b/crates/metrics/src/common/mpsc.rs @@ -173,7 +173,7 @@ impl MeteredSender { /// Calls the underlying [Sender](mpsc::Sender)'s `send`, incrementing the appropriate /// metrics depending on the result. - pub async fn send(&mut self, value: T) -> Result<(), SendError> { + pub async fn send(&self, value: T) -> Result<(), SendError> { match self.sender.send(value).await { Ok(()) => { self.metrics.messages_sent.increment(1); diff --git a/crates/net/discv4/src/lib.rs b/crates/net/discv4/src/lib.rs index 1a942a5b9..9a0cb9c11 100644 --- a/crates/net/discv4/src/lib.rs +++ b/crates/net/discv4/src/lib.rs @@ -992,7 +992,7 @@ impl Discv4Service { } /// Encodes the packet, sends it and returns the hash. - pub(crate) fn send_packet(&mut self, msg: Message, to: SocketAddr) -> B256 { + pub(crate) fn send_packet(&self, msg: Message, to: SocketAddr) -> B256 { let (payload, hash) = msg.encode(&self.secret_key); trace!(target: "discv4", r#type=?msg.msg_type(), ?to, ?hash, "sending packet"); let _ = self.egress.try_send((payload, to)).map_err(|err| { @@ -1277,7 +1277,7 @@ impl Discv4Service { /// Handler for incoming `EnrRequest` message fn on_enr_request( - &mut self, + &self, msg: EnrRequest, remote_addr: SocketAddr, id: PeerId, diff --git a/crates/net/discv4/src/test_utils.rs b/crates/net/discv4/src/test_utils.rs index dae3ea388..d4930f204 100644 --- a/crates/net/discv4/src/test_utils.rs +++ b/crates/net/discv4/src/test_utils.rs @@ -114,7 +114,7 @@ impl MockDiscovery { } /// Encodes the packet, sends it and returns the hash. - fn send_packet(&mut self, msg: Message, to: SocketAddr) -> B256 { + fn send_packet(&self, msg: Message, to: SocketAddr) -> B256 { let (payload, hash) = msg.encode(&self.secret_key); let _ = self.egress.try_send((payload, to)); hash diff --git a/crates/net/discv5/src/lib.rs b/crates/net/discv5/src/lib.rs index 8895f8a16..e9bc79dce 100644 --- a/crates/net/discv5/src/lib.rs +++ b/crates/net/discv5/src/lib.rs @@ -220,7 +220,7 @@ impl Discv5 { } /// Process an event from the underlying [`discv5::Discv5`] node. - pub fn on_discv5_update(&mut self, update: discv5::Event) -> Option { + pub fn on_discv5_update(&self, update: discv5::Event) -> Option { match update { discv5::Event::SocketUpdated(_) | discv5::Event::TalkRequest(_) | // `Discovered` not unique discovered peers @@ -254,7 +254,7 @@ impl Discv5 { /// Processes a discovered peer. Returns `true` if peer is added to pub fn on_discovered_peer( - &mut self, + &self, enr: &discv5::Enr, socket: SocketAddr, ) -> Option { @@ -724,7 +724,7 @@ mod tests { let remote_key = CombinedKey::generate_secp256k1(); let remote_enr = Enr::builder().tcp4(REMOTE_RLPX_PORT).build(&remote_key).unwrap(); - let mut discv5 = discv5_noop(); + let discv5 = discv5_noop(); // test let filtered_peer = discv5.on_discovered_peer(&remote_enr, remote_socket); diff --git a/crates/net/dns/src/lib.rs b/crates/net/dns/src/lib.rs index e5ddc0fd1..5000e524e 100644 --- a/crates/net/dns/src/lib.rs +++ b/crates/net/dns/src/lib.rs @@ -67,13 +67,13 @@ pub struct DnsDiscoveryHandle { impl DnsDiscoveryHandle { /// Starts syncing the given link to a tree. - pub fn sync_tree(&mut self, link: &str) -> Result<(), ParseDnsEntryError> { + pub fn sync_tree(&self, link: &str) -> Result<(), ParseDnsEntryError> { self.sync_tree_with_link(link.parse()?); Ok(()) } /// Starts syncing the given link to a tree. - pub fn sync_tree_with_link(&mut self, link: LinkEntry) { + pub fn sync_tree_with_link(&self, link: LinkEntry) { let _ = self.to_service.send(DnsDiscoveryCommand::SyncTree(link)); } diff --git a/crates/net/downloaders/src/bodies/bodies.rs b/crates/net/downloaders/src/bodies/bodies.rs index 985c545e9..8f97e09c7 100644 --- a/crates/net/downloaders/src/bodies/bodies.rs +++ b/crates/net/downloaders/src/bodies/bodies.rs @@ -68,7 +68,7 @@ where Provider: HeaderProvider + Unpin + 'static, { /// Returns the next contiguous request. - fn next_headers_request(&mut self) -> DownloadResult>> { + fn next_headers_request(&self) -> DownloadResult>> { let start_at = match self.in_progress_queue.last_requested_block_number { Some(num) => num + 1, None => *self.download_range.start(), diff --git a/crates/net/downloaders/src/file_client.rs b/crates/net/downloaders/src/file_client.rs index ee783a1a4..ef72a891b 100644 --- a/crates/net/downloaders/src/file_client.rs +++ b/crates/net/downloaders/src/file_client.rs @@ -224,7 +224,7 @@ impl FileClient { } /// Returns an iterator over headers in the client. - pub fn headers_iter(&mut self) -> impl Iterator { + pub fn headers_iter(&self) -> impl Iterator { self.headers.values() } diff --git a/crates/net/downloaders/src/headers/reverse_headers.rs b/crates/net/downloaders/src/headers/reverse_headers.rs index 273f97e58..a5cdb145b 100644 --- a/crates/net/downloaders/src/headers/reverse_headers.rs +++ b/crates/net/downloaders/src/headers/reverse_headers.rs @@ -536,7 +536,7 @@ where /// Handles the error of a bad response /// /// This will re-submit the request. - fn on_headers_error(&mut self, err: Box) { + fn on_headers_error(&self, err: Box) { let HeadersResponseError { request, peer_id, error } = *err; self.penalize_peer(peer_id, &error); @@ -581,7 +581,7 @@ where } /// Starts a request future - fn submit_request(&mut self, request: HeadersRequest, priority: Priority) { + fn submit_request(&self, request: HeadersRequest, priority: Priority) { trace!(target: "downloaders::headers", ?request, "Submitting headers request"); self.in_progress_queue.push(self.request_fut(request, priority)); self.metrics.in_flight_requests.increment(1.); diff --git a/crates/net/eth-wire/src/multiplex.rs b/crates/net/eth-wire/src/multiplex.rs index 8677ae77c..82eccd5c8 100644 --- a/crates/net/eth-wire/src/multiplex.rs +++ b/crates/net/eth-wire/src/multiplex.rs @@ -239,7 +239,7 @@ impl MultiplexInner { } /// Delegates a message to the matching protocol. - fn delegate_message(&mut self, cap: &SharedCapability, msg: BytesMut) -> bool { + fn delegate_message(&self, cap: &SharedCapability, msg: BytesMut) -> bool { for proto in &self.protocols { if proto.shared_cap == *cap { proto.send_raw(msg); diff --git a/crates/net/eth-wire/src/muxdemux.rs b/crates/net/eth-wire/src/muxdemux.rs index a9bbe2fdb..18112346e 100644 --- a/crates/net/eth-wire/src/muxdemux.rs +++ b/crates/net/eth-wire/src/muxdemux.rs @@ -171,7 +171,7 @@ impl MuxDemuxStream { /// Checks if all clones of this shared stream have been dropped, if true then returns // /// function to drop the stream. - fn can_drop(&mut self) -> bool { + fn can_drop(&self) -> bool { for tx in self.demux.values() { if !tx.is_closed() { return false diff --git a/crates/net/network/src/budget.rs b/crates/net/network/src/budget.rs index 319c8e311..e20d882fe 100644 --- a/crates/net/network/src/budget.rs +++ b/crates/net/network/src/budget.rs @@ -46,6 +46,7 @@ macro_rules! poll_nested_stream_with_budget { loop { match $poll_stream { Poll::Ready(Some(item)) => { + #[allow(unused_mut)] let mut f = $on_ready_some; f(item); diff --git a/crates/net/network/src/eth_requests.rs b/crates/net/network/src/eth_requests.rs index 57e83391d..3268ff898 100644 --- a/crates/net/network/src/eth_requests.rs +++ b/crates/net/network/src/eth_requests.rs @@ -139,7 +139,7 @@ where } fn on_headers_request( - &mut self, + &self, _peer_id: PeerId, request: GetBlockHeaders, response: oneshot::Sender>, @@ -150,7 +150,7 @@ where } fn on_bodies_request( - &mut self, + &self, _peer_id: PeerId, request: GetBlockBodies, response: oneshot::Sender>, @@ -187,7 +187,7 @@ where } fn on_receipts_request( - &mut self, + &self, _peer_id: PeerId, request: GetReceipts, response: oneshot::Sender>, diff --git a/crates/net/network/src/fetch/mod.rs b/crates/net/network/src/fetch/mod.rs index 3a529c97e..9ad50edb0 100644 --- a/crates/net/network/src/fetch/mod.rs +++ b/crates/net/network/src/fetch/mod.rs @@ -130,7 +130,7 @@ impl StateFetcher { /// Returns the _next_ idle peer that's ready to accept a request, /// prioritizing those with the lowest timeout/latency and those that recently responded with /// adequate data. - fn next_best_peer(&mut self) -> Option { + fn next_best_peer(&self) -> Option { let mut idle = self.peers.iter().filter(|(_, peer)| peer.state.is_idle()); let mut best_peer = idle.next()?; diff --git a/crates/net/network/src/manager.rs b/crates/net/network/src/manager.rs index 39d29ee71..0d2a33408 100644 --- a/crates/net/network/src/manager.rs +++ b/crates/net/network/src/manager.rs @@ -403,7 +403,7 @@ where } /// Handle an incoming request from the peer - fn on_eth_request(&mut self, peer_id: PeerId, req: PeerRequest) { + fn on_eth_request(&self, peer_id: PeerId, req: PeerRequest) { match req { PeerRequest::GetBlockHeaders { request, response } => { self.delegate_eth_request(IncomingEthRequest::GetBlockHeaders { diff --git a/crates/net/network/src/state.rs b/crates/net/network/src/state.rs index 0020b4927..309184ca3 100644 --- a/crates/net/network/src/state.rs +++ b/crates/net/network/src/state.rs @@ -234,7 +234,7 @@ where } /// Invoked when a new [`ForkId`] is activated. - pub(crate) fn update_fork_id(&mut self, fork_id: ForkId) { + pub(crate) fn update_fork_id(&self, fork_id: ForkId) { self.discovery.update_fork_id(fork_id) } diff --git a/crates/net/network/src/transactions/fetcher.rs b/crates/net/network/src/transactions/fetcher.rs index e82a20a31..f26b1abe2 100644 --- a/crates/net/network/src/transactions/fetcher.rs +++ b/crates/net/network/src/transactions/fetcher.rs @@ -239,7 +239,7 @@ impl TransactionFetcher { /// /// Returns left over hashes. pub fn pack_request( - &mut self, + &self, hashes_to_request: &mut RequestTxHashes, hashes_from_announcement: ValidAnnouncementData, ) -> RequestTxHashes { @@ -260,7 +260,7 @@ impl TransactionFetcher { /// response. If no, it's added to surplus hashes. If yes, it's added to hashes to the request /// and expected response size is accumulated. pub fn pack_request_eth68( - &mut self, + &self, hashes_to_request: &mut RequestTxHashes, hashes_from_announcement: impl HandleMempoolData + IntoIterator)>, @@ -328,7 +328,7 @@ impl TransactionFetcher { /// /// Returns left over hashes. pub fn pack_request_eth66( - &mut self, + &self, hashes_to_request: &mut RequestTxHashes, hashes_from_announcement: ValidAnnouncementData, ) -> RequestTxHashes { diff --git a/crates/node/events/src/node.rs b/crates/node/events/src/node.rs index 2689226ea..ba7ae8da4 100644 --- a/crates/node/events/src/node.rs +++ b/crates/node/events/src/node.rs @@ -232,7 +232,7 @@ impl NodeState { } } - fn handle_network_event(&mut self, _: NetworkEvent) { + fn handle_network_event(&self, _: NetworkEvent) { // NOTE(onbjerg): This used to log established/disconnecting sessions, but this is already // logged in the networking component. I kept this stub in case we want to catch other // networking events later on. diff --git a/crates/rpc/rpc-builder/src/lib.rs b/crates/rpc/rpc-builder/src/lib.rs index cea80398e..7d86a0056 100644 --- a/crates/rpc/rpc-builder/src/lib.rs +++ b/crates/rpc/rpc-builder/src/lib.rs @@ -1041,12 +1041,12 @@ where Network: NetworkInfo + Peers + Clone + 'static, { /// Instantiates AdminApi - pub fn admin_api(&mut self) -> AdminApi { + pub fn admin_api(&self) -> AdminApi { AdminApi::new(self.network.clone(), self.provider.chain_spec()) } /// Instantiates Web3Api - pub fn web3_api(&mut self) -> Web3Api { + pub fn web3_api(&self) -> Web3Api { Web3Api::new(self.network.clone()) } @@ -1443,7 +1443,7 @@ where } /// Instantiates RethApi - pub fn reth_api(&mut self) -> RethApi { + pub fn reth_api(&self) -> RethApi { RethApi::new(self.provider.clone(), Box::new(self.executor.clone())) } } diff --git a/crates/rpc/rpc/src/eth/api/sign.rs b/crates/rpc/rpc/src/eth/api/sign.rs index 66df0e8de..5cbdefa41 100644 --- a/crates/rpc/rpc/src/eth/api/sign.rs +++ b/crates/rpc/rpc/src/eth/api/sign.rs @@ -42,7 +42,7 @@ impl EthApi RethResult<()> { + pub fn produce_static_files(&self) -> RethResult<()> { let mut static_file_producer = self.static_file_producer.lock(); let provider = self.provider_factory.provider()?; diff --git a/crates/stages/src/stages/merkle.rs b/crates/stages/src/stages/merkle.rs index 186382e36..77fcf2e15 100644 --- a/crates/stages/src/stages/merkle.rs +++ b/crates/stages/src/stages/merkle.rs @@ -117,7 +117,7 @@ impl MerkleStage { /// Saves the hashing progress pub fn save_execution_checkpoint( - &mut self, + &self, provider: &DatabaseProviderRW, checkpoint: Option, ) -> Result<(), StageError> { diff --git a/crates/storage/libmdbx-rs/src/lib.rs b/crates/storage/libmdbx-rs/src/lib.rs index f8c251208..ba8c6b062 100644 --- a/crates/storage/libmdbx-rs/src/lib.rs +++ b/crates/storage/libmdbx-rs/src/lib.rs @@ -5,7 +5,7 @@ issue_tracker_base_url = "https://github.com/paradigmxyz/reth/issues/" )] #![cfg_attr(not(test), warn(unused_crate_dependencies))] -#![allow(missing_docs)] +#![allow(missing_docs, clippy::needless_pass_by_ref_mut)] #![cfg_attr(docsrs, feature(doc_cfg, doc_auto_cfg))] pub use crate::{ diff --git a/crates/storage/nippy-jar/src/lib.rs b/crates/storage/nippy-jar/src/lib.rs index 1cecdba40..1abbfba75 100644 --- a/crates/storage/nippy-jar/src/lib.rs +++ b/crates/storage/nippy-jar/src/lib.rs @@ -366,7 +366,7 @@ impl NippyJar { /// Writes all data and configuration to a file and the offset index to another. pub fn freeze( - mut self, + self, columns: Vec>>>, total_rows: u64, ) -> Result { @@ -392,7 +392,7 @@ impl NippyJar { } /// Freezes [`PerfectHashingFunction`], [`InclusionFilter`] and the offset index to file. - fn freeze_filters(&mut self) -> Result<(), NippyJarError> { + fn freeze_filters(&self) -> Result<(), NippyJarError> { debug!(target: "nippy-jar", path=?self.index_path(), "Writing offsets and offsets index to file."); let mut file = File::create(self.index_path())?; @@ -405,7 +405,7 @@ impl NippyJar { /// Safety checks before creating and returning a [`File`] handle to write data to. fn check_before_freeze( - &mut self, + &self, columns: &[impl IntoIterator>>], ) -> Result<(), NippyJarError> { if columns.len() != self.columns { @@ -427,7 +427,7 @@ impl NippyJar { } /// Writes all necessary configuration to file. - fn freeze_config(&mut self) -> Result<(), NippyJarError> { + fn freeze_config(&self) -> Result<(), NippyJarError> { Ok(bincode::serialize_into(File::create(self.config_path())?, &self)?) } } @@ -1200,7 +1200,7 @@ mod tests { fn append_two_rows(num_columns: usize, file_path: &Path, col1: &[Vec], col2: &[Vec]) { // Create and add 1 row { - let mut nippy = NippyJar::new_without_header(num_columns, file_path); + let nippy = NippyJar::new_without_header(num_columns, file_path); nippy.freeze_config().unwrap(); assert_eq!(nippy.max_row_size, 0); assert_eq!(nippy.rows, 0); diff --git a/crates/storage/nippy-jar/src/writer.rs b/crates/storage/nippy-jar/src/writer.rs index e1f4af10b..6417e6007 100644 --- a/crates/storage/nippy-jar/src/writer.rs +++ b/crates/storage/nippy-jar/src/writer.rs @@ -43,7 +43,7 @@ pub struct NippyJarWriter { impl NippyJarWriter { /// Creates a [`NippyJarWriter`] from [`NippyJar`]. - pub fn new(mut jar: NippyJar) -> Result { + pub fn new(jar: NippyJar) -> Result { let (data_file, offsets_file, is_created) = Self::create_or_open_files(jar.data_path(), &jar.offsets_path())?; diff --git a/crates/storage/provider/src/providers/static_file/writer.rs b/crates/storage/provider/src/providers/static_file/writer.rs index c61736b5e..d1aa8560f 100644 --- a/crates/storage/provider/src/providers/static_file/writer.rs +++ b/crates/storage/provider/src/providers/static_file/writer.rs @@ -225,7 +225,7 @@ impl StaticFileProviderRW { /// Verifies if the incoming block number matches the next expected block number /// for a static file. This ensures data continuity when adding new blocks. fn check_next_block_number( - &mut self, + &self, expected_block_number: u64, segment: StaticFileSegment, ) -> ProviderResult<()> { diff --git a/crates/storage/provider/src/test_utils/events.rs b/crates/storage/provider/src/test_utils/events.rs index 34c426661..baa6bc470 100644 --- a/crates/storage/provider/src/test_utils/events.rs +++ b/crates/storage/provider/src/test_utils/events.rs @@ -12,14 +12,14 @@ pub struct TestCanonStateSubscriptions { impl TestCanonStateSubscriptions { /// Adds new block commit to the queue that can be consumed with /// [`TestCanonStateSubscriptions::subscribe_to_canonical_state`] - pub fn add_next_commit(&mut self, new: Arc) { + pub fn add_next_commit(&self, new: Arc) { let event = CanonStateNotification::Commit { new }; self.canon_notif_tx.lock().as_mut().unwrap().retain(|tx| tx.send(event.clone()).is_ok()) } /// Adds reorg to the queue that can be consumed with /// [`TestCanonStateSubscriptions::subscribe_to_canonical_state`] - pub fn add_next_reorg(&mut self, old: Arc, new: Arc) { + pub fn add_next_reorg(&self, old: Arc, new: Arc) { let event = CanonStateNotification::Reorg { old, new }; self.canon_notif_tx.lock().as_mut().unwrap().retain(|tx| tx.send(event.clone()).is_ok()) } diff --git a/crates/transaction-pool/src/pool/txpool.rs b/crates/transaction-pool/src/pool/txpool.rs index 44a90f1cf..bcad71edb 100644 --- a/crates/transaction-pool/src/pool/txpool.rs +++ b/crates/transaction-pool/src/pool/txpool.rs @@ -428,7 +428,7 @@ impl TxPool { } /// Update sub-pools size metrics. - pub(crate) fn update_size_metrics(&mut self) { + pub(crate) fn update_size_metrics(&self) { let stats = self.size(); self.metrics.pending_pool_transactions.set(stats.pending as f64); self.metrics.pending_pool_size_bytes.set(stats.pending_size as f64); @@ -990,7 +990,7 @@ impl AllTransactions { } /// Updates the size metrics - pub(crate) fn update_size_metrics(&mut self) { + pub(crate) fn update_size_metrics(&self) { self.metrics.all_transactions_by_hash.set(self.by_hash.len() as f64); self.metrics.all_transactions_by_id.set(self.txs.len() as f64); }