mito2/worker/
handle_catchup.rs1use std::sync::Arc;
18
19use common_telemetry::tracing::warn;
20use common_telemetry::{debug, info};
21use snafu::ensure;
22use store_api::logstore::LogStore;
23use store_api::region_engine::RegionRole;
24use store_api::region_request::{AffectedRows, RegionCatchupRequest};
25use store_api::storage::RegionId;
26use tokio::time::Instant;
27
28use crate::error::{self, Result};
29use crate::region::opener::{replay_memtable, RegionOpener};
30use crate::region::MitoRegion;
31use crate::worker::RegionWorkerLoop;
32
33impl<S: LogStore> RegionWorkerLoop<S> {
34 pub(crate) async fn handle_catchup_request(
35 &mut self,
36 region_id: RegionId,
37 request: RegionCatchupRequest,
38 ) -> Result<AffectedRows> {
39 let Some(region) = self.regions.get_region(region_id) else {
40 return error::RegionNotFoundSnafu { region_id }.fail();
41 };
42
43 if region.is_writable() {
44 debug!("Region {region_id} is writable, skip catchup");
45 return Ok(0);
46 }
47 let version = region.version();
50 let is_empty_memtable = version.memtables.is_empty();
51
52 let region = if !is_empty_memtable || region.manifest_ctx.has_update().await? {
54 if !is_empty_memtable {
55 warn!("Region {} memtables is not empty, which should not happen, manifest version: {}, last entry id: {}",
56 region.region_id,
57 region.manifest_ctx.manifest_version().await,
58 region.version_control.current().last_entry_id
59 );
60 }
61 self.reopen_region(®ion).await?
62 } else {
63 region
64 };
65
66 if region.provider.is_remote_wal() {
67 let flushed_entry_id = region.version_control.current().last_entry_id;
68 let replay_from_entry_id = request
69 .checkpoint
70 .map(|c| c.entry_id)
71 .unwrap_or_default()
72 .max(flushed_entry_id);
73 info!("Trying to replay memtable for region: {region_id}, provider: {:?}, replay from entry id: {replay_from_entry_id}, flushed entry id: {flushed_entry_id}", region.provider);
74 let timer = Instant::now();
75 let wal_entry_reader =
76 self.wal
77 .wal_entry_reader(®ion.provider, region_id, request.location_id);
78 let on_region_opened = self.wal.on_region_opened();
79 let last_entry_id = replay_memtable(
80 ®ion.provider,
81 wal_entry_reader,
82 region_id,
83 replay_from_entry_id,
84 ®ion.version_control,
85 self.config.allow_stale_entries,
86 on_region_opened,
87 )
88 .await?;
89 info!(
90 "Elapsed: {:?}, region: {region_id}, provider: {:?} catchup finished. replay from entry id: {replay_from_entry_id}, flushed entry id: {flushed_entry_id}, last entry id: {last_entry_id}, expected: {:?}.",
91 timer.elapsed(),
92 region.provider,
93 request.entry_id
94 );
95 if let Some(expected_last_entry_id) = request.entry_id {
96 ensure!(
97 last_entry_id >= expected_last_entry_id,
99 error::UnexpectedSnafu {
100 reason: format!(
101 "failed to set region {} to writable, it was expected to replayed to {}, but actually replayed to {}",
102 region_id, expected_last_entry_id, last_entry_id,
103 ),
104 }
105 )
106 }
107 } else {
108 let version = region.version_control.current();
109 let mut flushed_entry_id = version.last_entry_id;
110
111 let latest_entry_id = self
112 .wal
113 .store()
114 .latest_entry_id(®ion.provider)
115 .unwrap_or_default();
116 warn!(
117 "Skips to replay memtable for region: {}, flushed entry id: {}, latest entry id: {}",
118 region.region_id, flushed_entry_id, latest_entry_id
119 );
120
121 if latest_entry_id > flushed_entry_id {
122 warn!(
123 "Found latest entry id is greater than flushed entry id, using latest entry id as flushed entry id, region: {}, latest entry id: {}, flushed entry id: {}",
124 region_id, latest_entry_id, flushed_entry_id
125 );
126 flushed_entry_id = latest_entry_id;
127 region.version_control.set_entry_id(flushed_entry_id);
128 }
129 let on_region_opened = self.wal.on_region_opened();
130 on_region_opened(region_id, flushed_entry_id, ®ion.provider).await?;
131 }
132
133 if request.set_writable {
134 region.set_role(RegionRole::Leader);
135 }
136
137 Ok(0)
138 }
139
140 pub(crate) async fn reopen_region(
142 &mut self,
143 region: &Arc<MitoRegion>,
144 ) -> Result<Arc<MitoRegion>> {
145 let region_id = region.region_id;
146 let manifest_version = region.manifest_ctx.manifest_version().await;
147 let flushed_entry_id = region.version_control.current().last_entry_id;
148 info!("Reopening the region: {region_id}, manifest version: {manifest_version}, flushed entry id: {flushed_entry_id}");
149 let reopened_region = Arc::new(
150 RegionOpener::new(
151 region_id,
152 region.table_dir(),
153 region.access_layer.path_type(),
154 self.memtable_builder_provider.clone(),
155 self.object_store_manager.clone(),
156 self.purge_scheduler.clone(),
157 self.puffin_manager_factory.clone(),
158 self.intermediate_manager.clone(),
159 self.time_provider.clone(),
160 self.file_ref_manager.clone(),
161 )
162 .cache(Some(self.cache_manager.clone()))
163 .options(region.version().options.clone())?
164 .skip_wal_replay(true)
165 .open(&self.config, &self.wal)
166 .await?,
167 );
168 debug_assert!(!reopened_region.is_writable());
169 self.regions.insert_region(reopened_region.clone());
170
171 Ok(reopened_region)
172 }
173}