#![allow(clippy::too_many_arguments)] use std::cmp::Reverse; use std::collections::{HashMap, HashSet}; use std::net::SocketAddr; use std::sync::Arc; use std::sync::atomic::Ordering; use std::time::{Duration, Instant}; use bytes::Bytes; use tokio::sync::mpsc::error::TrySendError; use tracing::{debug, warn}; use crate::config::{MeRouteNoWriterMode, MeWriterPickMode}; use crate::error::{ProxyError, Result}; use crate::network::IpFamily; use crate::protocol::constants::{RPC_CLOSE_CONN_U32, RPC_CLOSE_EXT_U32}; use super::MePool; use super::codec::WriterCommand; use super::pool::WriterContour; use super::registry::ConnMeta; use super::wire::build_proxy_req_payload; use rand::seq::SliceRandom; const IDLE_WRITER_PENALTY_MID_SECS: u64 = 45; const IDLE_WRITER_PENALTY_HIGH_SECS: u64 = 55; const HYBRID_GLOBAL_BURST_PERIOD_ROUNDS: u32 = 4; const HYBRID_RECENT_SUCCESS_WINDOW_MS: u64 = 120_000; const HYBRID_TIMEOUT_WARN_RATE_LIMIT_MS: u64 = 5_000; const HYBRID_RECOVERY_TRIGGER_MIN_INTERVAL_MS: u64 = 5_000; const PICK_PENALTY_WARM: u64 = 200; const PICK_PENALTY_DRAINING: u64 = 600; const PICK_PENALTY_STALE: u64 = 300; const PICK_PENALTY_DEGRADED: u64 = 250; impl MePool { /// Send RPC_PROXY_REQ. `tag_override`: per-user ad_tag (from access.user_ad_tags); if None, uses pool default. pub async fn send_proxy_req( self: &Arc, conn_id: u64, target_dc: i16, client_addr: SocketAddr, our_addr: SocketAddr, data: &[u8], proto_flags: u32, tag_override: Option<&[u8]>, ) -> Result<()> { let tag = tag_override.or(self.proxy_tag.as_deref()); let fallback_meta = ConnMeta { target_dc, client_addr, our_addr, proto_flags, }; let build_routed_payload = |effective_our_addr: SocketAddr| { ( build_proxy_req_payload( conn_id, client_addr, effective_our_addr, data, tag, proto_flags, ), ConnMeta { target_dc, client_addr, our_addr: effective_our_addr, proto_flags, }, ) }; let no_writer_mode = MeRouteNoWriterMode::from_u8( self.route_runtime .me_route_no_writer_mode .load(Ordering::Relaxed), ); let (routed_dc, unknown_target_dc) = self.resolve_target_dc_for_routing(target_dc as i32).await; let mut no_writer_deadline: Option = None; let mut emergency_attempts = 0u32; let mut async_recovery_triggered = false; let mut hybrid_recovery_round = 0u32; let mut hybrid_last_recovery_at: Option = None; let mut hybrid_total_deadline: Option = None; let hybrid_wait_step = self .route_runtime .me_route_no_writer_wait .max(Duration::from_millis(50)); let mut hybrid_wait_current = hybrid_wait_step; loop { let current_meta = self .registry .get_meta(conn_id) .await .unwrap_or_else(|| fallback_meta.clone()); let (current_payload, _) = build_routed_payload(current_meta.our_addr); if let Some(current) = self.registry.get_writer(conn_id).await { match current .tx .try_send(WriterCommand::Data(current_payload.clone())) { Ok(()) => { self.note_hybrid_route_success(); return Ok(()); } Err(TrySendError::Full(cmd)) => { if current.tx.send(cmd).await.is_ok() { self.note_hybrid_route_success(); return Ok(()); } warn!(writer_id = current.writer_id, "ME writer channel closed"); self.remove_writer_and_close_clients(current.writer_id) .await; continue; } Err(TrySendError::Closed(_)) => { warn!(writer_id = current.writer_id, "ME writer channel closed"); self.remove_writer_and_close_clients(current.writer_id) .await; continue; } } } let mut writers_snapshot = { let ws = self.writers.read().await; if ws.is_empty() { drop(ws); match no_writer_mode { MeRouteNoWriterMode::AsyncRecoveryFailfast => { let deadline = *no_writer_deadline.get_or_insert_with(|| { Instant::now() + self.route_runtime.me_route_no_writer_wait }); if !async_recovery_triggered && !unknown_target_dc { let triggered = self.trigger_async_recovery_for_target_dc(routed_dc).await; if !triggered { self.trigger_async_recovery_global().await; } async_recovery_triggered = true; } if self.wait_for_writer_until(deadline).await { continue; } self.stats.increment_me_no_writer_failfast_total(); return Err(ProxyError::Proxy( "No ME writer available in failfast window".into(), )); } MeRouteNoWriterMode::InlineRecoveryLegacy => { self.stats.increment_me_inline_recovery_total(); if !unknown_target_dc { for _ in 0..self.route_runtime.me_route_inline_recovery_attempts.max(1) { for family in self.family_order() { let map = match family { IpFamily::V4 => self.proxy_map_v4.read().await.clone(), IpFamily::V6 => self.proxy_map_v6.read().await.clone(), }; for (dc, addrs) in &map { for (ip, port) in addrs { let addr = SocketAddr::new(*ip, *port); let _ = self .connect_one_for_dc( addr, *dc, self.rng.as_ref(), ) .await; } } } if !self.writers.read().await.is_empty() { break; } } } if !self.writers.read().await.is_empty() { continue; } let deadline = *no_writer_deadline.get_or_insert_with(|| { Instant::now() + self.route_runtime.me_route_inline_recovery_wait }); if !self.wait_for_writer_until(deadline).await { if !self.writers.read().await.is_empty() { continue; } self.stats.increment_me_no_writer_failfast_total(); return Err(ProxyError::Proxy( "All ME connections dead (legacy wait timeout)".into(), )); } continue; } MeRouteNoWriterMode::HybridAsyncPersistent => { let total_deadline = *hybrid_total_deadline.get_or_insert_with(|| { Instant::now() + self.hybrid_total_wait_budget() }); if Instant::now() >= total_deadline { self.on_hybrid_timeout(total_deadline, routed_dc); return Err(ProxyError::Proxy( "ME writer not available within hybrid timeout".into(), )); } if !unknown_target_dc { self.maybe_trigger_hybrid_recovery( routed_dc, &mut hybrid_recovery_round, &mut hybrid_last_recovery_at, hybrid_wait_current, ) .await; } let deadline = Instant::now() + hybrid_wait_current; let _ = self.wait_for_writer_until(deadline).await; hybrid_wait_current = (hybrid_wait_current.saturating_mul(2)) .min(Duration::from_millis(400)); continue; } } } ws.clone() }; let mut candidate_indices = self .candidate_indices_for_dc(&writers_snapshot, routed_dc, false) .await; if candidate_indices.is_empty() { candidate_indices = self .candidate_indices_for_dc(&writers_snapshot, routed_dc, true) .await; } if candidate_indices.is_empty() { let pick_mode = self.writer_pick_mode(); match no_writer_mode { MeRouteNoWriterMode::AsyncRecoveryFailfast => { let deadline = *no_writer_deadline.get_or_insert_with(|| { Instant::now() + self.route_runtime.me_route_no_writer_wait }); if !async_recovery_triggered && !unknown_target_dc { let triggered = self.trigger_async_recovery_for_target_dc(routed_dc).await; if !triggered { self.trigger_async_recovery_global().await; } async_recovery_triggered = true; } if self.wait_for_candidate_until(routed_dc, deadline).await { continue; } self.stats .increment_me_writer_pick_no_candidate_total(pick_mode); self.stats.increment_me_no_writer_failfast_total(); return Err(ProxyError::Proxy( "No ME writers available for target DC in failfast window".into(), )); } MeRouteNoWriterMode::InlineRecoveryLegacy => { self.stats.increment_me_inline_recovery_total(); if unknown_target_dc { let deadline = *no_writer_deadline.get_or_insert_with(|| { Instant::now() + self.route_runtime.me_route_inline_recovery_wait }); if self.wait_for_candidate_until(routed_dc, deadline).await { continue; } self.stats .increment_me_writer_pick_no_candidate_total(pick_mode); self.stats.increment_me_no_writer_failfast_total(); return Err(ProxyError::Proxy( "No ME writers available for target DC".into(), )); } if emergency_attempts >= self.route_runtime.me_route_inline_recovery_attempts.max(1) { self.stats .increment_me_writer_pick_no_candidate_total(pick_mode); self.stats.increment_me_no_writer_failfast_total(); return Err(ProxyError::Proxy( "No ME writers available for target DC".into(), )); } emergency_attempts += 1; let mut endpoints = self.endpoint_candidates_for_target_dc(routed_dc).await; endpoints.shuffle(&mut rand::rng()); for addr in endpoints { if self .connect_one_for_dc(addr, routed_dc, self.rng.as_ref()) .await .is_ok() { break; } } tokio::time::sleep(Duration::from_millis(100 * emergency_attempts as u64)) .await; let ws2 = self.writers.read().await; writers_snapshot = ws2.clone(); drop(ws2); candidate_indices = self .candidate_indices_for_dc(&writers_snapshot, routed_dc, false) .await; if candidate_indices.is_empty() { candidate_indices = self .candidate_indices_for_dc(&writers_snapshot, routed_dc, true) .await; } if candidate_indices.is_empty() { self.stats .increment_me_writer_pick_no_candidate_total(pick_mode); return Err(ProxyError::Proxy( "No ME writers available for target DC".into(), )); } } MeRouteNoWriterMode::HybridAsyncPersistent => { let total_deadline = *hybrid_total_deadline.get_or_insert_with(|| { Instant::now() + self.hybrid_total_wait_budget() }); if Instant::now() >= total_deadline { self.on_hybrid_timeout(total_deadline, routed_dc); return Err(ProxyError::Proxy( "No ME writers available for target DC within hybrid timeout" .into(), )); } if !unknown_target_dc { self.maybe_trigger_hybrid_recovery( routed_dc, &mut hybrid_recovery_round, &mut hybrid_last_recovery_at, hybrid_wait_current, ) .await; } let deadline = Instant::now() + hybrid_wait_current; let _ = self.wait_for_candidate_until(routed_dc, deadline).await; hybrid_wait_current = (hybrid_wait_current.saturating_mul(2)).min(Duration::from_millis(400)); continue; } } } hybrid_wait_current = hybrid_wait_step; let pick_mode = self.writer_pick_mode(); let pick_sample_size = self.writer_pick_sample_size(); let writer_ids: Vec = candidate_indices .iter() .map(|idx| writers_snapshot[*idx].id) .collect(); let writer_idle_since = self .registry .writer_idle_since_for_writer_ids(&writer_ids) .await; let now_epoch_secs = Self::now_epoch_secs(); let start = self.rr.fetch_add(1, Ordering::Relaxed) as usize % candidate_indices.len(); let ordered_candidate_indices = if pick_mode == MeWriterPickMode::P2c { self.p2c_ordered_candidate_indices( &candidate_indices, &writers_snapshot, &writer_idle_since, now_epoch_secs, start, pick_sample_size, ) } else { if self .writer_selection_policy .me_deterministic_writer_sort .load(Ordering::Relaxed) { candidate_indices.sort_by(|lhs, rhs| { let left = &writers_snapshot[*lhs]; let right = &writers_snapshot[*rhs]; let left_key = ( self.writer_contour_rank_for_selection(left), (left.generation < self.current_generation()) as usize, left.degraded.load(Ordering::Relaxed) as usize, self.writer_idle_rank_for_selection( left, &writer_idle_since, now_epoch_secs, ), Reverse(left.tx.capacity()), left.addr, left.id, ); let right_key = ( self.writer_contour_rank_for_selection(right), (right.generation < self.current_generation()) as usize, right.degraded.load(Ordering::Relaxed) as usize, self.writer_idle_rank_for_selection( right, &writer_idle_since, now_epoch_secs, ), Reverse(right.tx.capacity()), right.addr, right.id, ); left_key.cmp(&right_key) }); } else { candidate_indices.sort_by_key(|idx| { let w = &writers_snapshot[*idx]; let degraded = w.degraded.load(Ordering::Relaxed); let stale = (w.generation < self.current_generation()) as usize; ( self.writer_contour_rank_for_selection(w), stale, degraded as usize, self.writer_idle_rank_for_selection( w, &writer_idle_since, now_epoch_secs, ), Reverse(w.tx.capacity()), ) }); } let mut ordered = Vec::::with_capacity(candidate_indices.len()); for offset in 0..candidate_indices.len() { ordered.push(candidate_indices[(start + offset) % candidate_indices.len()]); } ordered }; let mut fallback_blocking_idx: Option = None; for idx in ordered_candidate_indices { let w = &writers_snapshot[idx]; if !self.writer_accepts_new_binding(w) { continue; } let effective_our_addr = SocketAddr::new(w.source_ip, our_addr.port()); let (payload, meta) = build_routed_payload(effective_our_addr); match w.tx.clone().try_reserve_owned() { Ok(permit) => { if !self.registry.bind_writer(conn_id, w.id, meta).await { debug!( conn_id, writer_id = w.id, "ME writer disappeared before bind commit, pruning stale writer" ); drop(permit); self.remove_writer_and_close_clients(w.id).await; continue; } permit.send(WriterCommand::Data(payload.clone())); self.stats .increment_me_writer_pick_success_try_total(pick_mode); if w.generation < self.current_generation() { self.stats.increment_pool_stale_pick_total(); debug!( conn_id, writer_id = w.id, writer_generation = w.generation, current_generation = self.current_generation(), "Selected stale ME writer for fallback bind" ); } self.note_hybrid_route_success(); return Ok(()); } Err(TrySendError::Full(_)) => { if fallback_blocking_idx.is_none() { fallback_blocking_idx = Some(idx); } } Err(TrySendError::Closed(_)) => { self.stats.increment_me_writer_pick_closed_total(pick_mode); warn!(writer_id = w.id, "ME writer channel closed"); self.remove_writer_and_close_clients(w.id).await; continue; } } } let Some(blocking_idx) = fallback_blocking_idx else { self.stats.increment_me_writer_pick_full_total(pick_mode); continue; }; let w = writers_snapshot[blocking_idx].clone(); if !self.writer_accepts_new_binding(&w) { self.stats.increment_me_writer_pick_full_total(pick_mode); continue; } self.stats .increment_me_writer_pick_blocking_fallback_total(); let effective_our_addr = SocketAddr::new(w.source_ip, our_addr.port()); let (payload, meta) = build_routed_payload(effective_our_addr); let reserve_result = if let Some(timeout) = self.route_runtime.me_route_blocking_send_timeout { match tokio::time::timeout(timeout, w.tx.clone().reserve_owned()).await { Ok(result) => result, Err(_) => { self.stats.increment_me_writer_pick_full_total(pick_mode); continue; } } } else { w.tx.clone().reserve_owned().await }; match reserve_result { Ok(permit) => { if !self.registry.bind_writer(conn_id, w.id, meta).await { debug!( conn_id, writer_id = w.id, "ME writer disappeared before fallback bind commit, pruning stale writer" ); drop(permit); self.remove_writer_and_close_clients(w.id).await; continue; } permit.send(WriterCommand::Data(payload.clone())); self.stats .increment_me_writer_pick_success_fallback_total(pick_mode); if w.generation < self.current_generation() { self.stats.increment_pool_stale_pick_total(); } self.note_hybrid_route_success(); return Ok(()); } Err(_) => { self.stats.increment_me_writer_pick_closed_total(pick_mode); warn!(writer_id = w.id, "ME writer channel closed (blocking)"); self.remove_writer_and_close_clients(w.id).await; } } } } async fn wait_for_writer_until(&self, deadline: Instant) -> bool { let mut rx = self.writer_epoch.subscribe(); if !self.writers.read().await.is_empty() { return true; } let now = Instant::now(); if now >= deadline { return !self.writers.read().await.is_empty(); } let timeout = deadline.saturating_duration_since(now); if tokio::time::timeout(timeout, rx.changed()).await.is_ok() { return !self.writers.read().await.is_empty(); } !self.writers.read().await.is_empty() } async fn wait_for_candidate_until(&self, routed_dc: i32, deadline: Instant) -> bool { let mut rx = self.writer_epoch.subscribe(); loop { if self.has_candidate_for_target_dc(routed_dc).await { return true; } let now = Instant::now(); if now >= deadline { return self.has_candidate_for_target_dc(routed_dc).await; } if self.has_candidate_for_target_dc(routed_dc).await { return true; } let remaining = deadline.saturating_duration_since(Instant::now()); if remaining.is_zero() { return self.has_candidate_for_target_dc(routed_dc).await; } if tokio::time::timeout(remaining, rx.changed()).await.is_err() { return self.has_candidate_for_target_dc(routed_dc).await; } } } async fn has_candidate_for_target_dc(&self, routed_dc: i32) -> bool { let writers_snapshot = { let ws = self.writers.read().await; if ws.is_empty() { return false; } ws.clone() }; let mut candidate_indices = self .candidate_indices_for_dc(&writers_snapshot, routed_dc, false) .await; if candidate_indices.is_empty() { candidate_indices = self .candidate_indices_for_dc(&writers_snapshot, routed_dc, true) .await; } !candidate_indices.is_empty() } async fn trigger_async_recovery_for_target_dc(self: &Arc, routed_dc: i32) -> bool { let endpoints = self.endpoint_candidates_for_target_dc(routed_dc).await; if endpoints.is_empty() { return false; } self.stats.increment_me_async_recovery_trigger_total(); for addr in endpoints.into_iter().take(8) { self.trigger_immediate_refill_for_dc(addr, routed_dc); } true } async fn trigger_async_recovery_global(self: &Arc) { self.stats.increment_me_async_recovery_trigger_total(); let mut seen = HashSet::<(i32, SocketAddr)>::new(); for family in self.family_order() { let map_guard = match family { IpFamily::V4 => self.proxy_map_v4.read().await, IpFamily::V6 => self.proxy_map_v6.read().await, }; for (dc, addrs) in map_guard.iter() { for (ip, port) in addrs { let addr = SocketAddr::new(*ip, *port); if seen.insert((*dc, addr)) { self.trigger_immediate_refill_for_dc(addr, *dc); } if seen.len() >= 8 { return; } } } } } async fn endpoint_candidates_for_target_dc(&self, routed_dc: i32) -> Vec { self.preferred_endpoints_for_dc(routed_dc).await } async fn maybe_trigger_hybrid_recovery( self: &Arc, routed_dc: i32, hybrid_recovery_round: &mut u32, hybrid_last_recovery_at: &mut Option, hybrid_wait_step: Duration, ) { if !self.try_consume_hybrid_recovery_trigger_slot(HYBRID_RECOVERY_TRIGGER_MIN_INTERVAL_MS) { return; } if let Some(last) = *hybrid_last_recovery_at && last.elapsed() < hybrid_wait_step { return; } let round = *hybrid_recovery_round; let target_triggered = self.trigger_async_recovery_for_target_dc(routed_dc).await; if !target_triggered || round.is_multiple_of(HYBRID_GLOBAL_BURST_PERIOD_ROUNDS) { self.trigger_async_recovery_global().await; } *hybrid_recovery_round = round.saturating_add(1); *hybrid_last_recovery_at = Some(Instant::now()); } fn hybrid_total_wait_budget(&self) -> Duration { let base = self .route_runtime .me_route_hybrid_max_wait .max(Duration::from_millis(50)); let now_ms = Self::now_epoch_millis(); let last_success_ms = self .route_runtime .me_route_last_success_epoch_ms .load(Ordering::Relaxed); if last_success_ms != 0 && now_ms.saturating_sub(last_success_ms) <= HYBRID_RECENT_SUCCESS_WINDOW_MS { return base.saturating_mul(2); } base } fn note_hybrid_route_success(&self) { self.route_runtime .me_route_last_success_epoch_ms .store(Self::now_epoch_millis(), Ordering::Relaxed); } fn on_hybrid_timeout(&self, deadline: Instant, routed_dc: i32) { self.stats.increment_me_hybrid_timeout_total(); let now_ms = Self::now_epoch_millis(); let mut last_warn_ms = self .route_runtime .me_route_hybrid_timeout_warn_epoch_ms .load(Ordering::Relaxed); while now_ms.saturating_sub(last_warn_ms) >= HYBRID_TIMEOUT_WARN_RATE_LIMIT_MS { match self .route_runtime .me_route_hybrid_timeout_warn_epoch_ms .compare_exchange_weak(last_warn_ms, now_ms, Ordering::AcqRel, Ordering::Relaxed) { Ok(_) => { warn!( routed_dc, budget_ms = self.hybrid_total_wait_budget().as_millis() as u64, elapsed_ms = deadline.elapsed().as_millis() as u64, "ME hybrid route timeout reached" ); break; } Err(actual) => last_warn_ms = actual, } } } fn try_consume_hybrid_recovery_trigger_slot(&self, min_interval_ms: u64) -> bool { let now_ms = Self::now_epoch_millis(); let mut last_trigger_ms = self .route_runtime .me_async_recovery_last_trigger_epoch_ms .load(Ordering::Relaxed); loop { if now_ms.saturating_sub(last_trigger_ms) < min_interval_ms { return false; } match self .route_runtime .me_async_recovery_last_trigger_epoch_ms .compare_exchange_weak(last_trigger_ms, now_ms, Ordering::AcqRel, Ordering::Relaxed) { Ok(_) => return true, Err(actual) => last_trigger_ms = actual, } } } pub async fn send_close(self: &Arc, conn_id: u64) -> Result<()> { if let Some(w) = self.registry.get_writer(conn_id).await { let mut p = Vec::with_capacity(12); p.extend_from_slice(&RPC_CLOSE_EXT_U32.to_le_bytes()); p.extend_from_slice(&conn_id.to_le_bytes()); if w.tx .send(WriterCommand::DataAndFlush(Bytes::from(p))) .await .is_err() { debug!("ME close write failed"); self.remove_writer_and_close_clients(w.writer_id).await; } } else { debug!(conn_id, "ME close skipped (writer missing)"); } self.registry.unregister(conn_id).await; Ok(()) } pub async fn send_close_conn(self: &Arc, conn_id: u64) -> Result<()> { if let Some(w) = self.registry.get_writer(conn_id).await { let mut p = Vec::with_capacity(12); p.extend_from_slice(&RPC_CLOSE_CONN_U32.to_le_bytes()); p.extend_from_slice(&conn_id.to_le_bytes()); match w.tx.try_send(WriterCommand::DataAndFlush(Bytes::from(p))) { Ok(()) => {} Err(TrySendError::Full(cmd)) => { let _ = tokio::time::timeout(Duration::from_millis(50), w.tx.send(cmd)).await; } Err(TrySendError::Closed(_)) => { debug!(conn_id, "ME close_conn skipped: writer channel closed"); } } } else { debug!(conn_id, "ME close_conn skipped (writer missing)"); } self.registry.unregister(conn_id).await; Ok(()) } pub async fn shutdown_send_close_conn_all(self: &Arc) -> usize { let conn_ids = self.registry.active_conn_ids().await; let total = conn_ids.len(); for conn_id in conn_ids { let _ = self.send_close_conn(conn_id).await; } total } pub fn connection_count(&self) -> usize { self.conn_count.load(Ordering::Relaxed) } pub(super) async fn candidate_indices_for_dc( &self, writers: &[super::pool::MeWriter], routed_dc: i32, include_warm: bool, ) -> Vec { let preferred = self.preferred_endpoints_for_dc(routed_dc).await; if preferred.is_empty() { return Vec::new(); } let mut out = Vec::new(); for (idx, w) in writers.iter().enumerate() { if !self.writer_eligible_for_selection(w, include_warm) { continue; } if w.writer_dc == routed_dc && preferred.contains(&w.addr) { out.push(idx); } } out } fn writer_eligible_for_selection( &self, writer: &super::pool::MeWriter, include_warm: bool, ) -> bool { if !self.writer_accepts_new_binding(writer) { return false; } match WriterContour::from_u8(writer.contour.load(Ordering::Relaxed)) { WriterContour::Active => true, WriterContour::Warm => include_warm, WriterContour::Draining => true, } } fn writer_contour_rank_for_selection(&self, writer: &super::pool::MeWriter) -> usize { match WriterContour::from_u8(writer.contour.load(Ordering::Relaxed)) { WriterContour::Active => 0, WriterContour::Warm => 1, WriterContour::Draining => 2, } } fn writer_idle_rank_for_selection( &self, writer: &super::pool::MeWriter, idle_since_by_writer: &HashMap, now_epoch_secs: u64, ) -> usize { let Some(idle_since) = idle_since_by_writer.get(&writer.id).copied() else { return 0; }; let idle_age_secs = now_epoch_secs.saturating_sub(idle_since); if idle_age_secs >= IDLE_WRITER_PENALTY_HIGH_SECS { 2 } else if idle_age_secs >= IDLE_WRITER_PENALTY_MID_SECS { 1 } else { 0 } } fn writer_pick_score( &self, writer: &super::pool::MeWriter, idle_since_by_writer: &HashMap, now_epoch_secs: u64, ) -> u64 { let contour_penalty = match WriterContour::from_u8(writer.contour.load(Ordering::Relaxed)) { WriterContour::Active => 0, WriterContour::Warm => PICK_PENALTY_WARM, WriterContour::Draining => PICK_PENALTY_DRAINING, }; let stale_penalty = if writer.generation < self.current_generation() { PICK_PENALTY_STALE } else { 0 }; let degraded_penalty = if writer.degraded.load(Ordering::Relaxed) { PICK_PENALTY_DEGRADED } else { 0 }; let idle_penalty = (self.writer_idle_rank_for_selection(writer, idle_since_by_writer, now_epoch_secs) as u64) * 100; let queue_cap = self.writer_lifecycle.writer_cmd_channel_capacity.max(1) as u64; let queue_remaining = writer.tx.capacity() as u64; let queue_used = queue_cap.saturating_sub(queue_remaining.min(queue_cap)); let queue_util_pct = queue_used.saturating_mul(100) / queue_cap; let queue_penalty = queue_util_pct.saturating_mul(4); let rtt_penalty = ((writer.rtt_ema_ms_x10.load(Ordering::Relaxed) as u64).saturating_add(5) / 10) .min(400); contour_penalty .saturating_add(stale_penalty) .saturating_add(degraded_penalty) .saturating_add(idle_penalty) .saturating_add(queue_penalty) .saturating_add(rtt_penalty) } fn p2c_ordered_candidate_indices( &self, candidate_indices: &[usize], writers_snapshot: &[super::pool::MeWriter], idle_since_by_writer: &HashMap, now_epoch_secs: u64, start: usize, sample_size: usize, ) -> Vec { let total = candidate_indices.len(); if total == 0 { return Vec::new(); } let mut sampled = Vec::::with_capacity(sample_size.min(total)); let mut seen = HashSet::::with_capacity(total); for offset in 0..sample_size.min(total) { let idx = candidate_indices[(start + offset) % total]; if seen.insert(idx) { sampled.push(idx); } } sampled.sort_by_key(|idx| { let writer = &writers_snapshot[*idx]; ( self.writer_pick_score(writer, idle_since_by_writer, now_epoch_secs), writer.addr, writer.id, ) }); let mut ordered = Vec::::with_capacity(total); ordered.extend(sampled.iter().copied()); for offset in 0..total { let idx = candidate_indices[(start + offset) % total]; if seen.insert(idx) { ordered.push(idx); } } ordered } }