#![cfg_attr(feature = "proptest-impl", allow(dead_code))]
use std::{
net::{Ipv4Addr, SocketAddrV4},
sync::Arc,
time::Duration,
};
use chrono::Utc;
use futures::{
channel::{mpsc, oneshot},
future::{self, AbortHandle, Future, FutureExt},
};
use tokio::{
sync::broadcast::{self, error::TryRecvError},
task::JoinHandle,
};
use zebra_chain::block::Height;
use crate::{
constants,
peer::{
error::SharedPeerError, CancelHeartbeatTask, Client, ClientRequest, ConnectionInfo,
ErrorSlot,
},
peer_set::InventoryChange,
protocol::{
external::{types::Version, AddrInVersion},
types::{Nonce, PeerServices},
},
BoxError, VersionMessage,
};
#[cfg(test)]
mod vectors;
const MAX_PEER_CONNECTION_TIME: Duration = Duration::from_secs(10);
pub struct ClientTestHarness {
client_request_receiver: Option<mpsc::Receiver<ClientRequest>>,
shutdown_receiver: Option<oneshot::Receiver<CancelHeartbeatTask>>,
#[allow(dead_code)]
inv_receiver: Option<broadcast::Receiver<InventoryChange>>,
error_slot: ErrorSlot,
remote_version: Version,
connection_aborter: AbortHandle,
heartbeat_aborter: AbortHandle,
}
impl ClientTestHarness {
pub fn build() -> ClientTestHarnessBuilder {
ClientTestHarnessBuilder {
version: None,
connection_task: None,
heartbeat_task: None,
}
}
pub fn remote_version(&self) -> Version {
self.remote_version
}
pub fn wants_connection_heartbeats(&mut self) -> bool {
let receive_result = self
.shutdown_receiver
.as_mut()
.expect("heartbeat shutdown receiver endpoint has been dropped")
.try_recv();
match receive_result {
Ok(None) => true,
Ok(Some(CancelHeartbeatTask)) | Err(oneshot::Canceled) => false,
}
}
pub fn drop_heartbeat_shutdown_receiver(&mut self) {
let hearbeat_future = self
.shutdown_receiver
.take()
.expect("unexpected test failure: heartbeat shutdown receiver endpoint has already been dropped");
std::mem::drop(hearbeat_future);
}
pub fn close_outbound_client_request_receiver(&mut self) {
self.client_request_receiver
.as_mut()
.expect("request receiver endpoint has been dropped")
.close();
}
pub fn drop_outbound_client_request_receiver(&mut self) {
self.client_request_receiver
.take()
.expect("request receiver endpoint has already been dropped");
}
pub(crate) fn try_to_receive_outbound_client_request(&mut self) -> ReceiveRequestAttempt {
let receive_result = self
.client_request_receiver
.as_mut()
.expect("request receiver endpoint has been dropped")
.try_next();
match receive_result {
Ok(Some(request)) => ReceiveRequestAttempt::Request(request),
Ok(None) => ReceiveRequestAttempt::Closed,
Err(_) => ReceiveRequestAttempt::Empty,
}
}
#[allow(dead_code)]
pub fn drop_inventory_change_receiver(&mut self) {
self.inv_receiver
.take()
.expect("inventory change receiver endpoint has already been dropped");
}
#[allow(dead_code)]
#[allow(clippy::unwrap_in_result)]
pub(crate) fn try_to_receive_inventory_change(&mut self) -> Option<InventoryChange> {
let receive_result = self
.inv_receiver
.as_mut()
.expect("inventory change receiver endpoint has been dropped")
.try_recv();
match receive_result {
Ok(change) => Some(change),
Err(TryRecvError::Empty) => None,
Err(TryRecvError::Closed) => None,
Err(TryRecvError::Lagged(skipped_messages)) => unreachable!(
"unexpected lagged inventory receiver in tests, skipped {} messages",
skipped_messages,
),
}
}
pub fn current_error(&self) -> Option<SharedPeerError> {
self.error_slot.try_get_error()
}
pub fn set_error(&self, error: impl Into<SharedPeerError>) {
self.error_slot
.try_update_error(error.into())
.expect("unexpected earlier error in error slot")
}
pub async fn stop_connection_task(&self) {
self.connection_aborter.abort();
tokio::task::yield_now().await;
}
pub async fn stop_heartbeat_task(&self) {
self.heartbeat_aborter.abort();
tokio::task::yield_now().await;
}
}
pub(crate) enum ReceiveRequestAttempt {
Closed,
Empty,
Request(ClientRequest),
}
impl ReceiveRequestAttempt {
pub fn is_closed(&self) -> bool {
matches!(self, ReceiveRequestAttempt::Closed)
}
pub fn is_empty(&self) -> bool {
matches!(self, ReceiveRequestAttempt::Empty)
}
#[allow(dead_code)]
pub fn request(self) -> Option<ClientRequest> {
match self {
ReceiveRequestAttempt::Request(request) => Some(request),
ReceiveRequestAttempt::Closed | ReceiveRequestAttempt::Empty => None,
}
}
}
pub struct ClientTestHarnessBuilder<C = future::Ready<()>, H = future::Ready<()>> {
connection_task: Option<C>,
heartbeat_task: Option<H>,
version: Option<Version>,
}
impl<C, H> ClientTestHarnessBuilder<C, H>
where
C: Future<Output = ()> + Send + 'static,
H: Future<Output = ()> + Send + 'static,
{
pub fn with_version(mut self, version: Version) -> Self {
self.version = Some(version);
self
}
pub fn with_connection_task<NewC>(
self,
connection_task: NewC,
) -> ClientTestHarnessBuilder<NewC, H> {
ClientTestHarnessBuilder {
connection_task: Some(connection_task),
heartbeat_task: self.heartbeat_task,
version: self.version,
}
}
pub fn with_heartbeat_task<NewH>(
self,
heartbeat_task: NewH,
) -> ClientTestHarnessBuilder<C, NewH> {
ClientTestHarnessBuilder {
connection_task: self.connection_task,
heartbeat_task: Some(heartbeat_task),
version: self.version,
}
}
pub fn finish(self) -> (Client, ClientTestHarness) {
let (shutdown_sender, shutdown_receiver) = oneshot::channel();
let (client_request_sender, client_request_receiver) = mpsc::channel(1);
let (inv_sender, inv_receiver) = broadcast::channel(5);
let error_slot = ErrorSlot::default();
let remote_version = self.version.unwrap_or(Version(0));
let (connection_task, connection_aborter) =
Self::spawn_background_task_or_fallback(self.connection_task);
let (heartbeat_task, heartbeat_aborter) =
Self::spawn_background_task_or_fallback_with_result(self.heartbeat_task);
let negotiated_version =
std::cmp::min(remote_version, constants::CURRENT_NETWORK_PROTOCOL_VERSION);
let remote = VersionMessage {
version: remote_version,
services: PeerServices::default(),
timestamp: Utc::now(),
address_recv: AddrInVersion::new(
SocketAddrV4::new(Ipv4Addr::LOCALHOST, 1),
PeerServices::default(),
),
address_from: AddrInVersion::new(
SocketAddrV4::new(Ipv4Addr::LOCALHOST, 2),
PeerServices::default(),
),
nonce: Nonce::default(),
user_agent: "client test harness".to_string(),
start_height: Height(0),
relay: true,
};
let connection_info = Arc::new(ConnectionInfo {
connected_addr: crate::peer::ConnectedAddr::Isolated,
remote,
negotiated_version,
});
let client = Client {
connection_info,
shutdown_tx: Some(shutdown_sender),
server_tx: client_request_sender,
inv_collector: inv_sender,
error_slot: error_slot.clone(),
connection_task,
heartbeat_task,
};
let harness = ClientTestHarness {
client_request_receiver: Some(client_request_receiver),
shutdown_receiver: Some(shutdown_receiver),
inv_receiver: Some(inv_receiver),
error_slot,
remote_version,
connection_aborter,
heartbeat_aborter,
};
(client, harness)
}
fn spawn_background_task_or_fallback<T>(task_future: Option<T>) -> (JoinHandle<()>, AbortHandle)
where
T: Future<Output = ()> + Send + 'static,
{
match task_future {
Some(future) => Self::spawn_background_task(future),
None => Self::spawn_background_task(tokio::time::sleep(MAX_PEER_CONNECTION_TIME)),
}
}
fn spawn_background_task<T>(task_future: T) -> (JoinHandle<()>, AbortHandle)
where
T: Future<Output = ()> + Send + 'static,
{
let (task, abort_handle) = future::abortable(task_future);
let task_handle = tokio::spawn(task.map(|_result| ()));
(task_handle, abort_handle)
}
fn spawn_background_task_or_fallback_with_result<T>(
task_future: Option<T>,
) -> (JoinHandle<Result<(), BoxError>>, AbortHandle)
where
T: Future<Output = ()> + Send + 'static,
{
match task_future {
Some(future) => Self::spawn_background_task_with_result(future),
None => Self::spawn_background_task_with_result(tokio::time::sleep(
MAX_PEER_CONNECTION_TIME,
)),
}
}
fn spawn_background_task_with_result<T>(
task_future: T,
) -> (JoinHandle<Result<(), BoxError>>, AbortHandle)
where
T: Future<Output = ()> + Send + 'static,
{
let (task, abort_handle) = future::abortable(task_future);
let task_handle = tokio::spawn(task.map(|_result| Ok(())));
(task_handle, abort_handle)
}
}