1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209
//!
use std::io::{self, Write};
use crossbeam_channel::{Sender, Receiver};
use crate::networking::MessageHeader;
use super::bytes_slab::BytesSlab;
use super::bytes_exchange::MergeQueue;
use super::stream::Stream;
use timely_logging::Logger;
use crate::logging::{CommunicationEvent, CommunicationSetup, MessageEvent, StateEvent};
fn tcp_panic(context: &'static str, cause: io::Error) -> ! {
// NOTE: some downstream crates sniff out "timely communication error:" from
// the panic message. Avoid removing or rewording this message if possible.
// It'd be nice to instead use `panic_any` here with a structured error
// type, but the panic message for `panic_any` is no good (Box<dyn Any>).
panic!("timely communication error: {}: {}", context, cause)
}
/// Repeatedly reads from a TcpStream and carves out messages.
///
/// The intended communication pattern is a sequence of (header, message)^* for valid
/// messages, followed by a header for a zero length message indicating the end of stream.
///
/// If the stream ends without being shut down, or if reading from the stream fails, the
/// receive thread panics with a message that starts with "timely communication error:"
/// in an attempt to take down the computation and cause the failures to cascade.
pub fn recv_loop<S>(
mut reader: S,
targets: Vec<Receiver<MergeQueue>>,
worker_offset: usize,
process: usize,
remote: usize,
mut logger: Option<Logger<CommunicationEvent, CommunicationSetup>>)
where
S: Stream,
{
// Log the receive thread's start.
logger.as_mut().map(|l| l.log(StateEvent { send: false, process, remote, start: true }));
let mut targets: Vec<MergeQueue> = targets.into_iter().map(|x| x.recv().expect("Failed to receive MergeQueue")).collect();
let mut buffer = BytesSlab::new(20);
// Where we stash Bytes before handing them off.
let mut stageds = Vec::with_capacity(targets.len());
for _ in 0 .. targets.len() {
stageds.push(Vec::new());
}
// Each loop iteration adds to `self.Bytes` and consumes all complete messages.
// At the start of each iteration, `self.buffer[..self.length]` represents valid
// data, and the remaining capacity is available for reading from the reader.
//
// Once the buffer fills, we need to copy incomplete messages to a new shared
// allocation and place the existing Bytes into `self.in_progress`, so that it
// can be recovered once all readers have read what they need to.
let mut active = true;
while active {
buffer.ensure_capacity(1);
assert!(!buffer.empty().is_empty());
// Attempt to read some more bytes into self.buffer.
let read = match reader.read(&mut buffer.empty()) {
Err(x) => tcp_panic("reading data", x),
Ok(n) if n == 0 => {
tcp_panic(
"reading data",
std::io::Error::new(std::io::ErrorKind::UnexpectedEof, "socket closed"),
);
}
Ok(n) => n,
};
buffer.make_valid(read);
// Consume complete messages from the front of self.buffer.
while let Some(header) = MessageHeader::try_read(buffer.valid()) {
// TODO: Consolidate message sequences sent to the same worker?
let peeled_bytes = header.required_bytes();
let bytes = buffer.extract(peeled_bytes);
// Record message receipt.
logger.as_mut().map(|logger| {
logger.log(MessageEvent { is_send: false, header, });
});
if header.length > 0 {
stageds[header.target - worker_offset].push(bytes);
}
else {
// Shutting down; confirm absence of subsequent data.
active = false;
if !buffer.valid().is_empty() {
panic!("Clean shutdown followed by data.");
}
buffer.ensure_capacity(1);
if reader.read(&mut buffer.empty()).unwrap_or_else(|e| tcp_panic("reading EOF", e)) > 0 {
panic!("Clean shutdown followed by data.");
}
}
}
// Pass bytes along to targets.
for (index, staged) in stageds.iter_mut().enumerate() {
// FIXME: try to merge `staged` before handing it to BytesPush::extend
use crate::allocator::zero_copy::bytes_exchange::BytesPush;
targets[index].extend(staged.drain(..));
}
}
// Log the receive thread's end.
logger.as_mut().map(|l| l.log(StateEvent { send: false, process, remote, start: false, }));
}
/// Repeatedly sends messages into a TcpStream.
///
/// The intended communication pattern is a sequence of (header, message)^* for valid
/// messages, followed by a header for a zero length message indicating the end of stream.
///
/// If writing to the stream fails, the send thread panics with a message that starts with
/// "timely communication error:" in an attempt to take down the computation and cause the
/// failures to cascade.
pub fn send_loop<S: Stream>(
// TODO: Maybe we don't need BufWriter with consolidation in writes.
writer: S,
sources: Vec<Sender<MergeQueue>>,
process: usize,
remote: usize,
mut logger: Option<Logger<CommunicationEvent, CommunicationSetup>>)
{
// Log the send thread's start.
logger.as_mut().map(|l| l.log(StateEvent { send: true, process, remote, start: true, }));
let mut sources: Vec<MergeQueue> = sources.into_iter().map(|x| {
let buzzer = crate::buzzer::Buzzer::new();
let queue = MergeQueue::new(buzzer);
x.send(queue.clone()).expect("failed to send MergeQueue");
queue
}).collect();
let mut writer = ::std::io::BufWriter::with_capacity(1 << 16, writer);
let mut stash = Vec::new();
while !sources.is_empty() {
// TODO: Round-robin better, to release resources fairly when overloaded.
for source in sources.iter_mut() {
use crate::allocator::zero_copy::bytes_exchange::BytesPull;
source.drain_into(&mut stash);
}
if stash.is_empty() {
// No evidence of records to read, but sources not yet empty (at start of loop).
// We are going to flush our writer (to move buffered data), double check on the
// sources for emptiness and wait on a signal only if we are sure that there will
// still be a signal incoming.
//
// We could get awoken by more data, a channel closing, or spuriously perhaps.
writer.flush().unwrap_or_else(|e| tcp_panic("flushing writer", e));
sources.retain(|source| !source.is_complete());
if !sources.is_empty() {
std::thread::park();
}
}
else {
// TODO: Could do scatter/gather write here.
for mut bytes in stash.drain(..) {
// Record message sends.
logger.as_mut().map(|logger| {
let mut offset = 0;
while let Some(header) = MessageHeader::try_read(&mut bytes[offset..]) {
logger.log(MessageEvent { is_send: true, header, });
offset += header.required_bytes();
}
});
writer.write_all(&bytes[..]).unwrap_or_else(|e| tcp_panic("writing data", e));
}
}
}
// Write final zero-length header.
// Would be better with meaningful metadata, but as this stream merges many
// workers it isn't clear that there is anything specific to write here.
let header = MessageHeader {
channel: 0,
source: 0,
target: 0,
length: 0,
seqno: 0,
};
header.write_to(&mut writer).unwrap_or_else(|e| tcp_panic("writing data", e));
writer.flush().unwrap_or_else(|e| tcp_panic("flushing writer", e));
writer.get_mut().shutdown(::std::net::Shutdown::Write).unwrap_or_else(|e| tcp_panic("shutting down writer", e));
logger.as_mut().map(|logger| logger.log(MessageEvent { is_send: true, header }));
// Log the send thread's end.
logger.as_mut().map(|l| l.log(StateEvent { send: true, process, remote, start: false, }));
}