Updating docs
This commit is contained in:
parent
a54d855cbe
commit
2815d6d9e9
|
@ -811,7 +811,7 @@ dependencies = [
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "stm32h7xx-hal"
|
name = "stm32h7xx-hal"
|
||||||
version = "0.9.0"
|
version = "0.9.0"
|
||||||
source = "git+https://github.com/quartiq/stm32h7xx-hal.git?branch=feature/smoltcp-update#191b1d50a8a4d956492649630efaf563f59e35bf"
|
source = "git+https://github.com/quartiq/stm32h7xx-hal.git?rev=33aa67d#33aa67d74790cb9f680a4f281b72df0664bcf03c"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"bare-metal 1.0.0",
|
"bare-metal 1.0.0",
|
||||||
"cast",
|
"cast",
|
||||||
|
|
|
@ -0,0 +1,189 @@
|
||||||
|
#!/usr/bin/python3
|
||||||
|
"""
|
||||||
|
Author: Ryan Summers
|
||||||
|
|
||||||
|
Description: Provides a mechanism for measuring Stabilizer stream data throughput.
|
||||||
|
"""
|
||||||
|
import socket
|
||||||
|
import collections
|
||||||
|
import struct
|
||||||
|
import time
|
||||||
|
import logging
|
||||||
|
|
||||||
|
# Representation of a single UDP packet transmitted by Stabilizer.
|
||||||
|
Packet = collections.namedtuple('Packet', ['index', 'adc', 'dac'])
|
||||||
|
|
||||||
|
class Timer:
|
||||||
|
""" A basic timer for measuring elapsed time periods. """
|
||||||
|
|
||||||
|
def __init__(self, period=1.0):
|
||||||
|
""" Create the timer with the provided period. """
|
||||||
|
self.start_time = time.time()
|
||||||
|
self.trigger_time = self.start_time + period
|
||||||
|
self.period = period
|
||||||
|
self.started = False
|
||||||
|
|
||||||
|
|
||||||
|
def is_triggered(self):
|
||||||
|
""" Check if the timer period has elapsed. """
|
||||||
|
now = time.time()
|
||||||
|
return now >= self.trigger_time
|
||||||
|
|
||||||
|
|
||||||
|
def start(self):
|
||||||
|
""" Start the timer. """
|
||||||
|
self.start_time = time.time()
|
||||||
|
self.started = True
|
||||||
|
|
||||||
|
|
||||||
|
def is_started(self):
|
||||||
|
""" Check if the timer has started. """
|
||||||
|
return self.started
|
||||||
|
|
||||||
|
|
||||||
|
def arm(self):
|
||||||
|
""" Arm the timer trigger. """
|
||||||
|
self.trigger_time = time.time() + self.period
|
||||||
|
|
||||||
|
|
||||||
|
def elapsed(self):
|
||||||
|
""" Get the elapsed time since the timer was started. """
|
||||||
|
now = time.time()
|
||||||
|
return now - self.start_time
|
||||||
|
|
||||||
|
|
||||||
|
class PacketParser:
|
||||||
|
""" Utilize class used for parsing received UDP data. """
|
||||||
|
|
||||||
|
def __init__(self):
|
||||||
|
""" Initialize the parser. """
|
||||||
|
self.buf = b''
|
||||||
|
self.total_bytes = 0
|
||||||
|
|
||||||
|
|
||||||
|
def ingress(self, data):
|
||||||
|
""" Ingress received UDP data. """
|
||||||
|
self.total_bytes += len(data)
|
||||||
|
self.buf += data
|
||||||
|
|
||||||
|
|
||||||
|
def parse_all_packets(self):
|
||||||
|
""" Parse all received packets from the receive buffer.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
A list of received Packets.
|
||||||
|
"""
|
||||||
|
packets = []
|
||||||
|
while True:
|
||||||
|
new_packets = self._parse()
|
||||||
|
if new_packets:
|
||||||
|
packets += new_packets
|
||||||
|
else:
|
||||||
|
return packets
|
||||||
|
|
||||||
|
|
||||||
|
def _parse(self):
|
||||||
|
""" Attempt to parse packets from the received buffer. """
|
||||||
|
# Attempt to parse a block from the buffer.
|
||||||
|
if len(self.buf) < 4:
|
||||||
|
return None
|
||||||
|
|
||||||
|
start_id, num_blocks, data_size = struct.unpack_from('!HBB', self.buf)
|
||||||
|
|
||||||
|
packet_size = 4 + data_size * num_blocks * 8
|
||||||
|
|
||||||
|
if len(self.buf) < packet_size:
|
||||||
|
return None
|
||||||
|
|
||||||
|
self.buf = self.buf[4:]
|
||||||
|
|
||||||
|
packets = []
|
||||||
|
for offset in range(num_blocks):
|
||||||
|
adcs_dacs = struct.unpack_from(f'!{4 * data_size}H', self.buf)
|
||||||
|
adc = [
|
||||||
|
adcs_dacs[0:data_size],
|
||||||
|
adcs_dacs[data_size:2*data_size],
|
||||||
|
]
|
||||||
|
|
||||||
|
dac = [
|
||||||
|
adcs_dacs[2*data_size: 3*data_size],
|
||||||
|
adcs_dacs[3*data_size:],
|
||||||
|
]
|
||||||
|
|
||||||
|
self.buf = self.buf[8*data_size:]
|
||||||
|
packets.append(Packet(start_id + offset, adc, dac))
|
||||||
|
|
||||||
|
return packets
|
||||||
|
|
||||||
|
|
||||||
|
def check_index(previous_index, next_index):
|
||||||
|
""" Check if two indices are sequential. """
|
||||||
|
if previous_index == -1:
|
||||||
|
return True
|
||||||
|
|
||||||
|
# Handle index roll-over. Indices are only stored in 16-bit numbers.
|
||||||
|
if next_index < previous_index:
|
||||||
|
next_index += 65536
|
||||||
|
|
||||||
|
expected_index = previous_index + 1
|
||||||
|
|
||||||
|
return next_index == expected_index
|
||||||
|
|
||||||
|
|
||||||
|
def main():
|
||||||
|
""" Main program. """
|
||||||
|
connection = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
|
||||||
|
connection.bind(("", 1111))
|
||||||
|
|
||||||
|
logging.basicConfig(level=logging.INFO,
|
||||||
|
format='%(asctime)s.%(msecs)03d %(levelname)-8s %(message)s')
|
||||||
|
|
||||||
|
last_index = -1
|
||||||
|
|
||||||
|
drop_count = 0
|
||||||
|
good_blocks = 0
|
||||||
|
|
||||||
|
timer = Timer()
|
||||||
|
parser = PacketParser()
|
||||||
|
|
||||||
|
while True:
|
||||||
|
# Receive any data over UDP and parse it.
|
||||||
|
data = connection.recv(4096)
|
||||||
|
if data and not timer.is_started():
|
||||||
|
timer.start()
|
||||||
|
|
||||||
|
parser.ingress(data)
|
||||||
|
|
||||||
|
# Handle any received packets.
|
||||||
|
for packet in parser.parse_all_packets():
|
||||||
|
|
||||||
|
# Handle any dropped packets.
|
||||||
|
if not check_index(last_index, packet.index):
|
||||||
|
print(hex(last_index), hex(packet.index))
|
||||||
|
if packet.index < (last_index + 1):
|
||||||
|
dropped = packet.index + 65536 - (last_index + 1)
|
||||||
|
else:
|
||||||
|
dropped = packet.index - (last_index + 1)
|
||||||
|
|
||||||
|
drop_count += dropped
|
||||||
|
|
||||||
|
last_index = packet.index
|
||||||
|
good_blocks += 1
|
||||||
|
|
||||||
|
# Report the throughput periodically.
|
||||||
|
if timer.is_triggered():
|
||||||
|
drate = parser.total_bytes * 8 / 1e6 / timer.elapsed()
|
||||||
|
|
||||||
|
print(f'''
|
||||||
|
Data Rate: {drate:.3f} Mbps
|
||||||
|
Received Blocks: {good_blocks}
|
||||||
|
Dropped blocks: {drop_count}
|
||||||
|
|
||||||
|
Metadata: {parser.total_bytes / 1e6:.3f} MB in {timer.elapsed():.2f} s
|
||||||
|
----
|
||||||
|
''')
|
||||||
|
timer.arm()
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == '__main__':
|
||||||
|
main()
|
|
@ -95,8 +95,7 @@ const APP: () = {
|
||||||
stabilizer.net.mac_address,
|
stabilizer.net.mac_address,
|
||||||
);
|
);
|
||||||
|
|
||||||
let generator =
|
let generator = network.enable_streaming();
|
||||||
network.enable_streaming(StreamTarget::default().into());
|
|
||||||
|
|
||||||
// Spawn a settings update for default settings.
|
// Spawn a settings update for default settings.
|
||||||
c.spawn.settings_update().unwrap();
|
c.spawn.settings_update().unwrap();
|
||||||
|
|
|
@ -1,3 +1,20 @@
|
||||||
|
///! Stabilizer data stream capabilities
|
||||||
|
///!
|
||||||
|
///! # Design
|
||||||
|
///! Stabilizer data streamining utilizes UDP packets to send live data streams at high throughput.
|
||||||
|
///! Packets are always sent in a best-effort fashion, and data may be dropped. Each packet contains
|
||||||
|
///! an identifier that can be used to detect any dropped data.
|
||||||
|
///!
|
||||||
|
///! The current implementation utilizes an single-producer, single-consumer queue to send data
|
||||||
|
///! between a high priority task and the UDP transmitter.
|
||||||
|
///!
|
||||||
|
///! A "batch" of data is defined to be a single item in the SPSC queue sent to the UDP transmitter
|
||||||
|
///! thread. The transmitter thread then serializes as many sequential "batches" into a single UDP
|
||||||
|
///! packet as possible. The UDP packet is also given a header indicating the starting batch
|
||||||
|
///! sequence number and the number of batches present. If the UDP transmitter encounters a
|
||||||
|
///! non-sequential batch, it does not enqueue it into the packet and instead transmits any staged
|
||||||
|
///! data. The non-sequential batch is then transmitted in a new UDP packet. This method allows a
|
||||||
|
///! receiver to detect dropped batches (e.g. due to processing overhead).
|
||||||
use core::borrow::BorrowMut;
|
use core::borrow::BorrowMut;
|
||||||
use heapless::spsc::{Consumer, Producer, Queue};
|
use heapless::spsc::{Consumer, Producer, Queue};
|
||||||
use miniconf::MiniconfAtomic;
|
use miniconf::MiniconfAtomic;
|
||||||
|
@ -10,8 +27,10 @@ use crate::hardware::design_parameters::SAMPLE_BUFFER_SIZE;
|
||||||
// The number of data blocks that we will buffer in the queue.
|
// The number of data blocks that we will buffer in the queue.
|
||||||
const BLOCK_BUFFER_SIZE: usize = 30;
|
const BLOCK_BUFFER_SIZE: usize = 30;
|
||||||
|
|
||||||
|
// A factor that data may be subsampled at.
|
||||||
const SUBSAMPLE_RATE: usize = 1;
|
const SUBSAMPLE_RATE: usize = 1;
|
||||||
|
|
||||||
|
/// Represents the destination for the UDP stream to send data to.
|
||||||
#[derive(Copy, Clone, Debug, MiniconfAtomic, Deserialize)]
|
#[derive(Copy, Clone, Debug, MiniconfAtomic, Deserialize)]
|
||||||
pub struct StreamTarget {
|
pub struct StreamTarget {
|
||||||
pub ip: [u8; 4],
|
pub ip: [u8; 4],
|
||||||
|
@ -38,6 +57,23 @@ impl Into<SocketAddr> for StreamTarget {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// A basic "batch" of data.
|
||||||
|
// Note: In the future, the stream may be generic over this type.
|
||||||
|
#[derive(Debug, Copy, Clone)]
|
||||||
|
pub struct AdcDacData {
|
||||||
|
block_id: u16,
|
||||||
|
adcs: [[u16; SAMPLE_BUFFER_SIZE]; 2],
|
||||||
|
dacs: [[u16; SAMPLE_BUFFER_SIZE]; 2],
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Configure streaming on a device.
|
||||||
|
///
|
||||||
|
/// # Args
|
||||||
|
/// * `stack` - A reference to the shared network stack.
|
||||||
|
///
|
||||||
|
/// # Returns
|
||||||
|
/// (generator, stream) where `generator` can be used to enqueue "batches" for transmission. The
|
||||||
|
/// `stream` is the logically consumer (UDP transmitter) of the enqueued data.
|
||||||
pub fn setup_streaming(
|
pub fn setup_streaming(
|
||||||
stack: NetworkReference,
|
stack: NetworkReference,
|
||||||
) -> (BlockGenerator, DataStream) {
|
) -> (BlockGenerator, DataStream) {
|
||||||
|
@ -52,28 +88,34 @@ pub fn setup_streaming(
|
||||||
(generator, stream)
|
(generator, stream)
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Copy, Clone)]
|
/// The data generator for a stream.
|
||||||
pub struct AdcDacData {
|
|
||||||
block_id: u16,
|
|
||||||
adcs: [[u16; SAMPLE_BUFFER_SIZE]; 2],
|
|
||||||
dacs: [[u16; SAMPLE_BUFFER_SIZE]; 2],
|
|
||||||
}
|
|
||||||
|
|
||||||
pub struct BlockGenerator {
|
pub struct BlockGenerator {
|
||||||
queue: Producer<'static, AdcDacData, BLOCK_BUFFER_SIZE>,
|
queue: Producer<'static, AdcDacData, BLOCK_BUFFER_SIZE>,
|
||||||
current_id: u16,
|
current_id: u16,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl BlockGenerator {
|
impl BlockGenerator {
|
||||||
pub fn new(
|
/// Construct a new generator.
|
||||||
queue: Producer<'static, AdcDacData, BLOCK_BUFFER_SIZE>,
|
/// # Args
|
||||||
) -> Self {
|
/// * `queue` - The producer portion of the SPSC queue to enqueue data into.
|
||||||
|
///
|
||||||
|
/// # Returns
|
||||||
|
/// The generator to use.
|
||||||
|
fn new(queue: Producer<'static, AdcDacData, BLOCK_BUFFER_SIZE>) -> Self {
|
||||||
Self {
|
Self {
|
||||||
queue,
|
queue,
|
||||||
current_id: 0,
|
current_id: 0,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Schedule data to be sent by the generator.
|
||||||
|
///
|
||||||
|
/// # Note
|
||||||
|
/// If no space is available, the data batch may be silently dropped.
|
||||||
|
///
|
||||||
|
/// # Args
|
||||||
|
/// * `adcs` - The ADC data to transmit.
|
||||||
|
/// * `dacs` - The DAC data to transmit.
|
||||||
pub fn send(
|
pub fn send(
|
||||||
&mut self,
|
&mut self,
|
||||||
adcs: &[&mut [u16; SAMPLE_BUFFER_SIZE]; 2],
|
adcs: &[&mut [u16; SAMPLE_BUFFER_SIZE]; 2],
|
||||||
|
@ -90,26 +132,21 @@ impl BlockGenerator {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub struct DataStream {
|
/// Represents a single UDP packet sent by the stream.
|
||||||
stack: NetworkReference,
|
///
|
||||||
socket: Option<<NetworkReference as UdpClientStack>::UdpSocket>,
|
/// # Packet Format
|
||||||
queue: Consumer<'static, AdcDacData, BLOCK_BUFFER_SIZE>,
|
/// All data is sent in network-endian format. The format is as follows
|
||||||
remote: Option<SocketAddr>,
|
///
|
||||||
buffer: [u8; 1024],
|
/// Header:
|
||||||
}
|
/// [0..2]: Start block ID (u16)
|
||||||
|
/// [2..3]: Num Blocks present (u8) <N>
|
||||||
// Datapacket format:
|
/// [3..4]: Batch Size (u8) <BS>
|
||||||
//
|
///
|
||||||
// Header:
|
/// Following the header, batches are added sequentially. Each batch takes the form of:
|
||||||
// [0..2]: Start block ID (u16)
|
/// [<BS>*0..<BS>*2]: ADC0
|
||||||
// [2..3]: Num Blocks present (u8) <N>
|
/// [<BS>*2..<BS>*4]: ADC1
|
||||||
// [3..4]: Batch Size (u8) <BS>
|
/// [<BS>*4..<BS>*6]: DAC0
|
||||||
//
|
/// [<BS>*6..<BS>*8]: DAC1
|
||||||
// Following the header, batches are added sequentially. Each batch takes the form of:
|
|
||||||
// [<BS>*0..<BS>*2]: ADC0
|
|
||||||
// [<BS>*2..<BS>*4]: ADC1
|
|
||||||
// [<BS>*4..<BS>*6]: DAC0
|
|
||||||
// [<BS>*6..<BS>*8]: DAC1
|
|
||||||
struct DataPacket<'a> {
|
struct DataPacket<'a> {
|
||||||
buf: &'a mut [u8],
|
buf: &'a mut [u8],
|
||||||
subsample_rate: usize,
|
subsample_rate: usize,
|
||||||
|
@ -119,6 +156,11 @@ struct DataPacket<'a> {
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<'a> DataPacket<'a> {
|
impl<'a> DataPacket<'a> {
|
||||||
|
/// Construct a new packet.
|
||||||
|
///
|
||||||
|
/// # Args
|
||||||
|
/// * `buf` - The location to serialize the data packet into.
|
||||||
|
/// * `subsample_rate` - The factor at which to subsample data from batches.
|
||||||
pub fn new(buf: &'a mut [u8], subsample_rate: usize) -> Self {
|
pub fn new(buf: &'a mut [u8], subsample_rate: usize) -> Self {
|
||||||
Self {
|
Self {
|
||||||
buf,
|
buf,
|
||||||
|
@ -129,6 +171,13 @@ impl<'a> DataPacket<'a> {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Add a batch of data to the packet.
|
||||||
|
///
|
||||||
|
/// # Note
|
||||||
|
/// Serialization occurs as the packet is added.
|
||||||
|
///
|
||||||
|
/// # Args
|
||||||
|
/// * `batch` - The batch to add to the packet.
|
||||||
pub fn add_batch(&mut self, batch: &AdcDacData) -> Result<(), ()> {
|
pub fn add_batch(&mut self, batch: &AdcDacData) -> Result<(), ()> {
|
||||||
// Check that the block is sequential.
|
// Check that the block is sequential.
|
||||||
if let Some(id) = &self.start_id {
|
if let Some(id) = &self.start_id {
|
||||||
|
@ -170,6 +219,11 @@ impl<'a> DataPacket<'a> {
|
||||||
block_size_bytes * self.num_blocks as usize + header_length
|
block_size_bytes * self.num_blocks as usize + header_length
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Complete the packet and prepare it for transmission.
|
||||||
|
///
|
||||||
|
/// # Returns
|
||||||
|
/// The size of the packet. The user should utilize the original buffer provided for packet
|
||||||
|
/// construction to access the packet.
|
||||||
pub fn finish(self) -> usize {
|
pub fn finish(self) -> usize {
|
||||||
let block_sample_size = SAMPLE_BUFFER_SIZE / self.subsample_rate;
|
let block_sample_size = SAMPLE_BUFFER_SIZE / self.subsample_rate;
|
||||||
|
|
||||||
|
@ -183,15 +237,32 @@ impl<'a> DataPacket<'a> {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// The "consumer" portion of the data stream.
|
||||||
|
///
|
||||||
|
/// # Note
|
||||||
|
/// This is responsible for consuming data and sending it over UDP.
|
||||||
|
pub struct DataStream {
|
||||||
|
stack: NetworkReference,
|
||||||
|
socket: Option<<NetworkReference as UdpClientStack>::UdpSocket>,
|
||||||
|
queue: Consumer<'static, AdcDacData, BLOCK_BUFFER_SIZE>,
|
||||||
|
remote: SocketAddr,
|
||||||
|
buffer: [u8; 1024],
|
||||||
|
}
|
||||||
|
|
||||||
impl DataStream {
|
impl DataStream {
|
||||||
pub fn new(
|
/// Construct a new data streamer.
|
||||||
|
///
|
||||||
|
/// # Args
|
||||||
|
/// * `stack` - A reference to the shared network stack.
|
||||||
|
/// * `consumer` - The read side of the queue containing data to transmit.
|
||||||
|
fn new(
|
||||||
stack: NetworkReference,
|
stack: NetworkReference,
|
||||||
consumer: Consumer<'static, AdcDacData, BLOCK_BUFFER_SIZE>,
|
consumer: Consumer<'static, AdcDacData, BLOCK_BUFFER_SIZE>,
|
||||||
) -> Self {
|
) -> Self {
|
||||||
Self {
|
Self {
|
||||||
stack,
|
stack,
|
||||||
socket: None,
|
socket: None,
|
||||||
remote: None,
|
remote: StreamTarget::default().into(),
|
||||||
queue: consumer,
|
queue: consumer,
|
||||||
buffer: [0; 1024],
|
buffer: [0; 1024],
|
||||||
}
|
}
|
||||||
|
@ -230,24 +301,27 @@ impl DataStream {
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Configure the remote endpoint of the stream.
|
||||||
|
///
|
||||||
|
/// # Args
|
||||||
|
/// * `remote` - The destination to send stream data to.
|
||||||
pub fn set_remote(&mut self, remote: SocketAddr) {
|
pub fn set_remote(&mut self, remote: SocketAddr) {
|
||||||
// If the remote is identical to what we already have, do nothing.
|
// If the remote is identical to what we already have, do nothing.
|
||||||
if let Some(current_remote) = self.remote {
|
if remote == self.remote {
|
||||||
if current_remote == remote {
|
return;
|
||||||
return;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Open the new remote connection.
|
// Open the new remote connection.
|
||||||
self.open(remote).ok();
|
self.open(remote).ok();
|
||||||
self.remote = Some(remote);
|
self.remote = remote;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Process any data for transmission.
|
||||||
pub fn process(&mut self) {
|
pub fn process(&mut self) {
|
||||||
// If there's no socket available, try to connect to our remote.
|
// If there's no socket available, try to connect to our remote.
|
||||||
if self.socket.is_none() && self.remote.is_some() {
|
if self.socket.is_none() {
|
||||||
// If we still can't open the remote, continue.
|
// If we still can't open the remote, continue.
|
||||||
if self.open(self.remote.unwrap()).is_err() {
|
if self.open(self.remote).is_err() {
|
||||||
// Clear the queue out.
|
// Clear the queue out.
|
||||||
while self.queue.ready() {
|
while self.queue.ready() {
|
||||||
self.queue.dequeue();
|
self.queue.dequeue();
|
||||||
|
|
|
@ -113,11 +113,14 @@ where
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Enable live data streaming.
|
/// Enable live data streaming.
|
||||||
pub fn enable_streaming(&mut self, remote: SocketAddr) -> BlockGenerator {
|
pub fn enable_streaming(&mut self) -> BlockGenerator {
|
||||||
self.stream.set_remote(remote);
|
|
||||||
self.generator.take().unwrap()
|
self.generator.take().unwrap()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Direct the stream to the provided remote target.
|
||||||
|
///
|
||||||
|
/// # Args
|
||||||
|
/// * `remote` - The destination for the streamed data.
|
||||||
pub fn direct_stream(&mut self, remote: SocketAddr) {
|
pub fn direct_stream(&mut self, remote: SocketAddr) {
|
||||||
if self.generator.is_none() {
|
if self.generator.is_none() {
|
||||||
self.stream.set_remote(remote);
|
self.stream.set_remote(remote);
|
||||||
|
|
Loading…
Reference in New Issue