WIP merger tests
This commit is contained in:
+2
-2
@@ -14,7 +14,7 @@ pub mod bt {
|
|||||||
pub use backtrace::Backtrace;
|
pub use backtrace::Backtrace;
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, Debug, Serialize, Deserialize)]
|
#[derive(Clone, Debug, PartialEq, Serialize, Deserialize)]
|
||||||
pub enum Reason {
|
pub enum Reason {
|
||||||
InternalError,
|
InternalError,
|
||||||
BadRequest,
|
BadRequest,
|
||||||
@@ -24,7 +24,7 @@ pub enum Reason {
|
|||||||
/**
|
/**
|
||||||
The common error type for this application.
|
The common error type for this application.
|
||||||
*/
|
*/
|
||||||
#[derive(Serialize, Deserialize)]
|
#[derive(Clone, PartialEq, Serialize, Deserialize)]
|
||||||
pub struct Error {
|
pub struct Error {
|
||||||
msg: String,
|
msg: String,
|
||||||
trace_str: Option<String>,
|
trace_str: Option<String>,
|
||||||
|
|||||||
@@ -4,7 +4,8 @@ use crate::{response, response_err, BodyStream, ToPublicResponse};
|
|||||||
use futures_util::{Stream, StreamExt, TryStreamExt};
|
use futures_util::{Stream, StreamExt, TryStreamExt};
|
||||||
use http::{Method, Request, Response, StatusCode};
|
use http::{Method, Request, Response, StatusCode};
|
||||||
use hyper::Body;
|
use hyper::Body;
|
||||||
use items_2::{binned_collected, empty_events_dyn, empty_events_dyn_2, ChannelEvents, ChannelEventsMerger};
|
use items_2::merger::ChannelEventsMerger;
|
||||||
|
use items_2::{binned_collected, empty_events_dyn, empty_events_dyn_2, ChannelEvents};
|
||||||
use netpod::log::*;
|
use netpod::log::*;
|
||||||
use netpod::query::{BinnedQuery, ChannelStateEventsQuery, PlainEventsQuery};
|
use netpod::query::{BinnedQuery, ChannelStateEventsQuery, PlainEventsQuery};
|
||||||
use netpod::{AggKind, BinnedRange, FromUrl, NodeConfigCached};
|
use netpod::{AggKind, BinnedRange, FromUrl, NodeConfigCached};
|
||||||
@@ -239,7 +240,6 @@ impl EventsHandlerScylla {
|
|||||||
cl.ingest(item.as_collectable_mut());
|
cl.ingest(item.as_collectable_mut());
|
||||||
}
|
}
|
||||||
ChannelEvents::Status(..) => {}
|
ChannelEvents::Status(..) => {}
|
||||||
ChannelEvents::RangeComplete => {}
|
|
||||||
},
|
},
|
||||||
Err(e) => {
|
Err(e) => {
|
||||||
return Err(e.into());
|
return Err(e.into());
|
||||||
@@ -360,9 +360,12 @@ impl BinnedHandlerScylla {
|
|||||||
x
|
x
|
||||||
})
|
})
|
||||||
.map_err(|e| items_2::Error::from(format!("{e}")));
|
.map_err(|e| items_2::Error::from(format!("{e}")));
|
||||||
let data_stream = Box::pin(data_stream) as _;
|
todo!();
|
||||||
let state_stream = Box::pin(state_stream) as _;
|
type Items = Pin<Box<dyn Stream<Item = Result<ChannelEvents, items_2::Error>> + Send>>;
|
||||||
let merged_stream = ChannelEventsMerger::new(vec![data_stream, state_stream]);
|
let data_stream = Box::pin(data_stream) as Items;
|
||||||
|
let state_stream = Box::pin(state_stream) as Items;
|
||||||
|
let merged_stream = ChannelEventsMerger::new(todo!());
|
||||||
|
//let merged_stream = ChannelEventsMerger::new(vec![data_stream, state_stream]);
|
||||||
let merged_stream = Box::pin(merged_stream) as Pin<Box<dyn Stream<Item = _> + Send>>;
|
let merged_stream = Box::pin(merged_stream) as Pin<Box<dyn Stream<Item = _> + Send>>;
|
||||||
let binned_collected = binned_collected(
|
let binned_collected = binned_collected(
|
||||||
scalar_type.clone(),
|
scalar_type.clone(),
|
||||||
|
|||||||
+4
-4
@@ -58,27 +58,27 @@ pub fn bool_is_false(j: &bool) -> bool {
|
|||||||
*j == false
|
*j == false
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Serialize, Deserialize)]
|
#[derive(Clone, Debug, PartialEq, Serialize, Deserialize)]
|
||||||
pub enum RangeCompletableItem<T> {
|
pub enum RangeCompletableItem<T> {
|
||||||
RangeComplete,
|
RangeComplete,
|
||||||
Data(T),
|
Data(T),
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Serialize, Deserialize)]
|
#[derive(Clone, Debug, PartialEq, Serialize, Deserialize)]
|
||||||
pub enum StatsItem {
|
pub enum StatsItem {
|
||||||
EventDataReadStats(EventDataReadStats),
|
EventDataReadStats(EventDataReadStats),
|
||||||
RangeFilterStats(RangeFilterStats),
|
RangeFilterStats(RangeFilterStats),
|
||||||
DiskStats(DiskStats),
|
DiskStats(DiskStats),
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Serialize, Deserialize)]
|
#[derive(Clone, Debug, PartialEq, Serialize, Deserialize)]
|
||||||
pub enum StreamItem<T> {
|
pub enum StreamItem<T> {
|
||||||
DataItem(T),
|
DataItem(T),
|
||||||
Log(LogItem),
|
Log(LogItem),
|
||||||
Stats(StatsItem),
|
Stats(StatsItem),
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, Debug, Serialize, Deserialize)]
|
#[derive(Clone, Debug, PartialEq, Serialize, Deserialize)]
|
||||||
pub struct LogItem {
|
pub struct LogItem {
|
||||||
pub node_ix: u32,
|
pub node_ix: u32,
|
||||||
#[serde(with = "levelserde")]
|
#[serde(with = "levelserde")]
|
||||||
|
|||||||
@@ -573,6 +573,10 @@ impl<NTY: ScalarOps> Events for EventsDim0<NTY> {
|
|||||||
fn nty_id(&self) -> u32 {
|
fn nty_id(&self) -> u32 {
|
||||||
NTY::SUB
|
NTY::SUB
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn clone_dyn(&self) -> Box<dyn Events> {
|
||||||
|
Box::new(self.clone())
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub struct EventsDim0TimeBinner<NTY: ScalarOps> {
|
pub struct EventsDim0TimeBinner<NTY: ScalarOps> {
|
||||||
|
|||||||
+67
-277
@@ -1,5 +1,6 @@
|
|||||||
pub mod binsdim0;
|
pub mod binsdim0;
|
||||||
pub mod eventsdim0;
|
pub mod eventsdim0;
|
||||||
|
pub mod merger;
|
||||||
pub mod streams;
|
pub mod streams;
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
pub mod test;
|
pub mod test;
|
||||||
@@ -9,6 +10,9 @@ use futures_util::FutureExt;
|
|||||||
use futures_util::Stream;
|
use futures_util::Stream;
|
||||||
use futures_util::StreamExt;
|
use futures_util::StreamExt;
|
||||||
use items::FrameTypeInnerStatic;
|
use items::FrameTypeInnerStatic;
|
||||||
|
use items::RangeCompletableItem;
|
||||||
|
use items::Sitemty;
|
||||||
|
use items::StreamItem;
|
||||||
use items::SubFrId;
|
use items::SubFrId;
|
||||||
use netpod::log::*;
|
use netpod::log::*;
|
||||||
use netpod::timeunits::*;
|
use netpod::timeunits::*;
|
||||||
@@ -17,9 +21,7 @@ use serde::{Deserialize, Serialize, Serializer};
|
|||||||
use std::any::Any;
|
use std::any::Any;
|
||||||
use std::collections::VecDeque;
|
use std::collections::VecDeque;
|
||||||
use std::fmt;
|
use std::fmt;
|
||||||
use std::ops::ControlFlow;
|
|
||||||
use std::pin::Pin;
|
use std::pin::Pin;
|
||||||
use std::task::{Context, Poll};
|
|
||||||
use std::time::Duration;
|
use std::time::Duration;
|
||||||
use std::time::Instant;
|
use std::time::Instant;
|
||||||
use streams::Collectable;
|
use streams::Collectable;
|
||||||
@@ -139,6 +141,7 @@ pub enum ErrorKind {
|
|||||||
MismatchedType,
|
MismatchedType,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// TODO stack error better
|
||||||
#[derive(Debug, PartialEq)]
|
#[derive(Debug, PartialEq)]
|
||||||
pub struct Error {
|
pub struct Error {
|
||||||
#[allow(unused)]
|
#[allow(unused)]
|
||||||
@@ -167,6 +170,23 @@ impl From<String> for Error {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// TODO this discards structure
|
||||||
|
impl From<err::Error> for Error {
|
||||||
|
fn from(e: err::Error) -> Self {
|
||||||
|
Self {
|
||||||
|
msg: Some(format!("{e}")),
|
||||||
|
kind: ErrorKind::General,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// TODO this discards structure
|
||||||
|
impl From<Error> for err::Error {
|
||||||
|
fn from(e: Error) -> Self {
|
||||||
|
err::Error::with_msg_no_trace(format!("{e}"))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
impl std::error::Error for Error {}
|
impl std::error::Error for Error {}
|
||||||
|
|
||||||
impl serde::de::Error for Error {
|
impl serde::de::Error for Error {
|
||||||
@@ -276,6 +296,7 @@ pub trait Events: fmt::Debug + Any + Collectable + TimeBinnable + Send + erased_
|
|||||||
fn ts_min(&self) -> Option<u64>;
|
fn ts_min(&self) -> Option<u64>;
|
||||||
fn ts_max(&self) -> Option<u64>;
|
fn ts_max(&self) -> Option<u64>;
|
||||||
fn take_new_events_until_ts(&mut self, ts_end: u64) -> Box<dyn Events>;
|
fn take_new_events_until_ts(&mut self, ts_end: u64) -> Box<dyn Events>;
|
||||||
|
fn clone_dyn(&self) -> Box<dyn Events>;
|
||||||
fn partial_eq_dyn(&self, other: &dyn Events) -> bool;
|
fn partial_eq_dyn(&self, other: &dyn Events) -> bool;
|
||||||
fn serde_id(&self) -> &'static str;
|
fn serde_id(&self) -> &'static str;
|
||||||
fn nty_id(&self) -> u32;
|
fn nty_id(&self) -> u32;
|
||||||
@@ -483,8 +504,15 @@ impl<T: MergableEvents> MergableEvents for Box<T> {
|
|||||||
pub enum ChannelEvents {
|
pub enum ChannelEvents {
|
||||||
Events(Box<dyn Events>),
|
Events(Box<dyn Events>),
|
||||||
Status(ConnStatusEvent),
|
Status(ConnStatusEvent),
|
||||||
// TODO the RangeComplete event would probably fit better on some outer layer:
|
}
|
||||||
RangeComplete,
|
|
||||||
|
impl Clone for ChannelEvents {
|
||||||
|
fn clone(&self) -> Self {
|
||||||
|
match self {
|
||||||
|
Self::Events(arg0) => Self::Events(arg0.clone_dyn()),
|
||||||
|
Self::Status(arg0) => Self::Status(arg0.clone()),
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
mod serde_channel_events {
|
mod serde_channel_events {
|
||||||
@@ -510,7 +538,6 @@ mod serde_channel_events {
|
|||||||
ser.end()
|
ser.end()
|
||||||
}
|
}
|
||||||
ChannelEvents::Status(val) => serializer.serialize_newtype_variant(name, 1, "Status", val),
|
ChannelEvents::Status(val) => serializer.serialize_newtype_variant(name, 1, "Status", val),
|
||||||
ChannelEvents::RangeComplete => serializer.serialize_unit_variant(name, 2, "RangeComplete"),
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -627,7 +654,6 @@ impl MergableEvents for ChannelEvents {
|
|||||||
match self {
|
match self {
|
||||||
Events(k) => k.ts_min(),
|
Events(k) => k.ts_min(),
|
||||||
Status(k) => Some(k.ts),
|
Status(k) => Some(k.ts),
|
||||||
RangeComplete => None,
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -641,258 +667,6 @@ impl FrameTypeInnerStatic for ChannelEvents {
|
|||||||
const FRAME_TYPE_ID: u32 = items::ITEMS_2_CHANNEL_EVENTS_FRAME_TYPE_ID;
|
const FRAME_TYPE_ID: u32 = items::ITEMS_2_CHANNEL_EVENTS_FRAME_TYPE_ID;
|
||||||
}
|
}
|
||||||
|
|
||||||
type MergeInp = Pin<Box<dyn Stream<Item = Result<ChannelEvents, Error>> + Send>>;
|
|
||||||
|
|
||||||
pub struct ChannelEventsMerger {
|
|
||||||
inps: Vec<Option<MergeInp>>,
|
|
||||||
items: Vec<Option<ChannelEvents>>,
|
|
||||||
range_complete: bool,
|
|
||||||
done: bool,
|
|
||||||
done2: bool,
|
|
||||||
complete: bool,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl fmt::Debug for ChannelEventsMerger {
|
|
||||||
fn fmt(&self, fmt: &mut fmt::Formatter) -> fmt::Result {
|
|
||||||
let inps: Vec<_> = self.inps.iter().map(|x| x.is_some()).collect();
|
|
||||||
fmt.debug_struct(std::any::type_name::<Self>())
|
|
||||||
.field("inps", &inps)
|
|
||||||
.field("items", &self.items)
|
|
||||||
.field("range_complete", &self.range_complete)
|
|
||||||
.field("done", &self.done)
|
|
||||||
.field("done2", &self.done2)
|
|
||||||
.finish()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl ChannelEventsMerger {
|
|
||||||
pub fn new(inps: Vec<MergeInp>) -> Self {
|
|
||||||
let n = inps.len();
|
|
||||||
Self {
|
|
||||||
done: false,
|
|
||||||
done2: false,
|
|
||||||
complete: false,
|
|
||||||
inps: inps.into_iter().map(|x| Some(x)).collect(),
|
|
||||||
items: (0..n).into_iter().map(|_| None).collect(),
|
|
||||||
range_complete: false,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
fn process(mut self: Pin<&mut Self>, _cx: &mut Context) -> Result<ControlFlow<ChannelEvents>, Error> {
|
|
||||||
use ControlFlow::*;
|
|
||||||
let mut tslows = [None, None];
|
|
||||||
for (i1, itemopt) in self.items.iter_mut().enumerate() {
|
|
||||||
if let Some(item) = itemopt {
|
|
||||||
let t1 = item.ts_min();
|
|
||||||
if let Some(t1) = t1 {
|
|
||||||
if let Some((_, a)) = tslows[0] {
|
|
||||||
if t1 < a {
|
|
||||||
tslows[1] = tslows[0];
|
|
||||||
tslows[0] = Some((i1, t1));
|
|
||||||
} else {
|
|
||||||
if let Some((_, b)) = tslows[1] {
|
|
||||||
if t1 < b {
|
|
||||||
tslows[1] = Some((i1, t1));
|
|
||||||
} else {
|
|
||||||
// nothing to do
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
tslows[1] = Some((i1, t1));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
tslows[0] = Some((i1, t1));
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
match item {
|
|
||||||
ChannelEvents::Events(_) => {
|
|
||||||
trace!("events item without ts min discovered {item:?}");
|
|
||||||
itemopt.take();
|
|
||||||
return Ok(Continue(()));
|
|
||||||
}
|
|
||||||
ChannelEvents::Status(_) => {
|
|
||||||
return Err(format!("channel status without timestamp").into());
|
|
||||||
}
|
|
||||||
ChannelEvents::RangeComplete => {
|
|
||||||
trace!("--------------------- ChannelEvents::RangeComplete \n======================");
|
|
||||||
*itemopt = None;
|
|
||||||
self.range_complete = true;
|
|
||||||
return Ok(Continue(()));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if let Some((il0, _tl0)) = tslows[0] {
|
|
||||||
if let Some((_il1, tl1)) = tslows[1] {
|
|
||||||
let item = self.items[il0].as_mut().unwrap();
|
|
||||||
match item {
|
|
||||||
ChannelEvents::Events(item) => {
|
|
||||||
if let Some(th0) = item.ts_max() {
|
|
||||||
if th0 < tl1 {
|
|
||||||
let ret = self.items[il0].take().unwrap();
|
|
||||||
Ok(Break(ret))
|
|
||||||
} else {
|
|
||||||
let ritem = item.take_new_events_until_ts(tl1);
|
|
||||||
if item.len() == 0 {
|
|
||||||
// TODO should never be here
|
|
||||||
self.items[il0] = None;
|
|
||||||
}
|
|
||||||
Ok(Break(ChannelEvents::Events(ritem)))
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
// TODO should never be here because ts-max should always exist here.
|
|
||||||
let ritem = item.take_new_events_until_ts(tl1);
|
|
||||||
if item.len() == 0 {}
|
|
||||||
Ok(Break(ChannelEvents::Events(ritem)))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
ChannelEvents::Status(_) => {
|
|
||||||
let ret = self.items[il0].take().unwrap();
|
|
||||||
Ok(Break(ret))
|
|
||||||
}
|
|
||||||
ChannelEvents::RangeComplete => Err(format!("RangeComplete considered in merge-lowest").into()),
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
let item = self.items[il0].as_mut().unwrap();
|
|
||||||
match item {
|
|
||||||
ChannelEvents::Events(_) => {
|
|
||||||
let ret = self.items[il0].take().unwrap();
|
|
||||||
Ok(Break(ret))
|
|
||||||
}
|
|
||||||
ChannelEvents::Status(_) => {
|
|
||||||
let ret = self.items[il0].take().unwrap();
|
|
||||||
Ok(Break(ret))
|
|
||||||
}
|
|
||||||
ChannelEvents::RangeComplete => Err(format!("RangeComplete considered in merge-lowest").into()),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
Err(format!("after low ts search nothing found").into())
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
fn refill(mut self: Pin<&mut Self>, cx: &mut Context) -> ControlFlow<Poll<Error>> {
|
|
||||||
use ControlFlow::*;
|
|
||||||
use Poll::*;
|
|
||||||
let mut has_pending = false;
|
|
||||||
for i1 in 0..self.inps.len() {
|
|
||||||
let item = &self.items[i1];
|
|
||||||
if item.is_none() {
|
|
||||||
if let Some(inp) = &mut self.inps[i1] {
|
|
||||||
match inp.poll_next_unpin(cx) {
|
|
||||||
Ready(Some(Ok(k))) => {
|
|
||||||
if let ChannelEvents::Events(events) = &k {
|
|
||||||
if events.len() == 0 {
|
|
||||||
warn!("empty events item {events:?}");
|
|
||||||
} else {
|
|
||||||
trace!("\nrefilled with events {}\nREFILLED\n{:?}\n\n", events.len(), events);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
self.items[i1] = Some(k);
|
|
||||||
}
|
|
||||||
Ready(Some(Err(e))) => return Break(Ready(e)),
|
|
||||||
Ready(None) => {
|
|
||||||
self.inps[i1] = None;
|
|
||||||
}
|
|
||||||
Pending => {
|
|
||||||
has_pending = true;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if has_pending {
|
|
||||||
Break(Pending)
|
|
||||||
} else {
|
|
||||||
Continue(())
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
fn poll2(mut self: Pin<&mut Self>, cx: &mut Context) -> ControlFlow<Poll<Option<<Self as Stream>::Item>>> {
|
|
||||||
use ControlFlow::*;
|
|
||||||
use Poll::*;
|
|
||||||
let mut has_pending = false;
|
|
||||||
match Self::refill(Pin::new(&mut self), cx) {
|
|
||||||
Break(Ready(e)) => return Break(Ready(Some(Err(e)))),
|
|
||||||
Break(Pending) => {
|
|
||||||
has_pending = true;
|
|
||||||
}
|
|
||||||
Continue(()) => {}
|
|
||||||
}
|
|
||||||
let ninps = self.inps.iter().filter(|a| a.is_some()).count();
|
|
||||||
let nitems = self.items.iter().filter(|a| a.is_some()).count();
|
|
||||||
let nitemsmissing = self
|
|
||||||
.inps
|
|
||||||
.iter()
|
|
||||||
.zip(self.items.iter())
|
|
||||||
.filter(|(a, b)| a.is_some() && b.is_none())
|
|
||||||
.count();
|
|
||||||
if ninps == 0 && nitems == 0 {
|
|
||||||
self.done = true;
|
|
||||||
Break(Ready(None))
|
|
||||||
} else if nitemsmissing != 0 {
|
|
||||||
if !has_pending {
|
|
||||||
let e = Error::from(format!("missing but no pending"));
|
|
||||||
Break(Ready(Some(Err(e))))
|
|
||||||
} else {
|
|
||||||
Break(Pending)
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
match Self::process(Pin::new(&mut self), cx) {
|
|
||||||
Ok(Break(item)) => Break(Ready(Some(Ok(item)))),
|
|
||||||
Ok(Continue(())) => Continue(()),
|
|
||||||
Err(e) => Break(Ready(Some(Err(e)))),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl Stream for ChannelEventsMerger {
|
|
||||||
type Item = Result<ChannelEvents, Error>;
|
|
||||||
|
|
||||||
fn poll_next(mut self: Pin<&mut Self>, cx: &mut Context) -> Poll<Option<Self::Item>> {
|
|
||||||
use Poll::*;
|
|
||||||
//let self_name = std::any::type_name::<Self>();
|
|
||||||
//eprintln!("{self_name} poll_next");
|
|
||||||
loop {
|
|
||||||
break if self.complete {
|
|
||||||
panic!("poll after complete");
|
|
||||||
} else if self.done2 {
|
|
||||||
self.complete = true;
|
|
||||||
Ready(None)
|
|
||||||
} else if self.done {
|
|
||||||
self.done2 = true;
|
|
||||||
if self.range_complete {
|
|
||||||
trace!("MERGER EMITTING ChannelEvents::RangeComplete");
|
|
||||||
Ready(Some(Ok(ChannelEvents::RangeComplete)))
|
|
||||||
} else {
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
match Self::poll2(self.as_mut(), cx) {
|
|
||||||
ControlFlow::Continue(()) => continue,
|
|
||||||
ControlFlow::Break(k) => {
|
|
||||||
match &k {
|
|
||||||
Ready(Some(Ok(ChannelEvents::Events(item)))) => {
|
|
||||||
trace!("\n\nMERGER EMITTING\n{:?}\n\n", item);
|
|
||||||
}
|
|
||||||
Ready(Some(Ok(ChannelEvents::RangeComplete))) => {
|
|
||||||
trace!("\nLOGIC ERROR MERGER EMITTING PLAIN ChannelEvents::RangeComplete");
|
|
||||||
}
|
|
||||||
Ready(Some(Err(_))) => {
|
|
||||||
self.done = true;
|
|
||||||
}
|
|
||||||
_ => {}
|
|
||||||
}
|
|
||||||
k
|
|
||||||
}
|
|
||||||
}
|
|
||||||
};
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// TODO do this with some blanket impl:
|
// TODO do this with some blanket impl:
|
||||||
impl Collectable for Box<dyn Collectable> {
|
impl Collectable for Box<dyn Collectable> {
|
||||||
fn new_collector(&self, bin_count_exp: u32) -> Box<dyn streams::Collector> {
|
fn new_collector(&self, bin_count_exp: u32) -> Box<dyn streams::Collector> {
|
||||||
@@ -911,8 +685,9 @@ pub async fn binned_collected(
|
|||||||
agg_kind: AggKind,
|
agg_kind: AggKind,
|
||||||
edges: Vec<u64>,
|
edges: Vec<u64>,
|
||||||
timeout: Duration,
|
timeout: Duration,
|
||||||
inp: Pin<Box<dyn Stream<Item = Result<ChannelEvents, Error>> + Send>>,
|
inp: Pin<Box<dyn Stream<Item = Sitemty<ChannelEvents>> + Send>>,
|
||||||
) -> Result<Box<dyn ToJsonResult>, Error> {
|
) -> Result<Box<dyn ToJsonResult>, Error> {
|
||||||
|
info!("binned_collected");
|
||||||
let deadline = Instant::now() + timeout;
|
let deadline = Instant::now() + timeout;
|
||||||
let mut did_timeout = false;
|
let mut did_timeout = false;
|
||||||
let bin_count_exp = edges.len().max(2) as u32 - 1;
|
let bin_count_exp = edges.len().max(2) as u32 - 1;
|
||||||
@@ -925,10 +700,10 @@ pub async fn binned_collected(
|
|||||||
bin_count_exp: u32,
|
bin_count_exp: u32,
|
||||||
force: bool,
|
force: bool,
|
||||||
) -> Result<(), Error> {
|
) -> Result<(), Error> {
|
||||||
//info!("bins_ready_count: {}", binner.bins_ready_count());
|
info!("flush_binned bins_ready_count: {}", binner.bins_ready_count());
|
||||||
if force {
|
if force {
|
||||||
if binner.bins_ready_count() == 0 {
|
if binner.bins_ready_count() == 0 {
|
||||||
warn!("cycle the binner");
|
warn!("cycle the binner forced");
|
||||||
binner.cycle();
|
binner.cycle();
|
||||||
} else {
|
} else {
|
||||||
warn!("binner was some ready, do nothing");
|
warn!("binner was some ready, do nothing");
|
||||||
@@ -938,7 +713,7 @@ pub async fn binned_collected(
|
|||||||
let ready = binner.bins_ready();
|
let ready = binner.bins_ready();
|
||||||
match ready {
|
match ready {
|
||||||
Some(mut ready) => {
|
Some(mut ready) => {
|
||||||
trace!("binned_collected ready {ready:?}");
|
info!("binned_collected ready {ready:?}");
|
||||||
if coll.is_none() {
|
if coll.is_none() {
|
||||||
*coll = Some(ready.as_collectable_mut().new_collector(bin_count_exp));
|
*coll = Some(ready.as_collectable_mut().new_collector(bin_count_exp));
|
||||||
}
|
}
|
||||||
@@ -955,7 +730,9 @@ pub async fn binned_collected(
|
|||||||
let mut coll = None;
|
let mut coll = None;
|
||||||
let mut binner = None;
|
let mut binner = None;
|
||||||
let empty_item = empty_events_dyn_2(&scalar_type, &shape, &AggKind::TimeWeightedScalar);
|
let empty_item = empty_events_dyn_2(&scalar_type, &shape, &AggKind::TimeWeightedScalar);
|
||||||
let empty_stream = futures_util::stream::once(futures_util::future::ready(Ok(ChannelEvents::Events(empty_item))));
|
let empty_stream = futures_util::stream::once(futures_util::future::ready(Ok(StreamItem::DataItem(
|
||||||
|
RangeCompletableItem::Data(ChannelEvents::Events(empty_item)),
|
||||||
|
))));
|
||||||
let mut stream = empty_stream.chain(inp);
|
let mut stream = empty_stream.chain(inp);
|
||||||
loop {
|
loop {
|
||||||
let item = futures_util::select! {
|
let item = futures_util::select! {
|
||||||
@@ -972,22 +749,34 @@ pub async fn binned_collected(
|
|||||||
}
|
}
|
||||||
};
|
};
|
||||||
match item {
|
match item {
|
||||||
ChannelEvents::Events(events) => {
|
StreamItem::DataItem(k) => match k {
|
||||||
trace!("binned_collected sees\n{:?}", events);
|
RangeCompletableItem::RangeComplete => {
|
||||||
if binner.is_none() {
|
warn!("binned_collected TODO RangeComplete");
|
||||||
let bb = events.as_time_binnable().time_binner_new(edges.clone(), do_time_weight);
|
did_range_complete = true;
|
||||||
binner = Some(bb);
|
|
||||||
}
|
}
|
||||||
let binner = binner.as_mut().unwrap();
|
RangeCompletableItem::Data(k) => match k {
|
||||||
binner.ingest(events.as_time_binnable());
|
ChannelEvents::Events(events) => {
|
||||||
flush_binned(binner, &mut coll, bin_count_exp, false)?;
|
info!("binned_collected sees\n{:?}", events);
|
||||||
|
if binner.is_none() {
|
||||||
|
let bb = events.as_time_binnable().time_binner_new(edges.clone(), do_time_weight);
|
||||||
|
binner = Some(bb);
|
||||||
|
}
|
||||||
|
let binner = binner.as_mut().unwrap();
|
||||||
|
binner.ingest(events.as_time_binnable());
|
||||||
|
flush_binned(binner, &mut coll, bin_count_exp, false)?;
|
||||||
|
}
|
||||||
|
ChannelEvents::Status(item) => {
|
||||||
|
info!("{:?}", item);
|
||||||
|
}
|
||||||
|
},
|
||||||
|
},
|
||||||
|
StreamItem::Log(item) => {
|
||||||
|
// TODO collect also errors here?
|
||||||
|
info!("{:?}", item);
|
||||||
}
|
}
|
||||||
ChannelEvents::Status(_) => {
|
StreamItem::Stats(item) => {
|
||||||
warn!("binned_collected TODO Status");
|
// TODO do something with the stats
|
||||||
}
|
info!("{:?}", item);
|
||||||
ChannelEvents::RangeComplete => {
|
|
||||||
warn!("binned_collected TODO RangeComplete");
|
|
||||||
did_range_complete = true;
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -996,6 +785,7 @@ pub async fn binned_collected(
|
|||||||
binner.set_range_complete();
|
binner.set_range_complete();
|
||||||
}
|
}
|
||||||
if !did_timeout {
|
if !did_timeout {
|
||||||
|
warn!("cycle the binner");
|
||||||
binner.cycle();
|
binner.cycle();
|
||||||
}
|
}
|
||||||
flush_binned(&mut binner, &mut coll, bin_count_exp, false)?;
|
flush_binned(&mut binner, &mut coll, bin_count_exp, false)?;
|
||||||
|
|||||||
@@ -0,0 +1,289 @@
|
|||||||
|
use crate::{ChannelEvents, Error, MergableEvents};
|
||||||
|
use futures_util::{Stream, StreamExt};
|
||||||
|
use items::{RangeCompletableItem, Sitemty, StreamItem};
|
||||||
|
use netpod::log::*;
|
||||||
|
use std::fmt;
|
||||||
|
use std::ops::ControlFlow;
|
||||||
|
use std::pin::Pin;
|
||||||
|
use std::task::{Context, Poll};
|
||||||
|
|
||||||
|
type MergeInp = Pin<Box<dyn Stream<Item = Sitemty<ChannelEvents>> + Send>>;
|
||||||
|
|
||||||
|
pub struct ChannelEventsMerger {
|
||||||
|
inps: Vec<Option<MergeInp>>,
|
||||||
|
items: Vec<Option<ChannelEvents>>,
|
||||||
|
range_complete: bool,
|
||||||
|
done: bool,
|
||||||
|
done2: bool,
|
||||||
|
complete: bool,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl fmt::Debug for ChannelEventsMerger {
|
||||||
|
fn fmt(&self, fmt: &mut fmt::Formatter) -> fmt::Result {
|
||||||
|
let inps: Vec<_> = self.inps.iter().map(|x| x.is_some()).collect();
|
||||||
|
fmt.debug_struct(std::any::type_name::<Self>())
|
||||||
|
.field("inps", &inps)
|
||||||
|
.field("items", &self.items)
|
||||||
|
.field("range_complete", &self.range_complete)
|
||||||
|
.field("done", &self.done)
|
||||||
|
.field("done2", &self.done2)
|
||||||
|
.finish()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl ChannelEventsMerger {
|
||||||
|
pub fn new(inps: Vec<MergeInp>) -> Self {
|
||||||
|
let n = inps.len();
|
||||||
|
Self {
|
||||||
|
done: false,
|
||||||
|
done2: false,
|
||||||
|
complete: false,
|
||||||
|
inps: inps.into_iter().map(|x| Some(x)).collect(),
|
||||||
|
items: (0..n).into_iter().map(|_| None).collect(),
|
||||||
|
range_complete: false,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn process(mut self: Pin<&mut Self>, _cx: &mut Context) -> Result<ControlFlow<ChannelEvents>, Error> {
|
||||||
|
use ControlFlow::*;
|
||||||
|
let mut tslows = [None, None];
|
||||||
|
for (i1, itemopt) in self.items.iter_mut().enumerate() {
|
||||||
|
if let Some(item) = itemopt {
|
||||||
|
let t1 = item.ts_min();
|
||||||
|
if let Some(t1) = t1 {
|
||||||
|
if let Some((_, a)) = tslows[0] {
|
||||||
|
if t1 < a {
|
||||||
|
tslows[1] = tslows[0];
|
||||||
|
tslows[0] = Some((i1, t1));
|
||||||
|
} else {
|
||||||
|
if let Some((_, b)) = tslows[1] {
|
||||||
|
if t1 < b {
|
||||||
|
tslows[1] = Some((i1, t1));
|
||||||
|
} else {
|
||||||
|
// nothing to do
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
tslows[1] = Some((i1, t1));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
tslows[0] = Some((i1, t1));
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
match item {
|
||||||
|
ChannelEvents::Events(_) => {
|
||||||
|
trace!("events item without ts min discovered {item:?}");
|
||||||
|
itemopt.take();
|
||||||
|
return Ok(Continue(()));
|
||||||
|
}
|
||||||
|
ChannelEvents::Status(_) => {
|
||||||
|
return Err(format!("channel status without timestamp").into());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if let Some((il0, _tl0)) = tslows[0] {
|
||||||
|
if let Some((_il1, tl1)) = tslows[1] {
|
||||||
|
let item = self.items[il0].as_mut().unwrap();
|
||||||
|
match item {
|
||||||
|
ChannelEvents::Events(item) => {
|
||||||
|
if let Some(th0) = item.ts_max() {
|
||||||
|
if th0 < tl1 {
|
||||||
|
let ret = self.items[il0].take().unwrap();
|
||||||
|
Ok(Break(ret))
|
||||||
|
} else {
|
||||||
|
let ritem = item.take_new_events_until_ts(tl1);
|
||||||
|
if item.len() == 0 {
|
||||||
|
// TODO should never be here
|
||||||
|
self.items[il0] = None;
|
||||||
|
}
|
||||||
|
Ok(Break(ChannelEvents::Events(ritem)))
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
// TODO should never be here because ts-max should always exist here.
|
||||||
|
let ritem = item.take_new_events_until_ts(tl1);
|
||||||
|
if item.len() == 0 {}
|
||||||
|
Ok(Break(ChannelEvents::Events(ritem)))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
ChannelEvents::Status(_) => {
|
||||||
|
let ret = self.items[il0].take().unwrap();
|
||||||
|
Ok(Break(ret))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
let item = self.items[il0].as_mut().unwrap();
|
||||||
|
match item {
|
||||||
|
ChannelEvents::Events(_) => {
|
||||||
|
let ret = self.items[il0].take().unwrap();
|
||||||
|
Ok(Break(ret))
|
||||||
|
}
|
||||||
|
ChannelEvents::Status(_) => {
|
||||||
|
let ret = self.items[il0].take().unwrap();
|
||||||
|
Ok(Break(ret))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
Err(format!("after low ts search nothing found").into())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn refill(mut self: Pin<&mut Self>, cx: &mut Context) -> ControlFlow<Poll<Error>> {
|
||||||
|
use ControlFlow::*;
|
||||||
|
use Poll::*;
|
||||||
|
let mut has_pending = false;
|
||||||
|
for i1 in 0..self.inps.len() {
|
||||||
|
let item = &self.items[i1];
|
||||||
|
if item.is_none() {
|
||||||
|
while let Some(inp) = &mut self.inps[i1] {
|
||||||
|
match inp.poll_next_unpin(cx) {
|
||||||
|
Ready(Some(Ok(k))) => {
|
||||||
|
match k {
|
||||||
|
StreamItem::DataItem(k) => match k {
|
||||||
|
RangeCompletableItem::RangeComplete => {
|
||||||
|
trace!("--------------------- ChannelEvents::RangeComplete \n======================");
|
||||||
|
// TODO track range complete for all inputs, it's only complete if all inputs are complete.
|
||||||
|
self.range_complete = true;
|
||||||
|
eprintln!("TODO inp RangeComplete which does not fill slot");
|
||||||
|
}
|
||||||
|
RangeCompletableItem::Data(k) => {
|
||||||
|
match &k {
|
||||||
|
ChannelEvents::Events(events) => {
|
||||||
|
if events.len() == 0 {
|
||||||
|
warn!("empty events item {events:?}");
|
||||||
|
} else {
|
||||||
|
trace!(
|
||||||
|
"\nrefilled with events {}\nREFILLED\n{:?}\n\n",
|
||||||
|
events.len(),
|
||||||
|
events
|
||||||
|
);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
ChannelEvents::Status(_) => {
|
||||||
|
eprintln!("TODO inp Status which does not fill slot");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
self.items[i1] = Some(k);
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
},
|
||||||
|
StreamItem::Log(_) => {
|
||||||
|
eprintln!("TODO inp Log which does not fill slot");
|
||||||
|
}
|
||||||
|
StreamItem::Stats(_) => {
|
||||||
|
eprintln!("TODO inp Stats which does not fill slot");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Ready(Some(Err(e))) => return Break(Ready(e.into())),
|
||||||
|
Ready(None) => {
|
||||||
|
self.inps[i1] = None;
|
||||||
|
}
|
||||||
|
Pending => {
|
||||||
|
has_pending = true;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if has_pending {
|
||||||
|
Break(Pending)
|
||||||
|
} else {
|
||||||
|
Continue(())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn poll2(mut self: Pin<&mut Self>, cx: &mut Context) -> ControlFlow<Poll<Option<Result<ChannelEvents, Error>>>> {
|
||||||
|
use ControlFlow::*;
|
||||||
|
use Poll::*;
|
||||||
|
let mut has_pending = false;
|
||||||
|
match Self::refill(Pin::new(&mut self), cx) {
|
||||||
|
Break(Ready(e)) => return Break(Ready(Some(Err(e)))),
|
||||||
|
Break(Pending) => {
|
||||||
|
has_pending = true;
|
||||||
|
}
|
||||||
|
Continue(()) => {}
|
||||||
|
}
|
||||||
|
let ninps = self.inps.iter().filter(|a| a.is_some()).count();
|
||||||
|
let nitems = self.items.iter().filter(|a| a.is_some()).count();
|
||||||
|
let nitemsmissing = self
|
||||||
|
.inps
|
||||||
|
.iter()
|
||||||
|
.zip(self.items.iter())
|
||||||
|
.filter(|(a, b)| a.is_some() && b.is_none())
|
||||||
|
.count();
|
||||||
|
if ninps == 0 && nitems == 0 {
|
||||||
|
self.done = true;
|
||||||
|
Break(Ready(None))
|
||||||
|
} else if nitemsmissing != 0 {
|
||||||
|
if !has_pending {
|
||||||
|
let e = Error::from(format!("missing but no pending"));
|
||||||
|
Break(Ready(Some(Err(e))))
|
||||||
|
} else {
|
||||||
|
Break(Pending)
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
match Self::process(Pin::new(&mut self), cx) {
|
||||||
|
Ok(Break(item)) => Break(Ready(Some(Ok(item)))),
|
||||||
|
Ok(Continue(())) => Continue(()),
|
||||||
|
Err(e) => Break(Ready(Some(Err(e)))),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Stream for ChannelEventsMerger {
|
||||||
|
type Item = Sitemty<ChannelEvents>;
|
||||||
|
|
||||||
|
fn poll_next(mut self: Pin<&mut Self>, cx: &mut Context) -> Poll<Option<Self::Item>> {
|
||||||
|
use Poll::*;
|
||||||
|
const NAME: &str = "ChannelEventsMerger";
|
||||||
|
let span = span!(Level::TRACE, NAME);
|
||||||
|
let _spanguard = span.enter();
|
||||||
|
loop {
|
||||||
|
break if self.complete {
|
||||||
|
panic!("poll after complete");
|
||||||
|
} else if self.done2 {
|
||||||
|
self.complete = true;
|
||||||
|
Ready(None)
|
||||||
|
} else if self.done {
|
||||||
|
self.done2 = true;
|
||||||
|
if self.range_complete {
|
||||||
|
trace!("MERGER EMITTING ChannelEvents::RangeComplete");
|
||||||
|
Ready(Some(Ok(StreamItem::DataItem(RangeCompletableItem::RangeComplete))))
|
||||||
|
} else {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
match Self::poll2(self.as_mut(), cx) {
|
||||||
|
ControlFlow::Continue(()) => continue,
|
||||||
|
ControlFlow::Break(k) => match k {
|
||||||
|
Ready(Some(Ok(ChannelEvents::Events(item)))) => {
|
||||||
|
trace!("\n\nMERGER EMITTING\n{:?}\n\n", item);
|
||||||
|
Ready(Some(Ok(StreamItem::DataItem(RangeCompletableItem::Data(
|
||||||
|
ChannelEvents::Events(item),
|
||||||
|
)))))
|
||||||
|
}
|
||||||
|
Ready(Some(Ok(ChannelEvents::Status(item)))) => {
|
||||||
|
trace!("\n\nMERGER EMITTING\n{:?}\n\n", item);
|
||||||
|
Ready(Some(Ok(StreamItem::DataItem(RangeCompletableItem::Data(
|
||||||
|
ChannelEvents::Status(item),
|
||||||
|
)))))
|
||||||
|
}
|
||||||
|
Ready(Some(Err(e))) => {
|
||||||
|
self.done = true;
|
||||||
|
Ready(Some(Err(e.into())))
|
||||||
|
}
|
||||||
|
Ready(None) => {
|
||||||
|
self.done = true;
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
Pending => Pending,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
};
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
+121
-86
@@ -1,9 +1,11 @@
|
|||||||
use crate::binsdim0::BinsDim0CollectedResult;
|
use crate::binsdim0::BinsDim0CollectedResult;
|
||||||
use crate::eventsdim0::EventsDim0;
|
use crate::eventsdim0::EventsDim0;
|
||||||
use crate::{binned_collected, ChannelEvents, ChannelEventsMerger, Empty, IsoDateTime};
|
use crate::merger::ChannelEventsMerger;
|
||||||
|
use crate::{binned_collected, ChannelEvents, Empty, Events, IsoDateTime};
|
||||||
use crate::{ConnStatus, ConnStatusEvent, Error};
|
use crate::{ConnStatus, ConnStatusEvent, Error};
|
||||||
use chrono::{TimeZone, Utc};
|
use chrono::{TimeZone, Utc};
|
||||||
use futures_util::StreamExt;
|
use futures_util::StreamExt;
|
||||||
|
use items::{RangeCompletableItem, Sitemty, StreamItem};
|
||||||
use netpod::timeunits::*;
|
use netpod::timeunits::*;
|
||||||
use netpod::{AggKind, BinnedRange, NanoRange, ScalarType, Shape};
|
use netpod::{AggKind, BinnedRange, NanoRange, ScalarType, Shape};
|
||||||
use std::time::Duration;
|
use std::time::Duration;
|
||||||
@@ -11,20 +13,22 @@ use std::time::Duration;
|
|||||||
#[test]
|
#[test]
|
||||||
fn merge01() {
|
fn merge01() {
|
||||||
let fut = async {
|
let fut = async {
|
||||||
let mut events_vec1 = Vec::new();
|
let mut events_vec1: Vec<Sitemty<ChannelEvents>> = Vec::new();
|
||||||
let mut events_vec2 = Vec::new();
|
let mut events_vec2: Vec<Sitemty<ChannelEvents>> = Vec::new();
|
||||||
{
|
{
|
||||||
let mut events = EventsDim0::empty();
|
let mut events = EventsDim0::empty();
|
||||||
for i in 0..10 {
|
for i in 0..10 {
|
||||||
events.push(i * 100, i, i as f32 * 100.);
|
events.push(i * 100, i, i as f32 * 100.);
|
||||||
}
|
}
|
||||||
events_vec1.push(Ok(ChannelEvents::Events(Box::new(events.clone()))));
|
let cev = ChannelEvents::Events(Box::new(events.clone()));
|
||||||
events_vec2.push(Ok(ChannelEvents::Events(Box::new(events.clone()))));
|
events_vec1.push(Ok(StreamItem::DataItem(RangeCompletableItem::Data(cev))));
|
||||||
|
let cev = ChannelEvents::Events(Box::new(events.clone()));
|
||||||
|
events_vec2.push(Ok(StreamItem::DataItem(RangeCompletableItem::Data(cev))));
|
||||||
}
|
}
|
||||||
let inp1 = events_vec1;
|
let inp1 = events_vec1;
|
||||||
let inp1 = futures_util::stream::iter(inp1);
|
let inp1 = futures_util::stream::iter(inp1);
|
||||||
let inp1 = Box::pin(inp1);
|
let inp1 = Box::pin(inp1);
|
||||||
let inp2: Vec<Result<ChannelEvents, Error>> = Vec::new();
|
let inp2: Vec<Sitemty<ChannelEvents>> = Vec::new();
|
||||||
let inp2 = futures_util::stream::iter(inp2);
|
let inp2 = futures_util::stream::iter(inp2);
|
||||||
let inp2 = Box::pin(inp2);
|
let inp2 = Box::pin(inp2);
|
||||||
let mut merger = ChannelEventsMerger::new(vec![inp1, inp2]);
|
let mut merger = ChannelEventsMerger::new(vec![inp1, inp2]);
|
||||||
@@ -39,91 +43,106 @@ fn merge01() {
|
|||||||
#[test]
|
#[test]
|
||||||
fn merge02() {
|
fn merge02() {
|
||||||
let fut = async {
|
let fut = async {
|
||||||
let mut events_vec1 = Vec::new();
|
let events_vec1 = {
|
||||||
let mut events_vec2 = Vec::new();
|
let mut vec = Vec::new();
|
||||||
{
|
|
||||||
let mut events = EventsDim0::empty();
|
let mut events = EventsDim0::empty();
|
||||||
for i in 0..10 {
|
for i in 0..10 {
|
||||||
events.push(i * 100, i, i as f32 * 100.);
|
events.push(i * 100, i, i as f32 * 100.);
|
||||||
}
|
}
|
||||||
events_vec1.push(Ok(ChannelEvents::Events(Box::new(events.clone()))));
|
push_evd0(&mut vec, Box::new(events.clone()));
|
||||||
events_vec2.push(Ok(ChannelEvents::Events(Box::new(events.clone()))));
|
|
||||||
}
|
|
||||||
{
|
|
||||||
let mut events = EventsDim0::empty();
|
let mut events = EventsDim0::empty();
|
||||||
for i in 10..20 {
|
for i in 10..20 {
|
||||||
events.push(i * 100, i, i as f32 * 100.);
|
events.push(i * 100, i, i as f32 * 100.);
|
||||||
}
|
}
|
||||||
events_vec1.push(Ok(ChannelEvents::Events(Box::new(events.clone()))));
|
push_evd0(&mut vec, Box::new(events.clone()));
|
||||||
events_vec2.push(Ok(ChannelEvents::Events(Box::new(events.clone()))));
|
vec
|
||||||
}
|
};
|
||||||
|
let exp = events_vec1.clone();
|
||||||
let inp1 = events_vec1;
|
let inp1 = events_vec1;
|
||||||
let inp1 = futures_util::stream::iter(inp1);
|
let inp1 = futures_util::stream::iter(inp1);
|
||||||
let inp1 = Box::pin(inp1);
|
let inp1 = Box::pin(inp1);
|
||||||
let inp2: Vec<Result<ChannelEvents, Error>> = Vec::new();
|
let inp2: Vec<Sitemty<ChannelEvents>> = Vec::new();
|
||||||
let inp2 = futures_util::stream::iter(inp2);
|
let inp2 = futures_util::stream::iter(inp2);
|
||||||
let inp2 = Box::pin(inp2);
|
let inp2 = Box::pin(inp2);
|
||||||
let mut merger = ChannelEventsMerger::new(vec![inp1, inp2]);
|
let mut merger = ChannelEventsMerger::new(vec![inp1, inp2]);
|
||||||
let item = merger.next().await;
|
let item = merger.next().await;
|
||||||
assert_eq!(item.as_ref(), events_vec2.get(0));
|
assert_eq!(item.as_ref(), exp.get(0));
|
||||||
let item = merger.next().await;
|
let item = merger.next().await;
|
||||||
assert_eq!(item.as_ref(), events_vec2.get(1));
|
assert_eq!(item.as_ref(), exp.get(1));
|
||||||
let item = merger.next().await;
|
let item = merger.next().await;
|
||||||
assert_eq!(item.as_ref(), None);
|
assert_eq!(item.as_ref(), None);
|
||||||
};
|
};
|
||||||
tokio::runtime::Runtime::new().unwrap().block_on(fut);
|
tokio::runtime::Runtime::new().unwrap().block_on(fut);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn push_evd0(vec: &mut Vec<Sitemty<ChannelEvents>>, events: Box<dyn Events>) {
|
||||||
|
let cev = ChannelEvents::Events(events);
|
||||||
|
vec.push(Ok(StreamItem::DataItem(RangeCompletableItem::Data(cev))));
|
||||||
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn merge03() {
|
fn merge03() {
|
||||||
let fut = async {
|
let fut = async {
|
||||||
let mut events_vec1 = Vec::new();
|
let events_vec1 = {
|
||||||
{
|
let mut vec = Vec::new();
|
||||||
let mut events = EventsDim0::empty();
|
let mut events = EventsDim0::empty();
|
||||||
for i in 0..10 {
|
for i in 0..10 {
|
||||||
events.push(i * 100, i, i as f32 * 100.);
|
events.push(i * 100, i, i as f32 * 100.);
|
||||||
}
|
}
|
||||||
events_vec1.push(Ok(ChannelEvents::Events(Box::new(events))));
|
push_evd0(&mut vec, Box::new(events));
|
||||||
let mut events = EventsDim0::empty();
|
let mut events = EventsDim0::empty();
|
||||||
for i in 10..20 {
|
for i in 10..20 {
|
||||||
events.push(i * 100, i, i as f32 * 100.);
|
events.push(i * 100, i, i as f32 * 100.);
|
||||||
}
|
}
|
||||||
events_vec1.push(Ok(ChannelEvents::Events(Box::new(events.clone()))));
|
push_evd0(&mut vec, Box::new(events));
|
||||||
}
|
vec
|
||||||
let events_vec1 = events_vec1;
|
};
|
||||||
let mut events_vec2 = Vec::new();
|
let events_vec2 = {
|
||||||
{
|
let mut vec = Vec::new();
|
||||||
let mut events = EventsDim0::empty();
|
let mut events = EventsDim0::empty();
|
||||||
for i in 0..10 {
|
for i in 0..10 {
|
||||||
events.push(i * 100, i, i as f32 * 100.);
|
events.push(i * 100, i, i as f32 * 100.);
|
||||||
}
|
}
|
||||||
events_vec2.push(Ok(ChannelEvents::Events(Box::new(events.clone()))));
|
push_evd0(&mut vec, Box::new(events));
|
||||||
let mut events = EventsDim0::empty();
|
let mut events = EventsDim0::empty();
|
||||||
for i in 10..12 {
|
for i in 10..12 {
|
||||||
events.push(i * 100, i, i as f32 * 100.);
|
events.push(i * 100, i, i as f32 * 100.);
|
||||||
}
|
}
|
||||||
events_vec2.push(Ok(ChannelEvents::Events(Box::new(events.clone()))));
|
push_evd0(&mut vec, Box::new(events));
|
||||||
let mut events = EventsDim0::empty();
|
let mut events = EventsDim0::empty();
|
||||||
for i in 12..20 {
|
for i in 12..20 {
|
||||||
events.push(i * 100, i, i as f32 * 100.);
|
events.push(i * 100, i, i as f32 * 100.);
|
||||||
}
|
}
|
||||||
events_vec2.push(Ok(ChannelEvents::Events(Box::new(events.clone()))));
|
push_evd0(&mut vec, Box::new(events));
|
||||||
}
|
vec
|
||||||
let events_vec2 = events_vec2;
|
};
|
||||||
|
|
||||||
let inp2_events_a: Vec<Result<_, Error>> = vec![Ok(ChannelEvents::Status(ConnStatusEvent {
|
let inp2_events_a = {
|
||||||
ts: 1199,
|
let ev = ConnStatusEvent {
|
||||||
status: ConnStatus::Disconnect,
|
ts: 1199,
|
||||||
}))];
|
status: ConnStatus::Disconnect,
|
||||||
let inp2_events_b: Vec<Result<_, Error>> = vec![Ok(ChannelEvents::Status(ConnStatusEvent {
|
};
|
||||||
ts: 1199,
|
let item: Sitemty<ChannelEvents> = Ok(StreamItem::DataItem(RangeCompletableItem::Data(
|
||||||
status: ConnStatus::Disconnect,
|
ChannelEvents::Status(ev),
|
||||||
}))];
|
)));
|
||||||
|
vec![item]
|
||||||
|
};
|
||||||
|
|
||||||
|
let inp2_events_b = {
|
||||||
|
let ev = ConnStatusEvent {
|
||||||
|
ts: 1199,
|
||||||
|
status: ConnStatus::Disconnect,
|
||||||
|
};
|
||||||
|
let item: Sitemty<ChannelEvents> = Ok(StreamItem::DataItem(RangeCompletableItem::Data(
|
||||||
|
ChannelEvents::Status(ev),
|
||||||
|
)));
|
||||||
|
vec![item]
|
||||||
|
};
|
||||||
|
|
||||||
let inp1 = events_vec1;
|
let inp1 = events_vec1;
|
||||||
let inp1 = futures_util::stream::iter(inp1);
|
let inp1 = futures_util::stream::iter(inp1);
|
||||||
let inp1 = Box::pin(inp1);
|
let inp1 = Box::pin(inp1);
|
||||||
let inp2: Vec<Result<ChannelEvents, Error>> = inp2_events_a;
|
let inp2: Vec<Sitemty<ChannelEvents>> = inp2_events_a;
|
||||||
let inp2 = futures_util::stream::iter(inp2);
|
let inp2 = futures_util::stream::iter(inp2);
|
||||||
let inp2 = Box::pin(inp2);
|
let inp2 = Box::pin(inp2);
|
||||||
let mut merger = ChannelEventsMerger::new(vec![inp1, inp2]);
|
let mut merger = ChannelEventsMerger::new(vec![inp1, inp2]);
|
||||||
@@ -144,15 +163,17 @@ fn merge03() {
|
|||||||
#[test]
|
#[test]
|
||||||
fn bin01() {
|
fn bin01() {
|
||||||
let fut = async {
|
let fut = async {
|
||||||
let mut events_vec1 = Vec::new();
|
let inp1 = {
|
||||||
for j in 0..2 {
|
let mut vec = Vec::new();
|
||||||
let mut events = EventsDim0::empty();
|
for j in 0..2 {
|
||||||
for i in 10 * j..10 * (1 + j) {
|
let mut events = EventsDim0::empty();
|
||||||
events.push(SEC * i, i, 17f32);
|
for i in 10 * j..10 * (1 + j) {
|
||||||
|
events.push(SEC * i, i, 17f32);
|
||||||
|
}
|
||||||
|
push_evd0(&mut vec, Box::new(events));
|
||||||
}
|
}
|
||||||
events_vec1.push(Ok(ChannelEvents::Events(Box::new(events))));
|
vec
|
||||||
}
|
};
|
||||||
let inp1 = events_vec1;
|
|
||||||
let inp1 = futures_util::stream::iter(inp1);
|
let inp1 = futures_util::stream::iter(inp1);
|
||||||
let inp1 = Box::pin(inp1);
|
let inp1 = Box::pin(inp1);
|
||||||
let inp2 = Box::pin(futures_util::stream::empty()) as _;
|
let inp2 = Box::pin(futures_util::stream::empty()) as _;
|
||||||
@@ -165,36 +186,44 @@ fn bin01() {
|
|||||||
while let Some(item) = stream.next().await {
|
while let Some(item) = stream.next().await {
|
||||||
let item = item?;
|
let item = item?;
|
||||||
match item {
|
match item {
|
||||||
ChannelEvents::Events(events) => {
|
StreamItem::DataItem(item) => match item {
|
||||||
if binner.is_none() {
|
RangeCompletableItem::RangeComplete => todo!(),
|
||||||
let bb = events.as_time_binnable().time_binner_new(edges.clone(), do_time_weight);
|
RangeCompletableItem::Data(item) => match item {
|
||||||
binner = Some(bb);
|
ChannelEvents::Events(events) => {
|
||||||
}
|
if binner.is_none() {
|
||||||
let binner = binner.as_mut().unwrap();
|
let bb = events.as_time_binnable().time_binner_new(edges.clone(), do_time_weight);
|
||||||
binner.ingest(events.as_time_binnable());
|
binner = Some(bb);
|
||||||
eprintln!("bins_ready_count: {}", binner.bins_ready_count());
|
|
||||||
if binner.bins_ready_count() > 0 {
|
|
||||||
let ready = binner.bins_ready();
|
|
||||||
match ready {
|
|
||||||
Some(mut ready) => {
|
|
||||||
eprintln!("ready {ready:?}");
|
|
||||||
if coll.is_none() {
|
|
||||||
coll = Some(ready.as_collectable_mut().new_collector(bin_count_exp));
|
|
||||||
}
|
|
||||||
let cl = coll.as_mut().unwrap();
|
|
||||||
cl.ingest(ready.as_collectable_mut());
|
|
||||||
}
|
}
|
||||||
None => {
|
let binner = binner.as_mut().unwrap();
|
||||||
return Err(format!("bins_ready_count but no result").into());
|
binner.ingest(events.as_time_binnable());
|
||||||
|
eprintln!("bins_ready_count: {}", binner.bins_ready_count());
|
||||||
|
if binner.bins_ready_count() > 0 {
|
||||||
|
let ready = binner.bins_ready();
|
||||||
|
match ready {
|
||||||
|
Some(mut ready) => {
|
||||||
|
eprintln!("ready {ready:?}");
|
||||||
|
if coll.is_none() {
|
||||||
|
coll = Some(ready.as_collectable_mut().new_collector(bin_count_exp));
|
||||||
|
}
|
||||||
|
let cl = coll.as_mut().unwrap();
|
||||||
|
cl.ingest(ready.as_collectable_mut());
|
||||||
|
}
|
||||||
|
None => {
|
||||||
|
return Err(format!("bins_ready_count but no result").into());
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
ChannelEvents::Status(_) => {
|
||||||
|
eprintln!("TODO Status");
|
||||||
|
}
|
||||||
|
},
|
||||||
|
},
|
||||||
|
StreamItem::Log(_) => {
|
||||||
|
eprintln!("TODO Log");
|
||||||
}
|
}
|
||||||
ChannelEvents::Status(_) => {
|
StreamItem::Stats(_) => {
|
||||||
eprintln!("TODO Status");
|
eprintln!("TODO Stats");
|
||||||
}
|
|
||||||
ChannelEvents::RangeComplete => {
|
|
||||||
eprintln!("TODO RangeComplete");
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -249,26 +278,31 @@ fn bin02() {
|
|||||||
events.push(t, t, val(t));
|
events.push(t, t, val(t));
|
||||||
t += MS * 100;
|
t += MS * 100;
|
||||||
}
|
}
|
||||||
events_vec1.push(Ok(ChannelEvents::Events(Box::new(events))));
|
let cev = ChannelEvents::Events(Box::new(events));
|
||||||
|
events_vec1.push(Ok(StreamItem::DataItem(RangeCompletableItem::Data(cev))));
|
||||||
}
|
}
|
||||||
events_vec1.push(Ok(ChannelEvents::RangeComplete));
|
events_vec1.push(Ok(StreamItem::DataItem(RangeCompletableItem::RangeComplete)));
|
||||||
let inp1 = events_vec1;
|
let inp1 = events_vec1;
|
||||||
let inp1 = futures_util::stream::iter(inp1);
|
let inp1 = futures_util::stream::iter(inp1);
|
||||||
let inp1 = Box::pin(inp1);
|
let inp1 = Box::pin(inp1);
|
||||||
let inp2 = Box::pin(futures_util::stream::empty()) as _;
|
let inp2 = Box::pin(futures_util::stream::empty()) as _;
|
||||||
let stream = ChannelEventsMerger::new(vec![inp1, inp2]);
|
let stream = ChannelEventsMerger::new(vec![inp1, inp2]);
|
||||||
let range = NanoRange {
|
if false {
|
||||||
beg: TSBASE + SEC * 1,
|
// covering_range result is subject to adjustments, instead, manually choose bin edges
|
||||||
end: TSBASE + SEC * 10,
|
let range = NanoRange {
|
||||||
};
|
beg: TSBASE + SEC * 1,
|
||||||
let covering = BinnedRange::covering_range(range, 3).map_err(|e| format!("{e}"))?;
|
end: TSBASE + SEC * 10,
|
||||||
assert_eq!(covering.edges().len(), 10);
|
};
|
||||||
|
let covering = BinnedRange::covering_range(range, 3).map_err(|e| format!("{e}"))?;
|
||||||
|
assert_eq!(covering.edges().len(), 6);
|
||||||
|
}
|
||||||
|
let edges = (0..10).into_iter().map(|x| TSBASE + SEC * 1 + SEC * x).collect();
|
||||||
let stream = Box::pin(stream);
|
let stream = Box::pin(stream);
|
||||||
let collected = binned_collected(
|
let collected = binned_collected(
|
||||||
ScalarType::F32,
|
ScalarType::F32,
|
||||||
Shape::Scalar,
|
Shape::Scalar,
|
||||||
AggKind::TimeWeightedScalar,
|
AggKind::TimeWeightedScalar,
|
||||||
covering.edges(),
|
edges,
|
||||||
Duration::from_millis(2000),
|
Duration::from_millis(2000),
|
||||||
stream,
|
stream,
|
||||||
)
|
)
|
||||||
@@ -294,9 +328,10 @@ fn bin03() {
|
|||||||
events.push(t, t, val(t));
|
events.push(t, t, val(t));
|
||||||
t += MS * 100;
|
t += MS * 100;
|
||||||
}
|
}
|
||||||
events_vec1.push(Ok(ChannelEvents::Events(Box::new(events))));
|
let cev = ChannelEvents::Events(Box::new(events));
|
||||||
|
events_vec1.push(Ok(StreamItem::DataItem(RangeCompletableItem::Data(cev))));
|
||||||
}
|
}
|
||||||
events_vec1.push(Ok(ChannelEvents::RangeComplete));
|
events_vec1.push(Ok(StreamItem::DataItem(RangeCompletableItem::RangeComplete)));
|
||||||
let inp1 = events_vec1;
|
let inp1 = events_vec1;
|
||||||
let inp1 = futures_util::stream::iter(inp1).enumerate().then(|(i, k)| async move {
|
let inp1 = futures_util::stream::iter(inp1).enumerate().then(|(i, k)| async move {
|
||||||
if i == 4 {
|
if i == 4 {
|
||||||
|
|||||||
@@ -1640,7 +1640,7 @@ pub mod log {
|
|||||||
pub use tracing::{debug, error, event, info, span, trace, warn, Level};
|
pub use tracing::{debug, error, event, info, span, trace, warn, Level};
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Serialize, Deserialize)]
|
#[derive(Clone, Debug, PartialEq, Serialize, Deserialize)]
|
||||||
pub struct EventDataReadStats {
|
pub struct EventDataReadStats {
|
||||||
pub parsed_bytes: u64,
|
pub parsed_bytes: u64,
|
||||||
}
|
}
|
||||||
@@ -1655,7 +1655,7 @@ impl EventDataReadStats {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, Debug, Serialize, Deserialize)]
|
#[derive(Clone, Debug, PartialEq, Serialize, Deserialize)]
|
||||||
pub struct RangeFilterStats {
|
pub struct RangeFilterStats {
|
||||||
pub events_pre: u64,
|
pub events_pre: u64,
|
||||||
pub events_post: u64,
|
pub events_post: u64,
|
||||||
@@ -1672,7 +1672,7 @@ impl RangeFilterStats {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, Debug, Serialize, Deserialize)]
|
#[derive(Clone, Debug, PartialEq, Serialize, Deserialize)]
|
||||||
pub enum DiskStats {
|
pub enum DiskStats {
|
||||||
OpenStats(OpenStats),
|
OpenStats(OpenStats),
|
||||||
SeekStats(SeekStats),
|
SeekStats(SeekStats),
|
||||||
@@ -1680,7 +1680,7 @@ pub enum DiskStats {
|
|||||||
ReadExactStats(ReadExactStats),
|
ReadExactStats(ReadExactStats),
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, Debug, Serialize, Deserialize)]
|
#[derive(Clone, Debug, PartialEq, Serialize, Deserialize)]
|
||||||
pub struct OpenStats {
|
pub struct OpenStats {
|
||||||
pub duration: Duration,
|
pub duration: Duration,
|
||||||
}
|
}
|
||||||
@@ -1691,7 +1691,7 @@ impl OpenStats {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, Debug, Serialize, Deserialize)]
|
#[derive(Clone, Debug, PartialEq, Serialize, Deserialize)]
|
||||||
pub struct SeekStats {
|
pub struct SeekStats {
|
||||||
pub duration: Duration,
|
pub duration: Duration,
|
||||||
}
|
}
|
||||||
@@ -1702,7 +1702,7 @@ impl SeekStats {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, Debug, Serialize, Deserialize)]
|
#[derive(Clone, Debug, PartialEq, Serialize, Deserialize)]
|
||||||
pub struct ReadStats {
|
pub struct ReadStats {
|
||||||
pub duration: Duration,
|
pub duration: Duration,
|
||||||
}
|
}
|
||||||
@@ -1713,7 +1713,7 @@ impl ReadStats {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, Debug, Serialize, Deserialize)]
|
#[derive(Clone, Debug, PartialEq, Serialize, Deserialize)]
|
||||||
pub struct ReadExactStats {
|
pub struct ReadExactStats {
|
||||||
pub duration: Duration,
|
pub duration: Duration,
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -169,9 +169,6 @@ async fn events_conn_handler_inner_try(
|
|||||||
let item = items::scalarevents::ScalarEvents::<f64>::empty();
|
let item = items::scalarevents::ScalarEvents::<f64>::empty();
|
||||||
Ok(StreamItem::DataItem(RangeCompletableItem::Data(item)))
|
Ok(StreamItem::DataItem(RangeCompletableItem::Data(item)))
|
||||||
}
|
}
|
||||||
items_2::ChannelEvents::RangeComplete => {
|
|
||||||
Ok(StreamItem::DataItem(RangeCompletableItem::RangeComplete))
|
|
||||||
}
|
|
||||||
items_2::ChannelEvents::Status(_item) => todo!(),
|
items_2::ChannelEvents::Status(_item) => todo!(),
|
||||||
},
|
},
|
||||||
Err(e) => Err(e),
|
Err(e) => Err(e),
|
||||||
|
|||||||
@@ -383,6 +383,8 @@ pub async fn fetch_uncached_binned_events(
|
|||||||
)));
|
)));
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
// TODO as soon we encounter RangeComplete we just:
|
||||||
|
// complete = true;
|
||||||
match item {
|
match item {
|
||||||
Ok(ChannelEvents::Events(item)) => {
|
Ok(ChannelEvents::Events(item)) => {
|
||||||
time_binner.ingest(item.as_time_binnable());
|
time_binner.ingest(item.as_time_binnable());
|
||||||
@@ -394,9 +396,6 @@ pub async fn fetch_uncached_binned_events(
|
|||||||
"unexpected read of channel status events"
|
"unexpected read of channel status events"
|
||||||
)));
|
)));
|
||||||
}
|
}
|
||||||
Ok(ChannelEvents::RangeComplete) => {
|
|
||||||
complete = true;
|
|
||||||
}
|
|
||||||
Err(e) => return Err(e),
|
Err(e) => return Err(e),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user