Basic binner for new items
This commit is contained in:
@@ -5,7 +5,7 @@ use items::scalarevents::ScalarEvents;
|
|||||||
use items::waveevents::WaveEvents;
|
use items::waveevents::WaveEvents;
|
||||||
use items::{EventsDyn, RangeCompletableItem, Sitemty, StreamItem};
|
use items::{EventsDyn, RangeCompletableItem, Sitemty, StreamItem};
|
||||||
use netpod::log::*;
|
use netpod::log::*;
|
||||||
use netpod::query::{ChannelStateEvents, RawEventsQuery};
|
use netpod::query::{ChannelStateEventsQuery, RawEventsQuery};
|
||||||
use netpod::timeunits::DAY;
|
use netpod::timeunits::DAY;
|
||||||
use netpod::{Channel, Database, NanoRange, ScalarType, ScyllaConfig, Shape};
|
use netpod::{Channel, Database, NanoRange, ScalarType, ScyllaConfig, Shape};
|
||||||
use scylla::Session as ScySession;
|
use scylla::Session as ScySession;
|
||||||
@@ -576,7 +576,7 @@ pub async fn make_scylla_stream(
|
|||||||
}
|
}
|
||||||
|
|
||||||
pub async fn channel_state_events(
|
pub async fn channel_state_events(
|
||||||
evq: &ChannelStateEvents,
|
evq: &ChannelStateEventsQuery,
|
||||||
scyco: &ScyllaConfig,
|
scyco: &ScyllaConfig,
|
||||||
_dbconf: Database,
|
_dbconf: Database,
|
||||||
) -> Result<Vec<(u64, u32)>, Error> {
|
) -> Result<Vec<(u64, u32)>, Error> {
|
||||||
|
|||||||
@@ -3,7 +3,7 @@ use crate::err::Error;
|
|||||||
use dbconn::events_scylla::channel_state_events;
|
use dbconn::events_scylla::channel_state_events;
|
||||||
use http::{Method, Request, Response, StatusCode};
|
use http::{Method, Request, Response, StatusCode};
|
||||||
use hyper::Body;
|
use hyper::Body;
|
||||||
use netpod::query::ChannelStateEvents;
|
use netpod::query::ChannelStateEventsQuery;
|
||||||
use netpod::{FromUrl, NodeConfigCached, ACCEPT_ALL, APP_JSON};
|
use netpod::{FromUrl, NodeConfigCached, ACCEPT_ALL, APP_JSON};
|
||||||
use url::Url;
|
use url::Url;
|
||||||
|
|
||||||
@@ -27,7 +27,7 @@ impl ChannelStatusConnectionEvents {
|
|||||||
.map_or(accept_def, |k| k.to_str().unwrap_or(accept_def));
|
.map_or(accept_def, |k| k.to_str().unwrap_or(accept_def));
|
||||||
if accept == APP_JSON || accept == ACCEPT_ALL {
|
if accept == APP_JSON || accept == ACCEPT_ALL {
|
||||||
let url = Url::parse(&format!("dummy:{}", req.uri()))?;
|
let url = Url::parse(&format!("dummy:{}", req.uri()))?;
|
||||||
let q = ChannelStateEvents::from_url(&url)?;
|
let q = ChannelStateEventsQuery::from_url(&url)?;
|
||||||
match self.fetch_data(&q, node_config).await {
|
match self.fetch_data(&q, node_config).await {
|
||||||
Ok(k) => {
|
Ok(k) => {
|
||||||
let body = Body::from(serde_json::to_vec(&k)?);
|
let body = Body::from(serde_json::to_vec(&k)?);
|
||||||
@@ -46,7 +46,7 @@ impl ChannelStatusConnectionEvents {
|
|||||||
|
|
||||||
async fn fetch_data(
|
async fn fetch_data(
|
||||||
&self,
|
&self,
|
||||||
q: &ChannelStateEvents,
|
q: &ChannelStateEventsQuery,
|
||||||
node_config: &NodeConfigCached,
|
node_config: &NodeConfigCached,
|
||||||
) -> Result<Vec<(u64, u32)>, Error> {
|
) -> Result<Vec<(u64, u32)>, Error> {
|
||||||
let dbconf = &node_config.node_config.cluster.database;
|
let dbconf = &node_config.node_config.cluster.database;
|
||||||
|
|||||||
@@ -1,19 +1,19 @@
|
|||||||
use std::sync::Arc;
|
|
||||||
|
|
||||||
use crate::channelconfig::{chconf_from_events_binary, chconf_from_events_json};
|
use crate::channelconfig::{chconf_from_events_binary, chconf_from_events_json};
|
||||||
use crate::err::Error;
|
use crate::err::Error;
|
||||||
use crate::{response, response_err, BodyStream, ToPublicResponse};
|
use crate::{response, response_err, BodyStream, ToPublicResponse};
|
||||||
use futures_util::{StreamExt, TryStreamExt};
|
use futures_util::{Stream, StreamExt, TryStreamExt};
|
||||||
use http::{Method, Request, Response, StatusCode};
|
use http::{Method, Request, Response, StatusCode};
|
||||||
use hyper::Body;
|
use hyper::Body;
|
||||||
use items_2::ChannelEvents;
|
use items_2::{binned_collected, ChannelEvents, ChannelEventsMerger};
|
||||||
use netpod::log::*;
|
use netpod::log::*;
|
||||||
use netpod::query::PlainEventsQuery;
|
use netpod::query::{BinnedQuery, ChannelStateEventsQuery, PlainEventsQuery};
|
||||||
use netpod::{AggKind, FromUrl, NodeConfigCached};
|
use netpod::{AggKind, FromUrl, NodeConfigCached};
|
||||||
use netpod::{ACCEPT_ALL, APP_JSON, APP_OCTET};
|
use netpod::{ACCEPT_ALL, APP_JSON, APP_OCTET};
|
||||||
use scyllaconn::create_scy_session;
|
use scyllaconn::create_scy_session;
|
||||||
use scyllaconn::errconv::ErrConv;
|
use scyllaconn::errconv::ErrConv;
|
||||||
use scyllaconn::events::make_scylla_stream;
|
use scyllaconn::events::{channel_state_events, make_scylla_stream};
|
||||||
|
use std::pin::Pin;
|
||||||
|
use std::sync::Arc;
|
||||||
use url::Url;
|
use url::Url;
|
||||||
|
|
||||||
pub struct EventsHandler {}
|
pub struct EventsHandler {}
|
||||||
@@ -226,3 +226,90 @@ impl EventsHandlerScylla {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub struct BinnedHandlerScylla {}
|
||||||
|
|
||||||
|
impl BinnedHandlerScylla {
|
||||||
|
pub fn handler(req: &Request<Body>) -> Option<Self> {
|
||||||
|
if req.uri().path() == "/api/4/scylla/binned" {
|
||||||
|
Some(Self {})
|
||||||
|
} else {
|
||||||
|
None
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn handle(&self, req: Request<Body>, node_config: &NodeConfigCached) -> Result<Response<Body>, Error> {
|
||||||
|
if req.method() != Method::GET {
|
||||||
|
return Ok(response(StatusCode::NOT_ACCEPTABLE).body(Body::empty())?);
|
||||||
|
}
|
||||||
|
match self.fetch(req, node_config).await {
|
||||||
|
Ok(ret) => Ok(ret),
|
||||||
|
Err(e) => Ok(e.to_public_response()),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn fetch(&self, req: Request<Body>, node_config: &NodeConfigCached) -> Result<Response<Body>, Error> {
|
||||||
|
info!("BinnedHandlerScylla req: {:?}", req);
|
||||||
|
let accept_def = APP_JSON;
|
||||||
|
let accept = req
|
||||||
|
.headers()
|
||||||
|
.get(http::header::ACCEPT)
|
||||||
|
.map_or(accept_def, |k| k.to_str().unwrap_or(accept_def));
|
||||||
|
if accept.contains(APP_JSON) || accept.contains(ACCEPT_ALL) {
|
||||||
|
Ok(self.gather(req, node_config).await?)
|
||||||
|
} else {
|
||||||
|
let ret = response_err(StatusCode::NOT_ACCEPTABLE, format!("Unsupported Accept: {:?}", accept))?;
|
||||||
|
Ok(ret)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn gather(&self, req: Request<Body>, node_config: &NodeConfigCached) -> Result<Response<Body>, Error> {
|
||||||
|
let (head, _body) = req.into_parts();
|
||||||
|
warn!("TODO BinnedQuery needs to take AggKind");
|
||||||
|
let s1 = format!("dummy:{}", head.uri);
|
||||||
|
let url = Url::parse(&s1)?;
|
||||||
|
let evq = BinnedQuery::from_url(&url)?;
|
||||||
|
let pgclient = {
|
||||||
|
// TODO use common connection/pool:
|
||||||
|
info!("--------------- open postgres connection");
|
||||||
|
let pgconf = &node_config.node_config.cluster.database;
|
||||||
|
let u = {
|
||||||
|
let d = &pgconf;
|
||||||
|
format!("postgresql://{}:{}@{}:{}/{}", d.user, d.pass, d.host, d.port, d.name)
|
||||||
|
};
|
||||||
|
let (pgclient, pgconn) = tokio_postgres::connect(&u, tokio_postgres::NoTls).await.err_conv()?;
|
||||||
|
tokio::spawn(pgconn);
|
||||||
|
let pgclient = Arc::new(pgclient);
|
||||||
|
pgclient
|
||||||
|
};
|
||||||
|
if let Some(scyco) = &node_config.node_config.cluster.scylla {
|
||||||
|
let scy = create_scy_session(scyco).await?;
|
||||||
|
let mut query2 = PlainEventsQuery::new(evq.channel().clone(), evq.range().clone(), 0, None, false);
|
||||||
|
query2.set_timeout(evq.timeout());
|
||||||
|
let query2 = query2;
|
||||||
|
let stream = make_scylla_stream(&query2, scy.clone(), &pgclient, false).await?;
|
||||||
|
let query3 = ChannelStateEventsQuery::new(evq.channel().clone(), evq.range().clone());
|
||||||
|
let state_stream = channel_state_events(&query3, scy.clone()).await?;
|
||||||
|
// TODO let the stream itself use the items_2 error, do not convert here.
|
||||||
|
let data_stream = Box::pin(stream.map_err(|e| items_2::Error::from(format!("{e}"))));
|
||||||
|
let state_stream = Box::pin(state_stream.map_err(|e| items_2::Error::from(format!("{e}"))));
|
||||||
|
let merged_stream = ChannelEventsMerger::new(data_stream, state_stream);
|
||||||
|
let merged_stream = Box::pin(merged_stream) as Pin<Box<dyn Stream<Item = _> + Send>>;
|
||||||
|
let binned_collected = binned_collected(merged_stream)
|
||||||
|
.await
|
||||||
|
.map_err(|e| Error::with_msg_no_trace(format!("{e}")))?;
|
||||||
|
let res = binned_collected.to_json_result()?;
|
||||||
|
let res = res.to_json_bytes()?;
|
||||||
|
let ret = response(StatusCode::OK).body(Body::from(res))?;
|
||||||
|
{
|
||||||
|
let _ret = response(StatusCode::OK).body(BodyStream::wrapped(
|
||||||
|
futures_util::stream::iter([Ok(Vec::new())]),
|
||||||
|
format!("gather"),
|
||||||
|
))?;
|
||||||
|
}
|
||||||
|
Ok(ret)
|
||||||
|
} else {
|
||||||
|
return Err(Error::with_public_msg(format!("no scylla configured")));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|||||||
@@ -247,6 +247,8 @@ async fn http_service_try(req: Request<Body>, node_config: &NodeConfigCached) ->
|
|||||||
h.handle(req, &node_config).await
|
h.handle(req, &node_config).await
|
||||||
} else if let Some(h) = events::EventsHandlerScylla::handler(&req) {
|
} else if let Some(h) = events::EventsHandlerScylla::handler(&req) {
|
||||||
h.handle(req, &node_config).await
|
h.handle(req, &node_config).await
|
||||||
|
} else if let Some(h) = events::BinnedHandlerScylla::handler(&req) {
|
||||||
|
h.handle(req, &node_config).await
|
||||||
} else if let Some(h) = ChannelStatusConnectionEvents::handler(&req) {
|
} else if let Some(h) = ChannelStatusConnectionEvents::handler(&req) {
|
||||||
h.handle(req, &node_config).await
|
h.handle(req, &node_config).await
|
||||||
} else if path == "/api/4/binned" {
|
} else if path == "/api/4/binned" {
|
||||||
|
|||||||
@@ -1,5 +1,8 @@
|
|||||||
use crate::streams::{CollectableType, CollectorType, ToJsonResult};
|
use crate::streams::{Collectable, CollectableType, CollectorType, ToJsonResult};
|
||||||
use crate::{ts_offs_from_abs, AppendEmptyBin, Empty, IsoDateTime, RangeOverlapInfo, ScalarOps, TimeBins, WithLen};
|
use crate::{
|
||||||
|
ts_offs_from_abs, ts_offs_from_abs_with_anchor, AppendEmptyBin, Empty, IsoDateTime, RangeOverlapInfo, ScalarOps,
|
||||||
|
TimeBins, WithLen,
|
||||||
|
};
|
||||||
use crate::{TimeBinnable, TimeBinnableType, TimeBinnableTypeAggregator, TimeBinned, TimeBinner};
|
use crate::{TimeBinnable, TimeBinnableType, TimeBinnableTypeAggregator, TimeBinned, TimeBinner};
|
||||||
use chrono::{TimeZone, Utc};
|
use chrono::{TimeZone, Utc};
|
||||||
use err::Error;
|
use err::Error;
|
||||||
@@ -10,11 +13,10 @@ use num_traits::Zero;
|
|||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
use std::any::Any;
|
use std::any::Any;
|
||||||
use std::collections::VecDeque;
|
use std::collections::VecDeque;
|
||||||
use std::marker::PhantomData;
|
|
||||||
use std::{fmt, mem};
|
use std::{fmt, mem};
|
||||||
|
|
||||||
#[derive(Clone, Serialize, Deserialize)]
|
#[derive(Clone, PartialEq, Serialize, Deserialize)]
|
||||||
pub struct MinMaxAvgDim0Bins<NTY> {
|
pub struct BinsDim0<NTY> {
|
||||||
pub ts1s: VecDeque<u64>,
|
pub ts1s: VecDeque<u64>,
|
||||||
pub ts2s: VecDeque<u64>,
|
pub ts2s: VecDeque<u64>,
|
||||||
pub counts: VecDeque<u64>,
|
pub counts: VecDeque<u64>,
|
||||||
@@ -23,7 +25,7 @@ pub struct MinMaxAvgDim0Bins<NTY> {
|
|||||||
pub avgs: VecDeque<f32>,
|
pub avgs: VecDeque<f32>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<NTY> fmt::Debug for MinMaxAvgDim0Bins<NTY>
|
impl<NTY> fmt::Debug for BinsDim0<NTY>
|
||||||
where
|
where
|
||||||
NTY: fmt::Debug,
|
NTY: fmt::Debug,
|
||||||
{
|
{
|
||||||
@@ -43,7 +45,7 @@ where
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<NTY> MinMaxAvgDim0Bins<NTY> {
|
impl<NTY> BinsDim0<NTY> {
|
||||||
pub fn empty() -> Self {
|
pub fn empty() -> Self {
|
||||||
Self {
|
Self {
|
||||||
ts1s: VecDeque::new(),
|
ts1s: VecDeque::new(),
|
||||||
@@ -56,13 +58,13 @@ impl<NTY> MinMaxAvgDim0Bins<NTY> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<NTY> WithLen for MinMaxAvgDim0Bins<NTY> {
|
impl<NTY> WithLen for BinsDim0<NTY> {
|
||||||
fn len(&self) -> usize {
|
fn len(&self) -> usize {
|
||||||
self.ts1s.len()
|
self.ts1s.len()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<NTY> RangeOverlapInfo for MinMaxAvgDim0Bins<NTY> {
|
impl<NTY> RangeOverlapInfo for BinsDim0<NTY> {
|
||||||
fn ends_before(&self, range: NanoRange) -> bool {
|
fn ends_before(&self, range: NanoRange) -> bool {
|
||||||
if let Some(&max) = self.ts2s.back() {
|
if let Some(&max) = self.ts2s.back() {
|
||||||
max <= range.beg
|
max <= range.beg
|
||||||
@@ -88,7 +90,7 @@ impl<NTY> RangeOverlapInfo for MinMaxAvgDim0Bins<NTY> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<NTY> Empty for MinMaxAvgDim0Bins<NTY> {
|
impl<NTY> Empty for BinsDim0<NTY> {
|
||||||
fn empty() -> Self {
|
fn empty() -> Self {
|
||||||
Self {
|
Self {
|
||||||
ts1s: Default::default(),
|
ts1s: Default::default(),
|
||||||
@@ -101,7 +103,7 @@ impl<NTY> Empty for MinMaxAvgDim0Bins<NTY> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<NTY: ScalarOps> AppendEmptyBin for MinMaxAvgDim0Bins<NTY> {
|
impl<NTY: ScalarOps> AppendEmptyBin for BinsDim0<NTY> {
|
||||||
fn append_empty_bin(&mut self, ts1: u64, ts2: u64) {
|
fn append_empty_bin(&mut self, ts1: u64, ts2: u64) {
|
||||||
self.ts1s.push_back(ts1);
|
self.ts1s.push_back(ts1);
|
||||||
self.ts2s.push_back(ts2);
|
self.ts2s.push_back(ts2);
|
||||||
@@ -112,7 +114,7 @@ impl<NTY: ScalarOps> AppendEmptyBin for MinMaxAvgDim0Bins<NTY> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<NTY: ScalarOps> TimeBins for MinMaxAvgDim0Bins<NTY> {
|
impl<NTY: ScalarOps> TimeBins for BinsDim0<NTY> {
|
||||||
fn ts_min(&self) -> Option<u64> {
|
fn ts_min(&self) -> Option<u64> {
|
||||||
self.ts1s.front().map(Clone::clone)
|
self.ts1s.front().map(Clone::clone)
|
||||||
}
|
}
|
||||||
@@ -130,9 +132,9 @@ impl<NTY: ScalarOps> TimeBins for MinMaxAvgDim0Bins<NTY> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<NTY: ScalarOps> TimeBinnableType for MinMaxAvgDim0Bins<NTY> {
|
impl<NTY: ScalarOps> TimeBinnableType for BinsDim0<NTY> {
|
||||||
type Output = MinMaxAvgDim0Bins<NTY>;
|
type Output = BinsDim0<NTY>;
|
||||||
type Aggregator = MinMaxAvgDim0BinsAggregator<NTY>;
|
type Aggregator = BinsDim0Aggregator<NTY>;
|
||||||
|
|
||||||
fn aggregator(range: NanoRange, x_bin_count: usize, do_time_weight: bool) -> Self::Aggregator {
|
fn aggregator(range: NanoRange, x_bin_count: usize, do_time_weight: bool) -> Self::Aggregator {
|
||||||
let self_name = std::any::type_name::<Self>();
|
let self_name = std::any::type_name::<Self>();
|
||||||
@@ -144,24 +146,18 @@ impl<NTY: ScalarOps> TimeBinnableType for MinMaxAvgDim0Bins<NTY> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub struct MinMaxAvgBinsCollected<NTY> {
|
#[derive(Debug, Serialize)]
|
||||||
_m1: PhantomData<NTY>,
|
pub struct BinsDim0CollectedResult<NTY> {
|
||||||
}
|
|
||||||
|
|
||||||
impl<NTY> MinMaxAvgBinsCollected<NTY> {
|
|
||||||
pub fn new() -> Self {
|
|
||||||
Self { _m1: PhantomData }
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Serialize)]
|
|
||||||
pub struct MinMaxAvgBinsCollectedResult<NTY> {
|
|
||||||
#[serde(rename = "tsAnchor")]
|
#[serde(rename = "tsAnchor")]
|
||||||
ts_anchor_sec: u64,
|
ts_anchor_sec: u64,
|
||||||
#[serde(rename = "tsMs")]
|
#[serde(rename = "ts1Ms")]
|
||||||
ts_off_ms: VecDeque<u64>,
|
ts1_off_ms: VecDeque<u64>,
|
||||||
#[serde(rename = "tsNs")]
|
#[serde(rename = "ts2Ms")]
|
||||||
ts_off_ns: VecDeque<u64>,
|
ts2_off_ms: VecDeque<u64>,
|
||||||
|
#[serde(rename = "ts1Ns")]
|
||||||
|
ts1_off_ns: VecDeque<u64>,
|
||||||
|
#[serde(rename = "ts2Ns")]
|
||||||
|
ts2_off_ns: VecDeque<u64>,
|
||||||
counts: VecDeque<u64>,
|
counts: VecDeque<u64>,
|
||||||
mins: VecDeque<NTY>,
|
mins: VecDeque<NTY>,
|
||||||
maxs: VecDeque<NTY>,
|
maxs: VecDeque<NTY>,
|
||||||
@@ -174,41 +170,47 @@ pub struct MinMaxAvgBinsCollectedResult<NTY> {
|
|||||||
continue_at: Option<IsoDateTime>,
|
continue_at: Option<IsoDateTime>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<NTY: ScalarOps> ToJsonResult for MinMaxAvgBinsCollectedResult<NTY> {
|
impl<NTY: ScalarOps> ToJsonResult for BinsDim0CollectedResult<NTY> {
|
||||||
fn to_json_result(&self) -> Result<Box<dyn crate::streams::ToJsonBytes>, Error> {
|
fn to_json_result(&self) -> Result<Box<dyn crate::streams::ToJsonBytes>, Error> {
|
||||||
let k = serde_json::to_value(self)?;
|
let k = serde_json::to_value(self)?;
|
||||||
Ok(Box::new(k))
|
Ok(Box::new(k))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub struct MinMaxAvgBinsCollector<NTY> {
|
pub struct BinsDim0Collector<NTY> {
|
||||||
timed_out: bool,
|
timed_out: bool,
|
||||||
range_complete: bool,
|
range_complete: bool,
|
||||||
vals: MinMaxAvgDim0Bins<NTY>,
|
vals: BinsDim0<NTY>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<NTY> MinMaxAvgBinsCollector<NTY> {
|
impl<NTY> BinsDim0Collector<NTY> {
|
||||||
pub fn new() -> Self {
|
pub fn new() -> Self {
|
||||||
Self {
|
Self {
|
||||||
timed_out: false,
|
timed_out: false,
|
||||||
range_complete: false,
|
range_complete: false,
|
||||||
vals: MinMaxAvgDim0Bins::<NTY>::empty(),
|
vals: BinsDim0::<NTY>::empty(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<NTY> WithLen for MinMaxAvgBinsCollector<NTY> {
|
impl<NTY> WithLen for BinsDim0Collector<NTY> {
|
||||||
fn len(&self) -> usize {
|
fn len(&self) -> usize {
|
||||||
self.vals.ts1s.len()
|
self.vals.ts1s.len()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<NTY: ScalarOps> CollectorType for MinMaxAvgBinsCollector<NTY> {
|
impl<NTY: ScalarOps> CollectorType for BinsDim0Collector<NTY> {
|
||||||
type Input = MinMaxAvgDim0Bins<NTY>;
|
type Input = BinsDim0<NTY>;
|
||||||
type Output = MinMaxAvgBinsCollectedResult<NTY>;
|
type Output = BinsDim0CollectedResult<NTY>;
|
||||||
|
|
||||||
fn ingest(&mut self, _src: &mut Self::Input) {
|
fn ingest(&mut self, src: &mut Self::Input) {
|
||||||
err::todo();
|
// TODO could be optimized by non-contiguous container.
|
||||||
|
self.vals.ts1s.append(&mut src.ts1s);
|
||||||
|
self.vals.ts2s.append(&mut src.ts2s);
|
||||||
|
self.vals.counts.append(&mut src.counts);
|
||||||
|
self.vals.mins.append(&mut src.mins);
|
||||||
|
self.vals.maxs.append(&mut src.maxs);
|
||||||
|
self.vals.avgs.append(&mut src.avgs);
|
||||||
}
|
}
|
||||||
|
|
||||||
fn set_range_complete(&mut self) {
|
fn set_range_complete(&mut self) {
|
||||||
@@ -221,7 +223,7 @@ impl<NTY: ScalarOps> CollectorType for MinMaxAvgBinsCollector<NTY> {
|
|||||||
|
|
||||||
fn result(&mut self) -> Result<Self::Output, Error> {
|
fn result(&mut self) -> Result<Self::Output, Error> {
|
||||||
let bin_count = self.vals.ts1s.len() as u32;
|
let bin_count = self.vals.ts1s.len() as u32;
|
||||||
// TODO could save the copy:
|
// TODO save the clone:
|
||||||
let mut ts_all = self.vals.ts1s.clone();
|
let mut ts_all = self.vals.ts1s.clone();
|
||||||
if self.vals.ts2s.len() > 0 {
|
if self.vals.ts2s.len() > 0 {
|
||||||
ts_all.push_back(*self.vals.ts2s.back().unwrap());
|
ts_all.push_back(*self.vals.ts2s.back().unwrap());
|
||||||
@@ -242,15 +244,18 @@ impl<NTY: ScalarOps> CollectorType for MinMaxAvgBinsCollector<NTY> {
|
|||||||
if ts_all.as_slices().1.len() != 0 {
|
if ts_all.as_slices().1.len() != 0 {
|
||||||
panic!();
|
panic!();
|
||||||
}
|
}
|
||||||
let tst = ts_offs_from_abs(ts_all.as_slices().0);
|
let tst1 = ts_offs_from_abs(self.vals.ts1s.as_slices().0);
|
||||||
|
let tst2 = ts_offs_from_abs_with_anchor(tst1.0, self.vals.ts2s.as_slices().0);
|
||||||
let counts = mem::replace(&mut self.vals.counts, VecDeque::new());
|
let counts = mem::replace(&mut self.vals.counts, VecDeque::new());
|
||||||
let mins = mem::replace(&mut self.vals.mins, VecDeque::new());
|
let mins = mem::replace(&mut self.vals.mins, VecDeque::new());
|
||||||
let maxs = mem::replace(&mut self.vals.maxs, VecDeque::new());
|
let maxs = mem::replace(&mut self.vals.maxs, VecDeque::new());
|
||||||
let avgs = mem::replace(&mut self.vals.avgs, VecDeque::new());
|
let avgs = mem::replace(&mut self.vals.avgs, VecDeque::new());
|
||||||
let ret = MinMaxAvgBinsCollectedResult::<NTY> {
|
let ret = BinsDim0CollectedResult::<NTY> {
|
||||||
ts_anchor_sec: tst.0,
|
ts_anchor_sec: tst1.0,
|
||||||
ts_off_ms: tst.1,
|
ts1_off_ms: tst1.1,
|
||||||
ts_off_ns: tst.2,
|
ts1_off_ns: tst1.2,
|
||||||
|
ts2_off_ms: tst2.0,
|
||||||
|
ts2_off_ns: tst2.1,
|
||||||
counts,
|
counts,
|
||||||
mins,
|
mins,
|
||||||
maxs,
|
maxs,
|
||||||
@@ -263,15 +268,15 @@ impl<NTY: ScalarOps> CollectorType for MinMaxAvgBinsCollector<NTY> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<NTY: ScalarOps> CollectableType for MinMaxAvgDim0Bins<NTY> {
|
impl<NTY: ScalarOps> CollectableType for BinsDim0<NTY> {
|
||||||
type Collector = MinMaxAvgBinsCollector<NTY>;
|
type Collector = BinsDim0Collector<NTY>;
|
||||||
|
|
||||||
fn new_collector() -> Self::Collector {
|
fn new_collector() -> Self::Collector {
|
||||||
Self::Collector::new()
|
Self::Collector::new()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub struct MinMaxAvgDim0BinsAggregator<NTY> {
|
pub struct BinsDim0Aggregator<NTY> {
|
||||||
range: NanoRange,
|
range: NanoRange,
|
||||||
count: u64,
|
count: u64,
|
||||||
min: NTY,
|
min: NTY,
|
||||||
@@ -282,7 +287,7 @@ pub struct MinMaxAvgDim0BinsAggregator<NTY> {
|
|||||||
sum: f32,
|
sum: f32,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<NTY: ScalarOps> MinMaxAvgDim0BinsAggregator<NTY> {
|
impl<NTY: ScalarOps> BinsDim0Aggregator<NTY> {
|
||||||
pub fn new(range: NanoRange, _do_time_weight: bool) -> Self {
|
pub fn new(range: NanoRange, _do_time_weight: bool) -> Self {
|
||||||
Self {
|
Self {
|
||||||
range,
|
range,
|
||||||
@@ -296,9 +301,9 @@ impl<NTY: ScalarOps> MinMaxAvgDim0BinsAggregator<NTY> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<NTY: ScalarOps> TimeBinnableTypeAggregator for MinMaxAvgDim0BinsAggregator<NTY> {
|
impl<NTY: ScalarOps> TimeBinnableTypeAggregator for BinsDim0Aggregator<NTY> {
|
||||||
type Input = MinMaxAvgDim0Bins<NTY>;
|
type Input = BinsDim0<NTY>;
|
||||||
type Output = MinMaxAvgDim0Bins<NTY>;
|
type Output = BinsDim0<NTY>;
|
||||||
|
|
||||||
fn range(&self) -> &NanoRange {
|
fn range(&self) -> &NanoRange {
|
||||||
&self.range
|
&self.range
|
||||||
@@ -348,9 +353,9 @@ impl<NTY: ScalarOps> TimeBinnableTypeAggregator for MinMaxAvgDim0BinsAggregator<
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<NTY: ScalarOps> TimeBinnable for MinMaxAvgDim0Bins<NTY> {
|
impl<NTY: ScalarOps> TimeBinnable for BinsDim0<NTY> {
|
||||||
fn time_binner_new(&self, edges: Vec<u64>, do_time_weight: bool) -> Box<dyn TimeBinner> {
|
fn time_binner_new(&self, edges: Vec<u64>, do_time_weight: bool) -> Box<dyn TimeBinner> {
|
||||||
let ret = MinMaxAvgDim0BinsTimeBinner::<NTY>::new(edges.into(), do_time_weight);
|
let ret = BinsDim0TimeBinner::<NTY>::new(edges.into(), do_time_weight);
|
||||||
Box::new(ret)
|
Box::new(ret)
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -359,14 +364,14 @@ impl<NTY: ScalarOps> TimeBinnable for MinMaxAvgDim0Bins<NTY> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub struct MinMaxAvgDim0BinsTimeBinner<NTY: ScalarOps> {
|
pub struct BinsDim0TimeBinner<NTY: ScalarOps> {
|
||||||
edges: VecDeque<u64>,
|
edges: VecDeque<u64>,
|
||||||
do_time_weight: bool,
|
do_time_weight: bool,
|
||||||
agg: Option<MinMaxAvgDim0BinsAggregator<NTY>>,
|
agg: Option<BinsDim0Aggregator<NTY>>,
|
||||||
ready: Option<<MinMaxAvgDim0BinsAggregator<NTY> as TimeBinnableTypeAggregator>::Output>,
|
ready: Option<<BinsDim0Aggregator<NTY> as TimeBinnableTypeAggregator>::Output>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<NTY: ScalarOps> MinMaxAvgDim0BinsTimeBinner<NTY> {
|
impl<NTY: ScalarOps> BinsDim0TimeBinner<NTY> {
|
||||||
fn new(edges: VecDeque<u64>, do_time_weight: bool) -> Self {
|
fn new(edges: VecDeque<u64>, do_time_weight: bool) -> Self {
|
||||||
Self {
|
Self {
|
||||||
edges,
|
edges,
|
||||||
@@ -390,7 +395,7 @@ impl<NTY: ScalarOps> MinMaxAvgDim0BinsTimeBinner<NTY> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<NTY: ScalarOps> TimeBinner for MinMaxAvgDim0BinsTimeBinner<NTY> {
|
impl<NTY: ScalarOps> TimeBinner for BinsDim0TimeBinner<NTY> {
|
||||||
fn ingest(&mut self, item: &dyn TimeBinnable) {
|
fn ingest(&mut self, item: &dyn TimeBinnable) {
|
||||||
let self_name = std::any::type_name::<Self>();
|
let self_name = std::any::type_name::<Self>();
|
||||||
if item.len() == 0 {
|
if item.len() == 0 {
|
||||||
@@ -427,7 +432,7 @@ impl<NTY: ScalarOps> TimeBinner for MinMaxAvgDim0BinsTimeBinner<NTY> {
|
|||||||
let agg = if let Some(agg) = self.agg.as_mut() {
|
let agg = if let Some(agg) = self.agg.as_mut() {
|
||||||
agg
|
agg
|
||||||
} else {
|
} else {
|
||||||
self.agg = Some(MinMaxAvgDim0BinsAggregator::new(
|
self.agg = Some(BinsDim0Aggregator::new(
|
||||||
// We know here that we have enough edges for another bin.
|
// We know here that we have enough edges for another bin.
|
||||||
// and `next_bin_range` will pop the first edge.
|
// and `next_bin_range` will pop the first edge.
|
||||||
self.next_bin_range().unwrap(),
|
self.next_bin_range().unwrap(),
|
||||||
@@ -438,7 +443,7 @@ impl<NTY: ScalarOps> TimeBinner for MinMaxAvgDim0BinsTimeBinner<NTY> {
|
|||||||
if let Some(item) = item
|
if let Some(item) = item
|
||||||
.as_any()
|
.as_any()
|
||||||
// TODO make statically sure that we attempt to cast to the correct type here:
|
// TODO make statically sure that we attempt to cast to the correct type here:
|
||||||
.downcast_ref::<<MinMaxAvgDim0BinsAggregator<NTY> as TimeBinnableTypeAggregator>::Input>()
|
.downcast_ref::<<BinsDim0Aggregator<NTY> as TimeBinnableTypeAggregator>::Input>()
|
||||||
{
|
{
|
||||||
agg.ingest(item);
|
agg.ingest(item);
|
||||||
} else {
|
} else {
|
||||||
@@ -502,7 +507,7 @@ impl<NTY: ScalarOps> TimeBinner for MinMaxAvgDim0BinsTimeBinner<NTY> {
|
|||||||
self.push_in_progress(true);
|
self.push_in_progress(true);
|
||||||
if self.bins_ready_count() == n {
|
if self.bins_ready_count() == n {
|
||||||
if let Some(_range) = self.next_bin_range() {
|
if let Some(_range) = self.next_bin_range() {
|
||||||
let bins = MinMaxAvgDim0Bins::<NTY>::empty();
|
let bins = BinsDim0::<NTY>::empty();
|
||||||
err::todo();
|
err::todo();
|
||||||
//bins.append_zero(range.beg, range.end);
|
//bins.append_zero(range.beg, range.end);
|
||||||
match self.ready.as_mut() {
|
match self.ready.as_mut() {
|
||||||
@@ -524,7 +529,7 @@ impl<NTY: ScalarOps> TimeBinner for MinMaxAvgDim0BinsTimeBinner<NTY> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<NTY: ScalarOps> TimeBinned for MinMaxAvgDim0Bins<NTY> {
|
impl<NTY: ScalarOps> TimeBinned for BinsDim0<NTY> {
|
||||||
fn as_time_binnable_dyn(&self) -> &dyn TimeBinnable {
|
fn as_time_binnable_dyn(&self) -> &dyn TimeBinnable {
|
||||||
self as &dyn TimeBinnable
|
self as &dyn TimeBinnable
|
||||||
}
|
}
|
||||||
@@ -576,4 +581,8 @@ impl<NTY: ScalarOps> TimeBinned for MinMaxAvgDim0Bins<NTY> {
|
|||||||
Err(msg)
|
Err(msg)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn as_collectable_mut(&mut self) -> &mut dyn Collectable {
|
||||||
|
self
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
use crate::binsdim0::MinMaxAvgDim0Bins;
|
use crate::binsdim0::BinsDim0;
|
||||||
use crate::streams::{CollectableType, CollectorType, ToJsonResult};
|
use crate::streams::{CollectableType, CollectorType, ToJsonResult};
|
||||||
use crate::{pulse_offs_from_abs, ts_offs_from_abs, RangeOverlapInfo};
|
use crate::{pulse_offs_from_abs, ts_offs_from_abs, RangeOverlapInfo};
|
||||||
use crate::{Empty, Events, ScalarOps, WithLen};
|
use crate::{Empty, Events, ScalarOps, WithLen};
|
||||||
@@ -87,7 +87,7 @@ impl<NTY: ScalarOps> RangeOverlapInfo for EventsDim0<NTY> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
impl<NTY: ScalarOps> TimeBinnableType for EventsDim0<NTY> {
|
impl<NTY: ScalarOps> TimeBinnableType for EventsDim0<NTY> {
|
||||||
type Output = MinMaxAvgDim0Bins<NTY>;
|
type Output = BinsDim0<NTY>;
|
||||||
type Aggregator = EventValuesAggregator<NTY>;
|
type Aggregator = EventValuesAggregator<NTY>;
|
||||||
|
|
||||||
fn aggregator(range: NanoRange, x_bin_count: usize, do_time_weight: bool) -> Self::Aggregator {
|
fn aggregator(range: NanoRange, x_bin_count: usize, do_time_weight: bool) -> Self::Aggregator {
|
||||||
@@ -122,7 +122,7 @@ impl<NTY> WithLen for EventValuesCollector<NTY> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Serialize)]
|
#[derive(Debug, Serialize)]
|
||||||
pub struct EventValuesCollectorOutput<NTY> {
|
pub struct EventValuesCollectorOutput<NTY> {
|
||||||
#[serde(rename = "tsAnchor")]
|
#[serde(rename = "tsAnchor")]
|
||||||
ts_anchor_sec: u64,
|
ts_anchor_sec: u64,
|
||||||
@@ -338,7 +338,7 @@ impl<NTY: ScalarOps> EventValuesAggregator<NTY> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn result_reset_unweight(&mut self, range: NanoRange, _expand: bool) -> MinMaxAvgDim0Bins<NTY> {
|
fn result_reset_unweight(&mut self, range: NanoRange, _expand: bool) -> BinsDim0<NTY> {
|
||||||
let (min, max, avg) = if self.sumc > 0 {
|
let (min, max, avg) = if self.sumc > 0 {
|
||||||
let avg = self.sum / self.sumc as f32;
|
let avg = self.sum / self.sumc as f32;
|
||||||
(self.min.clone(), self.max.clone(), avg)
|
(self.min.clone(), self.max.clone(), avg)
|
||||||
@@ -349,7 +349,7 @@ impl<NTY: ScalarOps> EventValuesAggregator<NTY> {
|
|||||||
};
|
};
|
||||||
(g.clone(), g.clone(), g.as_prim_f32())
|
(g.clone(), g.clone(), g.as_prim_f32())
|
||||||
};
|
};
|
||||||
let ret = MinMaxAvgDim0Bins {
|
let ret = BinsDim0 {
|
||||||
ts1s: [self.range.beg].into(),
|
ts1s: [self.range.beg].into(),
|
||||||
ts2s: [self.range.end].into(),
|
ts2s: [self.range.end].into(),
|
||||||
counts: [self.count].into(),
|
counts: [self.count].into(),
|
||||||
@@ -365,7 +365,7 @@ impl<NTY: ScalarOps> EventValuesAggregator<NTY> {
|
|||||||
ret
|
ret
|
||||||
}
|
}
|
||||||
|
|
||||||
fn result_reset_time_weight(&mut self, range: NanoRange, expand: bool) -> MinMaxAvgDim0Bins<NTY> {
|
fn result_reset_time_weight(&mut self, range: NanoRange, expand: bool) -> BinsDim0<NTY> {
|
||||||
// TODO check callsite for correct expand status.
|
// TODO check callsite for correct expand status.
|
||||||
if expand {
|
if expand {
|
||||||
debug!("result_reset_time_weight calls apply_event_time_weight");
|
debug!("result_reset_time_weight calls apply_event_time_weight");
|
||||||
@@ -383,7 +383,7 @@ impl<NTY: ScalarOps> EventValuesAggregator<NTY> {
|
|||||||
};
|
};
|
||||||
(g.clone(), g.clone(), g.as_prim_f32())
|
(g.clone(), g.clone(), g.as_prim_f32())
|
||||||
};
|
};
|
||||||
let ret = MinMaxAvgDim0Bins {
|
let ret = BinsDim0 {
|
||||||
ts1s: [self.range.beg].into(),
|
ts1s: [self.range.beg].into(),
|
||||||
ts2s: [self.range.end].into(),
|
ts2s: [self.range.end].into(),
|
||||||
counts: [self.count].into(),
|
counts: [self.count].into(),
|
||||||
@@ -402,14 +402,16 @@ impl<NTY: ScalarOps> EventValuesAggregator<NTY> {
|
|||||||
|
|
||||||
impl<NTY: ScalarOps> TimeBinnableTypeAggregator for EventValuesAggregator<NTY> {
|
impl<NTY: ScalarOps> TimeBinnableTypeAggregator for EventValuesAggregator<NTY> {
|
||||||
type Input = EventsDim0<NTY>;
|
type Input = EventsDim0<NTY>;
|
||||||
type Output = MinMaxAvgDim0Bins<NTY>;
|
type Output = BinsDim0<NTY>;
|
||||||
|
|
||||||
fn range(&self) -> &NanoRange {
|
fn range(&self) -> &NanoRange {
|
||||||
&self.range
|
&self.range
|
||||||
}
|
}
|
||||||
|
|
||||||
fn ingest(&mut self, item: &Self::Input) {
|
fn ingest(&mut self, item: &Self::Input) {
|
||||||
debug!("ingest len {}", item.len());
|
for ts in &item.tss {
|
||||||
|
eprintln!("EventValuesAggregator ingest {ts:20}");
|
||||||
|
}
|
||||||
if self.do_time_weight {
|
if self.do_time_weight {
|
||||||
self.ingest_time_weight(item)
|
self.ingest_time_weight(item)
|
||||||
} else {
|
} else {
|
||||||
@@ -512,6 +514,8 @@ pub struct ScalarEventsTimeBinner<NTY: ScalarOps> {
|
|||||||
|
|
||||||
impl<NTY: ScalarOps> ScalarEventsTimeBinner<NTY> {
|
impl<NTY: ScalarOps> ScalarEventsTimeBinner<NTY> {
|
||||||
fn new(edges: VecDeque<u64>, do_time_weight: bool) -> Self {
|
fn new(edges: VecDeque<u64>, do_time_weight: bool) -> Self {
|
||||||
|
let self_name = std::any::type_name::<Self>();
|
||||||
|
eprintln!("{self_name}::new edges {edges:?}");
|
||||||
Self {
|
Self {
|
||||||
edges,
|
edges,
|
||||||
do_time_weight,
|
do_time_weight,
|
||||||
@@ -661,7 +665,7 @@ impl<NTY: ScalarOps> TimeBinner for ScalarEventsTimeBinner<NTY> {
|
|||||||
self.push_in_progress(true);
|
self.push_in_progress(true);
|
||||||
if self.bins_ready_count() == n {
|
if self.bins_ready_count() == n {
|
||||||
if let Some(_range) = self.next_bin_range() {
|
if let Some(_range) = self.next_bin_range() {
|
||||||
let bins = MinMaxAvgDim0Bins::<NTY>::empty();
|
let bins = BinsDim0::<NTY>::empty();
|
||||||
error!("TODO eventsdim0 time binner append");
|
error!("TODO eventsdim0 time binner append");
|
||||||
err::todo();
|
err::todo();
|
||||||
//bins.append_zero(range.beg, range.end);
|
//bins.append_zero(range.beg, range.end);
|
||||||
|
|||||||
@@ -6,6 +6,7 @@ pub mod test;
|
|||||||
|
|
||||||
use chrono::{DateTime, TimeZone, Utc};
|
use chrono::{DateTime, TimeZone, Utc};
|
||||||
use futures_util::Stream;
|
use futures_util::Stream;
|
||||||
|
use futures_util::StreamExt;
|
||||||
use netpod::log::*;
|
use netpod::log::*;
|
||||||
use netpod::timeunits::*;
|
use netpod::timeunits::*;
|
||||||
use netpod::{AggKind, NanoRange, ScalarType, Shape};
|
use netpod::{AggKind, NanoRange, ScalarType, Shape};
|
||||||
@@ -17,6 +18,7 @@ use std::ops::ControlFlow;
|
|||||||
use std::pin::Pin;
|
use std::pin::Pin;
|
||||||
use std::task::{Context, Poll};
|
use std::task::{Context, Poll};
|
||||||
use streams::Collectable;
|
use streams::Collectable;
|
||||||
|
use streams::ToJsonResult;
|
||||||
|
|
||||||
pub fn bool_is_false(x: &bool) -> bool {
|
pub fn bool_is_false(x: &bool) -> bool {
|
||||||
*x == false
|
*x == false
|
||||||
@@ -35,6 +37,17 @@ pub fn ts_offs_from_abs(tss: &[u64]) -> (u64, VecDeque<u64>, VecDeque<u64>) {
|
|||||||
(ts_anchor_sec, ts_off_ms, ts_off_ns)
|
(ts_anchor_sec, ts_off_ms, ts_off_ns)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub fn ts_offs_from_abs_with_anchor(ts_anchor_sec: u64, tss: &[u64]) -> (VecDeque<u64>, VecDeque<u64>) {
|
||||||
|
let ts_anchor_ns = ts_anchor_sec * SEC;
|
||||||
|
let ts_off_ms: VecDeque<_> = tss.iter().map(|&k| (k - ts_anchor_ns) / MS).collect();
|
||||||
|
let ts_off_ns = tss
|
||||||
|
.iter()
|
||||||
|
.zip(ts_off_ms.iter().map(|&k| k * MS))
|
||||||
|
.map(|(&j, k)| (j - ts_anchor_ns - k))
|
||||||
|
.collect();
|
||||||
|
(ts_off_ms, ts_off_ns)
|
||||||
|
}
|
||||||
|
|
||||||
// TODO take iterator instead of slice, because a VecDeque can't produce a slice in general.
|
// TODO take iterator instead of slice, because a VecDeque can't produce a slice in general.
|
||||||
pub fn pulse_offs_from_abs(pulse: &[u64]) -> (u64, VecDeque<u64>) {
|
pub fn pulse_offs_from_abs(pulse: &[u64]) -> (u64, VecDeque<u64>) {
|
||||||
let pulse_anchor = pulse.first().map_or(0, |k| *k);
|
let pulse_anchor = pulse.first().map_or(0, |k| *k);
|
||||||
@@ -112,6 +125,7 @@ struct Ts(u64);
|
|||||||
|
|
||||||
#[derive(Debug, PartialEq)]
|
#[derive(Debug, PartialEq)]
|
||||||
pub enum ErrorKind {
|
pub enum ErrorKind {
|
||||||
|
General,
|
||||||
#[allow(unused)]
|
#[allow(unused)]
|
||||||
MismatchedType,
|
MismatchedType,
|
||||||
}
|
}
|
||||||
@@ -120,11 +134,27 @@ pub enum ErrorKind {
|
|||||||
pub struct Error {
|
pub struct Error {
|
||||||
#[allow(unused)]
|
#[allow(unused)]
|
||||||
kind: ErrorKind,
|
kind: ErrorKind,
|
||||||
|
msg: Option<String>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl fmt::Display for Error {
|
||||||
|
fn fmt(&self, fmt: &mut fmt::Formatter) -> fmt::Result {
|
||||||
|
write!(fmt, "{self:?}")
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl From<ErrorKind> for Error {
|
impl From<ErrorKind> for Error {
|
||||||
fn from(kind: ErrorKind) -> Self {
|
fn from(kind: ErrorKind) -> Self {
|
||||||
Self { kind }
|
Self { kind, msg: None }
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl From<String> for Error {
|
||||||
|
fn from(msg: String) -> Self {
|
||||||
|
Self {
|
||||||
|
msg: Some(msg),
|
||||||
|
kind: ErrorKind::General,
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -190,9 +220,9 @@ pub fn make_iso_ts(tss: &[u64]) -> Vec<IsoDateTime> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
pub trait TimeBinner: Send {
|
pub trait TimeBinner: Send {
|
||||||
|
fn ingest(&mut self, item: &dyn TimeBinnable);
|
||||||
fn bins_ready_count(&self) -> usize;
|
fn bins_ready_count(&self) -> usize;
|
||||||
fn bins_ready(&mut self) -> Option<Box<dyn TimeBinned>>;
|
fn bins_ready(&mut self) -> Option<Box<dyn TimeBinned>>;
|
||||||
fn ingest(&mut self, item: &dyn TimeBinnable);
|
|
||||||
|
|
||||||
/// If there is a bin in progress with non-zero count, push it to the result set.
|
/// If there is a bin in progress with non-zero count, push it to the result set.
|
||||||
/// With push_empty == true, a bin in progress is pushed even if it contains no counts.
|
/// With push_empty == true, a bin in progress is pushed even if it contains no counts.
|
||||||
@@ -206,13 +236,13 @@ pub trait TimeBinner: Send {
|
|||||||
|
|
||||||
/// Provides a time-binned representation of the implementing type.
|
/// Provides a time-binned representation of the implementing type.
|
||||||
/// In contrast to `TimeBinnableType` this is meant for trait objects.
|
/// In contrast to `TimeBinnableType` this is meant for trait objects.
|
||||||
pub trait TimeBinnable: WithLen + RangeOverlapInfo + Any + Send {
|
pub trait TimeBinnable: fmt::Debug + WithLen + RangeOverlapInfo + Any + Send {
|
||||||
fn time_binner_new(&self, edges: Vec<u64>, do_time_weight: bool) -> Box<dyn TimeBinner>;
|
fn time_binner_new(&self, edges: Vec<u64>, do_time_weight: bool) -> Box<dyn TimeBinner>;
|
||||||
fn as_any(&self) -> &dyn Any;
|
fn as_any(&self) -> &dyn Any;
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Container of some form of events, for use as trait object.
|
/// Container of some form of events, for use as trait object.
|
||||||
pub trait Events: fmt::Debug + Any + Collectable + TimeBinnable {
|
pub trait Events: fmt::Debug + Any + Collectable + TimeBinnable + Send {
|
||||||
fn as_time_binnable_dyn(&self) -> &dyn TimeBinnable;
|
fn as_time_binnable_dyn(&self) -> &dyn TimeBinnable;
|
||||||
fn verify(&self);
|
fn verify(&self);
|
||||||
fn output_info(&self);
|
fn output_info(&self);
|
||||||
@@ -229,8 +259,9 @@ impl PartialEq for Box<dyn Events> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/// Data in time-binned form.
|
/// Data in time-binned form.
|
||||||
pub trait TimeBinned: TimeBinnable {
|
pub trait TimeBinned: Any + TimeBinnable {
|
||||||
fn as_time_binnable_dyn(&self) -> &dyn TimeBinnable;
|
fn as_time_binnable_dyn(&self) -> &dyn TimeBinnable;
|
||||||
|
fn as_collectable_mut(&mut self) -> &mut dyn Collectable;
|
||||||
fn edges_slice(&self) -> (&[u64], &[u64]);
|
fn edges_slice(&self) -> (&[u64], &[u64]);
|
||||||
fn counts(&self) -> &[u64];
|
fn counts(&self) -> &[u64];
|
||||||
fn mins(&self) -> Vec<f32>;
|
fn mins(&self) -> Vec<f32>;
|
||||||
@@ -297,7 +328,7 @@ pub fn empty_binned_dyn(scalar_type: &ScalarType, shape: &Shape, agg_kind: &AggK
|
|||||||
Shape::Scalar => match agg_kind {
|
Shape::Scalar => match agg_kind {
|
||||||
AggKind::TimeWeightedScalar => {
|
AggKind::TimeWeightedScalar => {
|
||||||
use ScalarType::*;
|
use ScalarType::*;
|
||||||
type K<T> = binsdim0::MinMaxAvgDim0Bins<T>;
|
type K<T> = binsdim0::BinsDim0<T>;
|
||||||
match scalar_type {
|
match scalar_type {
|
||||||
U8 => Box::new(K::<u8>::empty()),
|
U8 => Box::new(K::<u8>::empty()),
|
||||||
U16 => Box::new(K::<u16>::empty()),
|
U16 => Box::new(K::<u16>::empty()),
|
||||||
@@ -323,7 +354,7 @@ pub fn empty_binned_dyn(scalar_type: &ScalarType, shape: &Shape, agg_kind: &AggK
|
|||||||
Shape::Wave(_n) => match agg_kind {
|
Shape::Wave(_n) => match agg_kind {
|
||||||
AggKind::DimXBins1 => {
|
AggKind::DimXBins1 => {
|
||||||
use ScalarType::*;
|
use ScalarType::*;
|
||||||
type K<T> = binsdim0::MinMaxAvgDim0Bins<T>;
|
type K<T> = binsdim0::BinsDim0<T>;
|
||||||
match scalar_type {
|
match scalar_type {
|
||||||
U8 => Box::new(K::<u8>::empty()),
|
U8 => Box::new(K::<u8>::empty()),
|
||||||
F32 => Box::new(K::<f32>::empty()),
|
F32 => Box::new(K::<f32>::empty()),
|
||||||
@@ -418,8 +449,8 @@ impl MergableEvents for ChannelEvents {
|
|||||||
}
|
}
|
||||||
|
|
||||||
pub struct ChannelEventsMerger {
|
pub struct ChannelEventsMerger {
|
||||||
inp1: Pin<Box<dyn Stream<Item = Result<ChannelEvents, Error>>>>,
|
inp1: Pin<Box<dyn Stream<Item = Result<ChannelEvents, Error>> + Send>>,
|
||||||
inp2: Pin<Box<dyn Stream<Item = Result<ChannelEvents, Error>>>>,
|
inp2: Pin<Box<dyn Stream<Item = Result<ChannelEvents, Error>> + Send>>,
|
||||||
inp1_done: bool,
|
inp1_done: bool,
|
||||||
inp2_done: bool,
|
inp2_done: bool,
|
||||||
inp1_item: Option<ChannelEvents>,
|
inp1_item: Option<ChannelEvents>,
|
||||||
@@ -432,8 +463,8 @@ pub struct ChannelEventsMerger {
|
|||||||
|
|
||||||
impl ChannelEventsMerger {
|
impl ChannelEventsMerger {
|
||||||
pub fn new(
|
pub fn new(
|
||||||
inp1: Pin<Box<dyn Stream<Item = Result<ChannelEvents, Error>>>>,
|
inp1: Pin<Box<dyn Stream<Item = Result<ChannelEvents, Error>> + Send>>,
|
||||||
inp2: Pin<Box<dyn Stream<Item = Result<ChannelEvents, Error>>>>,
|
inp2: Pin<Box<dyn Stream<Item = Result<ChannelEvents, Error>> + Send>>,
|
||||||
) -> Self {
|
) -> Self {
|
||||||
Self {
|
Self {
|
||||||
done: false,
|
done: false,
|
||||||
@@ -682,3 +713,84 @@ impl Collectable for Box<dyn Collectable> {
|
|||||||
Collectable::as_any_mut(self.as_mut())
|
Collectable::as_any_mut(self.as_mut())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub async fn binned_collected(
|
||||||
|
inp: Pin<Box<dyn Stream<Item = Result<ChannelEvents, Error>> + Send>>,
|
||||||
|
) -> Result<Box<dyn ToJsonResult>, Error> {
|
||||||
|
let mut coll = None;
|
||||||
|
let mut binner = None;
|
||||||
|
let edges: Vec<_> = (0..10).into_iter().map(|t| SEC * 10 * t).collect();
|
||||||
|
let do_time_weight = true;
|
||||||
|
let mut inp = inp;
|
||||||
|
while let Some(item) = inp.next().await {
|
||||||
|
let item = item?;
|
||||||
|
match item {
|
||||||
|
ChannelEvents::Events(events) => {
|
||||||
|
if binner.is_none() {
|
||||||
|
let bb = events
|
||||||
|
.as_time_binnable_dyn()
|
||||||
|
.time_binner_new(edges.clone(), do_time_weight);
|
||||||
|
binner = Some(bb);
|
||||||
|
}
|
||||||
|
let binner = binner.as_mut().unwrap();
|
||||||
|
binner.ingest(events.as_time_binnable_dyn());
|
||||||
|
eprintln!("bins_ready_count: {}", binner.bins_ready_count());
|
||||||
|
if binner.bins_ready_count() > 0 {
|
||||||
|
let ready = binner.bins_ready();
|
||||||
|
match ready {
|
||||||
|
Some(mut ready) => {
|
||||||
|
eprintln!("ready {ready:?}");
|
||||||
|
if coll.is_none() {
|
||||||
|
coll = Some(ready.as_collectable_mut().new_collector());
|
||||||
|
}
|
||||||
|
let cl = coll.as_mut().unwrap();
|
||||||
|
cl.ingest(ready.as_collectable_mut());
|
||||||
|
}
|
||||||
|
None => {
|
||||||
|
return Err(format!("bins_ready_count but no result").into());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
ChannelEvents::Status(_) => {
|
||||||
|
eprintln!("TODO Status");
|
||||||
|
}
|
||||||
|
ChannelEvents::RangeComplete => {
|
||||||
|
eprintln!("TODO RangeComplete");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if let Some(mut binner) = binner {
|
||||||
|
binner.cycle();
|
||||||
|
// TODO merge with the same logic above in the loop.
|
||||||
|
if binner.bins_ready_count() > 0 {
|
||||||
|
let ready = binner.bins_ready();
|
||||||
|
match ready {
|
||||||
|
Some(mut ready) => {
|
||||||
|
eprintln!("ready {ready:?}");
|
||||||
|
if coll.is_none() {
|
||||||
|
coll = Some(ready.as_collectable_mut().new_collector());
|
||||||
|
}
|
||||||
|
let cl = coll.as_mut().unwrap();
|
||||||
|
cl.ingest(ready.as_collectable_mut());
|
||||||
|
}
|
||||||
|
None => {
|
||||||
|
return Err(format!("bins_ready_count but no result").into());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
match coll {
|
||||||
|
Some(mut coll) => {
|
||||||
|
let res = coll.result().map_err(|e| format!("{e}"))?;
|
||||||
|
//let res = res.to_json_result().map_err(|e| format!("{e}"))?;
|
||||||
|
//let res = res.to_json_bytes().map_err(|e| format!("{e}"))?;
|
||||||
|
eprintln!("res {res:?}");
|
||||||
|
Ok(res)
|
||||||
|
}
|
||||||
|
None => {
|
||||||
|
//empty_binned_dyn(scalar_type, shape, agg_kind)
|
||||||
|
Err(format!("TODO produce empty result"))?
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|||||||
@@ -2,6 +2,7 @@ use crate::WithLen;
|
|||||||
use err::Error;
|
use err::Error;
|
||||||
use serde::Serialize;
|
use serde::Serialize;
|
||||||
use std::any::Any;
|
use std::any::Any;
|
||||||
|
use std::fmt;
|
||||||
|
|
||||||
pub trait CollectorType: Send + Unpin + WithLen {
|
pub trait CollectorType: Send + Unpin + WithLen {
|
||||||
type Input: Collectable;
|
type Input: Collectable;
|
||||||
@@ -66,7 +67,7 @@ pub trait ToJsonBytes {
|
|||||||
fn to_json_bytes(&self) -> Result<Vec<u8>, Error>;
|
fn to_json_bytes(&self) -> Result<Vec<u8>, Error>;
|
||||||
}
|
}
|
||||||
|
|
||||||
pub trait ToJsonResult {
|
pub trait ToJsonResult: fmt::Debug + Send {
|
||||||
fn to_json_result(&self) -> Result<Box<dyn ToJsonBytes>, Error>;
|
fn to_json_result(&self) -> Result<Box<dyn ToJsonBytes>, Error>;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -2,6 +2,7 @@ use crate::eventsdim0::EventsDim0;
|
|||||||
use crate::{ChannelEvents, ChannelEventsMerger, ConnStatus, Empty};
|
use crate::{ChannelEvents, ChannelEventsMerger, ConnStatus, Empty};
|
||||||
use crate::{ConnStatusEvent, Error};
|
use crate::{ConnStatusEvent, Error};
|
||||||
use futures_util::StreamExt;
|
use futures_util::StreamExt;
|
||||||
|
use netpod::timeunits::SEC;
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn merge01() {
|
fn merge01() {
|
||||||
@@ -135,3 +136,97 @@ fn merge03() {
|
|||||||
};
|
};
|
||||||
tokio::runtime::Runtime::new().unwrap().block_on(fut);
|
tokio::runtime::Runtime::new().unwrap().block_on(fut);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn bin01() {
|
||||||
|
let fut = async {
|
||||||
|
let mut events_vec1 = Vec::new();
|
||||||
|
for j in 0..2 {
|
||||||
|
let mut events = EventsDim0::empty();
|
||||||
|
for i in 10 * j..10 * (1 + j) {
|
||||||
|
events.push(SEC * i, i, 17f32);
|
||||||
|
}
|
||||||
|
events_vec1.push(Ok(ChannelEvents::Events(Box::new(events))));
|
||||||
|
}
|
||||||
|
let inp1 = events_vec1;
|
||||||
|
let inp1 = futures_util::stream::iter(inp1);
|
||||||
|
let inp1 = Box::pin(inp1);
|
||||||
|
let inp2 = Box::pin(futures_util::stream::empty());
|
||||||
|
let mut stream = ChannelEventsMerger::new(inp1, inp2);
|
||||||
|
let mut coll = None;
|
||||||
|
let mut binner = None;
|
||||||
|
let edges: Vec<_> = (0..10).into_iter().map(|t| SEC * 10 * t).collect();
|
||||||
|
let do_time_weight = true;
|
||||||
|
while let Some(item) = stream.next().await {
|
||||||
|
let item = item?;
|
||||||
|
match item {
|
||||||
|
ChannelEvents::Events(events) => {
|
||||||
|
if binner.is_none() {
|
||||||
|
let bb = events
|
||||||
|
.as_time_binnable_dyn()
|
||||||
|
.time_binner_new(edges.clone(), do_time_weight);
|
||||||
|
binner = Some(bb);
|
||||||
|
}
|
||||||
|
let binner = binner.as_mut().unwrap();
|
||||||
|
binner.ingest(events.as_time_binnable_dyn());
|
||||||
|
eprintln!("bins_ready_count: {}", binner.bins_ready_count());
|
||||||
|
if binner.bins_ready_count() > 0 {
|
||||||
|
let ready = binner.bins_ready();
|
||||||
|
match ready {
|
||||||
|
Some(mut ready) => {
|
||||||
|
eprintln!("ready {ready:?}");
|
||||||
|
if coll.is_none() {
|
||||||
|
coll = Some(ready.as_collectable_mut().new_collector());
|
||||||
|
}
|
||||||
|
let cl = coll.as_mut().unwrap();
|
||||||
|
cl.ingest(ready.as_collectable_mut());
|
||||||
|
}
|
||||||
|
None => {
|
||||||
|
return Err(format!("bins_ready_count but no result").into());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
ChannelEvents::Status(_) => {
|
||||||
|
eprintln!("TODO Status");
|
||||||
|
}
|
||||||
|
ChannelEvents::RangeComplete => {
|
||||||
|
eprintln!("TODO RangeComplete");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if let Some(mut binner) = binner {
|
||||||
|
binner.cycle();
|
||||||
|
// TODO merge with the same logic above in the loop.
|
||||||
|
if binner.bins_ready_count() > 0 {
|
||||||
|
let ready = binner.bins_ready();
|
||||||
|
match ready {
|
||||||
|
Some(mut ready) => {
|
||||||
|
eprintln!("ready {ready:?}");
|
||||||
|
if coll.is_none() {
|
||||||
|
coll = Some(ready.as_collectable_mut().new_collector());
|
||||||
|
}
|
||||||
|
let cl = coll.as_mut().unwrap();
|
||||||
|
cl.ingest(ready.as_collectable_mut());
|
||||||
|
}
|
||||||
|
None => {
|
||||||
|
return Err(format!("bins_ready_count but no result").into());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
match coll {
|
||||||
|
Some(mut coll) => {
|
||||||
|
let res = coll.result().map_err(|e| format!("{e}"))?;
|
||||||
|
//let res = res.to_json_result().map_err(|e| format!("{e}"))?;
|
||||||
|
//let res = res.to_json_bytes().map_err(|e| format!("{e}"))?;
|
||||||
|
eprintln!("res {res:?}");
|
||||||
|
}
|
||||||
|
None => {
|
||||||
|
panic!();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Ok::<_, Error>(())
|
||||||
|
};
|
||||||
|
tokio::runtime::Runtime::new().unwrap().block_on(fut).unwrap();
|
||||||
|
}
|
||||||
|
|||||||
@@ -521,12 +521,12 @@ pub fn agg_kind_from_binning_scheme(pairs: &BTreeMap<String, String>) -> Result<
|
|||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, Debug)]
|
#[derive(Clone, Debug)]
|
||||||
pub struct ChannelStateEvents {
|
pub struct ChannelStateEventsQuery {
|
||||||
channel: Channel,
|
channel: Channel,
|
||||||
range: NanoRange,
|
range: NanoRange,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl ChannelStateEvents {
|
impl ChannelStateEventsQuery {
|
||||||
pub fn new(channel: Channel, range: NanoRange) -> Self {
|
pub fn new(channel: Channel, range: NanoRange) -> Self {
|
||||||
Self { channel, range }
|
Self { channel, range }
|
||||||
}
|
}
|
||||||
@@ -548,19 +548,19 @@ impl ChannelStateEvents {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl HasBackend for ChannelStateEvents {
|
impl HasBackend for ChannelStateEventsQuery {
|
||||||
fn backend(&self) -> &str {
|
fn backend(&self) -> &str {
|
||||||
&self.channel.backend
|
&self.channel.backend
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl HasTimeout for ChannelStateEvents {
|
impl HasTimeout for ChannelStateEventsQuery {
|
||||||
fn timeout(&self) -> Duration {
|
fn timeout(&self) -> Duration {
|
||||||
Duration::from_millis(6000)
|
Duration::from_millis(6000)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl FromUrl for ChannelStateEvents {
|
impl FromUrl for ChannelStateEventsQuery {
|
||||||
fn from_url(url: &Url) -> Result<Self, Error> {
|
fn from_url(url: &Url) -> Result<Self, Error> {
|
||||||
let pairs = get_url_query_pairs(url);
|
let pairs = get_url_query_pairs(url);
|
||||||
Self::from_pairs(&pairs)
|
Self::from_pairs(&pairs)
|
||||||
@@ -582,7 +582,7 @@ impl FromUrl for ChannelStateEvents {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl AppendToUrl for ChannelStateEvents {
|
impl AppendToUrl for ChannelStateEventsQuery {
|
||||||
fn append_to_url(&self, url: &mut Url) {
|
fn append_to_url(&self, url: &mut Url) {
|
||||||
let date_fmt = "%Y-%m-%dT%H:%M:%S.%3fZ";
|
let date_fmt = "%Y-%m-%dT%H:%M:%S.%3fZ";
|
||||||
self.channel.append_to_url(url);
|
self.channel.append_to_url(url);
|
||||||
|
|||||||
@@ -2,7 +2,7 @@ use crate::errconv::ErrConv;
|
|||||||
use crate::events::EventsStreamScylla;
|
use crate::events::EventsStreamScylla;
|
||||||
use err::Error;
|
use err::Error;
|
||||||
use futures_util::{Future, Stream, StreamExt};
|
use futures_util::{Future, Stream, StreamExt};
|
||||||
use items_2::binsdim0::MinMaxAvgDim0Bins;
|
use items_2::binsdim0::BinsDim0;
|
||||||
use items_2::{empty_binned_dyn, empty_events_dyn, ChannelEvents, TimeBinned};
|
use items_2::{empty_binned_dyn, empty_events_dyn, ChannelEvents, TimeBinned};
|
||||||
use netpod::log::*;
|
use netpod::log::*;
|
||||||
use netpod::query::{CacheUsage, PlainEventsQuery, RawEventsQuery};
|
use netpod::query::{CacheUsage, PlainEventsQuery, RawEventsQuery};
|
||||||
@@ -82,7 +82,7 @@ pub async fn read_cached_scylla(
|
|||||||
match &chn.shape {
|
match &chn.shape {
|
||||||
Scalar => match &chn.scalar_type {
|
Scalar => match &chn.scalar_type {
|
||||||
F64 => {
|
F64 => {
|
||||||
let ret = MinMaxAvgDim0Bins::<f64> {
|
let ret = BinsDim0::<f64> {
|
||||||
ts1s,
|
ts1s,
|
||||||
ts2s,
|
ts2s,
|
||||||
counts,
|
counts,
|
||||||
|
|||||||
@@ -2,11 +2,11 @@ use crate::errconv::ErrConv;
|
|||||||
use err::Error;
|
use err::Error;
|
||||||
use futures_util::{Future, FutureExt, Stream, StreamExt};
|
use futures_util::{Future, FutureExt, Stream, StreamExt};
|
||||||
use items_2::eventsdim0::EventsDim0;
|
use items_2::eventsdim0::EventsDim0;
|
||||||
use items_2::{ChannelEvents, Empty, Events};
|
use items_2::{ChannelEvents, ConnStatus, ConnStatusEvent, Empty, Events};
|
||||||
use netpod::log::*;
|
use netpod::log::*;
|
||||||
use netpod::query::{ChannelStateEvents, PlainEventsQuery};
|
use netpod::query::{ChannelStateEventsQuery, PlainEventsQuery};
|
||||||
use netpod::timeunits::*;
|
use netpod::timeunits::*;
|
||||||
use netpod::{Channel, Database, NanoRange, ScalarType, Shape};
|
use netpod::{Channel, NanoRange, ScalarType, Shape};
|
||||||
use scylla::Session as ScySession;
|
use scylla::Session as ScySession;
|
||||||
use std::collections::VecDeque;
|
use std::collections::VecDeque;
|
||||||
use std::pin::Pin;
|
use std::pin::Pin;
|
||||||
@@ -568,39 +568,57 @@ pub async fn make_scylla_stream(
|
|||||||
}
|
}
|
||||||
|
|
||||||
pub async fn channel_state_events(
|
pub async fn channel_state_events(
|
||||||
evq: &ChannelStateEvents,
|
evq: &ChannelStateEventsQuery,
|
||||||
scy: Arc<ScySession>,
|
scy: Arc<ScySession>,
|
||||||
_dbconf: Database,
|
) -> Result<Pin<Box<dyn Stream<Item = Result<ChannelEvents, Error>> + Send>>, Error> {
|
||||||
) -> Result<Vec<(u64, u32)>, Error> {
|
let (tx, rx) = async_channel::bounded(8);
|
||||||
let mut ret = Vec::new();
|
let evq = evq.clone();
|
||||||
let div = DAY;
|
let fut = async move {
|
||||||
let mut ts_msp = evq.range().beg / div * div;
|
let div = DAY;
|
||||||
loop {
|
let mut ts_msp = evq.range().beg / div * div;
|
||||||
let series = (evq
|
loop {
|
||||||
.channel()
|
let series = (evq
|
||||||
.series()
|
.channel()
|
||||||
.ok_or(Error::with_msg_no_trace(format!("series id not given"))))?;
|
.series()
|
||||||
let params = (series as i64, ts_msp as i64);
|
.ok_or(Error::with_msg_no_trace(format!("series id not given"))))?;
|
||||||
let mut res = scy
|
let params = (series as i64, ts_msp as i64);
|
||||||
.query_iter(
|
let mut res = scy
|
||||||
"select ts_lsp, kind from channel_status where series = ? and ts_msp = ?",
|
.query_iter(
|
||||||
params,
|
"select ts_lsp, kind from channel_status where series = ? and ts_msp = ?",
|
||||||
)
|
params,
|
||||||
.await
|
)
|
||||||
.err_conv()?;
|
.await
|
||||||
while let Some(row) = res.next().await {
|
.err_conv()?;
|
||||||
let row = row.err_conv()?;
|
while let Some(row) = res.next().await {
|
||||||
let (ts_lsp, kind): (i64, i32) = row.into_typed().err_conv()?;
|
let row = row.err_conv()?;
|
||||||
let ts = ts_msp + ts_lsp as u64;
|
let (ts_lsp, kind): (i64, i32) = row.into_typed().err_conv()?;
|
||||||
let kind = kind as u32;
|
let ts = ts_msp + ts_lsp as u64;
|
||||||
if ts >= evq.range().beg && ts < evq.range().end {
|
let kind = kind as u32;
|
||||||
ret.push((ts, kind));
|
if ts >= evq.range().beg && ts < evq.range().end {
|
||||||
|
let status = match kind {
|
||||||
|
1 => ConnStatus::Connect,
|
||||||
|
2 => ConnStatus::Disconnect,
|
||||||
|
_ => {
|
||||||
|
let e = Error::with_msg_no_trace(format!("bad status kind {kind}"));
|
||||||
|
let e2 = Error::with_msg_no_trace(format!("bad status kind {kind}"));
|
||||||
|
let _ = tx.send(Err(e)).await;
|
||||||
|
return Err(e2);
|
||||||
|
}
|
||||||
|
};
|
||||||
|
let ev = ConnStatusEvent { ts, status };
|
||||||
|
tx.send(Ok(ChannelEvents::Status(ev)))
|
||||||
|
.await
|
||||||
|
.map_err(|e| format!("{e}"))?;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
ts_msp += DAY;
|
||||||
|
if ts_msp >= evq.range().end {
|
||||||
|
break;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
ts_msp += DAY;
|
Ok(())
|
||||||
if ts_msp >= evq.range().end {
|
};
|
||||||
break;
|
// TODO join the task (better: rewrite as proper stream)
|
||||||
}
|
tokio::spawn(fut);
|
||||||
}
|
Ok(Box::pin(rx))
|
||||||
Ok(ret)
|
|
||||||
}
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user