file_name
large_stringlengths
4
140
prefix
large_stringlengths
0
12.1k
suffix
large_stringlengths
0
12k
middle
large_stringlengths
0
7.51k
fim_type
large_stringclasses
4 values
binder.rs
capability::{ CapabilityProvider, CapabilitySource, ComponentCapability, InternalCapability, OptionalTask, }, channel, model::{ component::{BindReason, WeakComponentInstance}, error::ModelError, hooks::{Event, EventPayload, EventType, Hook, HooksRegistration}, model::Model, routing::report_routing_failure, }, }, async_trait::async_trait, cm_rust::{CapabilityName, CapabilityPath, ProtocolDecl}, fuchsia_async as fasync, fuchsia_zircon as zx, lazy_static::lazy_static, moniker::{AbsoluteMoniker, AbsoluteMonikerBase, ExtendedMoniker}, std::{ path::PathBuf, sync::{Arc, Weak}, }, }; lazy_static! { pub static ref BINDER_SERVICE: CapabilityName = "fuchsia.component.Binder".into(); pub static ref BINDER_CAPABILITY: ComponentCapability = ComponentCapability::Protocol(ProtocolDecl { name: BINDER_SERVICE.clone(), source_path: Some(CapabilityPath { basename: "fuchsia.component.Binder".into(), dirname: "svc".into() }), }); } /// Implementation of `fuchsia.component.Binder` FIDL protocol. pub struct BinderCapabilityProvider { source: WeakComponentInstance, target: WeakComponentInstance, host: Arc<BinderCapabilityHost>, } impl BinderCapabilityProvider { pub fn new( source: WeakComponentInstance, target: WeakComponentInstance, host: Arc<BinderCapabilityHost>, ) -> Self { Self { source, target, host } } } #[async_trait] impl CapabilityProvider for BinderCapabilityProvider { async fn open( self: Box<Self>, _flags: u32, _open_mode: u32, _relative_path: PathBuf, server_end: &mut zx::Channel, ) -> Result<OptionalTask, ModelError> { let host = self.host.clone(); let target = self.target.clone(); let source = self.source.clone(); let server_end = channel::take_channel(server_end); Ok(fasync::Task::spawn(async move { if let Err(err) = host.bind(source).await { let res = target.upgrade().map_err(|e| ModelError::from(e)); match res { Ok(target) => { report_routing_failure(&target, &*BINDER_CAPABILITY, &err, server_end) .await; } Err(err) => { log::warn!("failed to upgrade reference to {}: {}", target.moniker, err); } } } }) .into()) } } // A `Hook` that serves the `fuchsia.component.Binder` FIDL protocol. #[derive(Clone)] pub struct BinderCapabilityHost { model: Weak<Model>, } impl BinderCapabilityHost { pub fn new(model: Weak<Model>) -> Self { Self { model } } pub fn hooks(self: &Arc<Self>) -> Vec<HooksRegistration> { vec![HooksRegistration::new( "BinderCapabilityHost", vec![EventType::CapabilityRouted], Arc::downgrade(self) as Weak<dyn Hook>, )] } pub async fn bind(&self, source: WeakComponentInstance) -> Result<(), ModelError> { let source = source.upgrade().map_err(|e| ModelError::from(e))?; source.bind(&BindReason::Binder).await?; Ok(()) } async fn on_scoped_framework_capability_routed_async<'a>( self: Arc<Self>, source: WeakComponentInstance, target_moniker: AbsoluteMoniker, capability: &'a InternalCapability, capability_provider: Option<Box<dyn CapabilityProvider>>, ) -> Result<Option<Box<dyn CapabilityProvider>>, ModelError> { // If some other capability has already been installed, then there's nothing to // do here. if capability_provider.is_none() && capability.matches_protocol(&BINDER_SERVICE) { let model = self.model.upgrade().ok_or(ModelError::ModelNotAvailable)?; let target = WeakComponentInstance::new(&model.look_up(&target_moniker.to_partial()).await?); Ok(Some(Box::new(BinderCapabilityProvider::new(source, target, self.clone())) as Box<dyn CapabilityProvider>)) } else { Ok(capability_provider) } } } #[async_trait] impl Hook for BinderCapabilityHost { async fn on(self: Arc<Self>, event: &Event) -> Result<(), ModelError> { if let Ok(EventPayload::CapabilityRouted { source: CapabilitySource::Framework { capability, component }, capability_provider, }) = &event.result { let target_moniker = match &event.target_moniker { ExtendedMoniker::ComponentManager => { Err(ModelError::UnexpectedComponentManagerMoniker) } ExtendedMoniker::ComponentInstance(moniker) => Ok(moniker), }?; let mut capability_provider = capability_provider.lock().await; *capability_provider = self .on_scoped_framework_capability_routed_async( component.clone(), target_moniker.clone(), &capability, capability_provider.take(), ) .await?; } Ok(()) } } #[cfg(test)] mod tests { use { super::*, crate::{ builtin_environment::BuiltinEnvironment, capability::CapabilityProvider, model::{ events::{source::EventSource, stream::EventStream}, testing::test_helpers::*, }, }, cm_rust::{self, CapabilityName, ComponentDecl, EventMode}, cm_rust_testing::*, fidl::{client::Client, handle::AsyncChannel}, fuchsia_zircon as zx, futures::{lock::Mutex, StreamExt}, matches::assert_matches, moniker::AbsoluteMoniker, std::path::PathBuf, }; struct BinderCapabilityTestFixture { builtin_environment: Arc<Mutex<BuiltinEnvironment>>, } impl BinderCapabilityTestFixture { async fn
(components: Vec<(&'static str, ComponentDecl)>) -> Self { let TestModelResult { builtin_environment, .. } = TestEnvironmentBuilder::new().set_components(components).build().await; BinderCapabilityTestFixture { builtin_environment } } async fn new_event_stream( &self, events: Vec<CapabilityName>, mode: EventMode, ) -> (EventSource, EventStream) { new_event_stream(self.builtin_environment.clone(), events, mode).await } async fn provider( &self, source: AbsoluteMoniker, target: AbsoluteMoniker, ) -> Box<BinderCapabilityProvider> { let builtin_environment = self.builtin_environment.lock().await; let host = builtin_environment.binder_capability_host.clone(); let source = builtin_environment .model .look_up(&source.to_partial()) .await .expect("failed to look up source moniker"); let target = builtin_environment .model .look_up(&target.to_partial()) .await .expect("failed to look up target moniker"); Box::new(BinderCapabilityProvider::new( WeakComponentInstance::new(&source), WeakComponentInstance::new(&target), host, )) } } #[fuchsia::test] async fn component_starts_on_open() { let fixture = BinderCapabilityTestFixture::new(vec![ ( "root", ComponentDeclBuilder::new() .add_lazy_child("source") .add_lazy_child("target") .build(), ), ("source", component_decl_with_test_runner()), ("target", component_decl_with_test_runner()), ]) .await; let (_event_source, mut event_stream) = fixture .new_event_stream( vec![EventType::Resolved.into(), EventType::Started.into()], EventMode::Async, ) .await; let (_client_end, mut server_end) = zx::Channel::create().expect("failed to create channels"); let moniker: AbsoluteMoniker = vec!["source:0"].into(); let () = fixture .provider(moniker.clone(), vec!["target:0"].into()) .await .open(0, 0, PathBuf::new(), &mut server_end) .await .expect("failed to call open()") .take() .expect("task is empty") .await; assert!(event_stream.wait_until(EventType::Resolved, moniker.clone()).await.is_some()); assert!(event_stream.wait_until(EventType::Started, moniker.clone()).await.is_some()); } // TODO(yaneury): Figure out a way to test this behavior. #[ignore] #[fuchsia::test] async fn channel_is_closed_if_component_does_not_exist() { let fixture = BinderCapabilityTestFixture::new(vec![( "root", ComponentDeclBuilder::new() .add_lazy_child("target") .add_lazy_child("unresolvable") .build(), )]) .await; let (client_end, mut server_end) = zx::Channel::create().expect("failed to create channels"); let moniker: AbsoluteMoniker = AbsoluteMoniker::from(vec!["foo:0"]); let () = fixture .provider(moniker, vec![].into()) .await .open(0, 0, PathBuf::new(), &mut server_end) .await .expect("failed to call open()") .take() .expect("task is empty") .await; let client
new
identifier_name
consolidate_data.py
[..., :3], [1.0/3.0, 1.0/3.0, 1.0/3.0]), mi, mx) # https://github.com/CSBDeep/CSBDeep/blob/master/csbdeep/utils/utils.py def normalize_percentile(x, pmin=3, pmax=99.8, axis=None, clip=False, eps=1e-20, dtype=np.float32): mi = np.percentile(x, pmin, axis=axis, keepdims=True) ma = np.percentile(x, pmax, axis=axis, keepdims=True) logger.debug("min/max %s %s %s %s", mi, ma, np.min(x), np.max(x)) return normalize_min_max(x, mi, ma, clip=clip, eps=eps, dtype=dtype) def normalize_min_max(x, mi, ma, clip=False, eps=1e-20, dtype=np.float32): if mi is None: mi = np.min(x) if ma is None: ma = np.max(x) if dtype is not None: x = x.astype(dtype, copy=False) mi = dtype(mi) if np.isscalar(mi) else mi.astype(dtype, copy=False) ma = dtype(ma) if np.isscalar(ma) else ma.astype(dtype, copy=False) eps = dtype(eps) try: import numexpr x = numexpr.evaluate("(x - mi) / ( ma - mi + eps )") except ImportError: x = (x - mi) / ( ma - mi + eps ) if clip: x = np.clip(x, 0, 1) return x def normalize(args, raw, sample, mn, mx): logger.debug( "%s before norm %s: min %s, max %s, mean %s, std %s, median %s", sample, args.normalize, np.min(raw), np.max(raw), np.mean(raw), np.std(raw), np.median(raw)) if args.normalize == "minmax":
elif args.normalize == "percentile": raw = normalize_percentile(raw, mn, mx) logger.debug( "%s after norm %s: min %s, max %s, mean %s, std %s, median %s", sample, args.normalize, np.min(raw), np.max(raw), np.mean(raw), np.std(raw), np.median(raw)) return raw def preprocess(args, raw, sample): logger.debug("%s before preproc %s: skew %s", sample, args.preprocess, scipy.stats.skew(raw.ravel())) if args.preprocess is None or args.preprocess == "no": pass elif args.preprocess == "square": raw = np.square(raw) elif args.preprocess == "cuberoot": raw = np.cbrt(raw) logger.debug("%s after preproc %s: skew %s", sample, args.preprocess, scipy.stats.skew(raw.ravel())) return raw def pad(args, array, mode): if args.padding != 0: array = np.pad(array, ((args.padding, args.padding), (args.padding, args.padding), (args.padding, args.padding)), mode) return array def get_arguments(): parser = argparse.ArgumentParser() parser.add_argument('-i', '--in-dir', dest='in_dir', required=True, help='location of input files') parser.add_argument('-o', '--out-dir', dest='out_dir', required=True, help='where to place output files') parser.add_argument('--out-format', dest='out_format', default="hdf", help='format of output files') parser.add_argument('-p', '--parallel', default=1, type=int) parser.add_argument('--raw-gfp-min', dest='raw_gfp_min', type=int) parser.add_argument('--raw-gfp-max', dest='raw_gfp_max', type=int) parser.add_argument('--raw-bf-min', dest='raw_bf_min', type=int) parser.add_argument('--raw-bf-max', dest='raw_bf_max', type=int) parser.add_argument('--normalize', default='minmax', choices=['minmax', 'percentile', 'meanstd']) parser.add_argument('--preprocess', default='no', choices=['no', 'square', 'cuberoot']) parser.add_argument('--padding', default=0, type=int) args = parser.parse_args() os.makedirs(args.out_dir, exist_ok=True) return args def main(): logging.basicConfig(level='INFO') args = get_arguments() files = map( lambda fn: fn.split("/")[-1].split(".")[0].split("_")[6], glob.glob(os.path.join(args.in_dir, 'BBBC010_v2_images/*.tif'))) files = sorted(list(set(files))) print(files) if args.parallel > 1: Parallel(n_jobs=args.parallel, verbose=1, backend='multiprocessing') \ (delayed(work)(args, f) for f in files) else: for f in files: work(args, f) def work(args, sample): logger.info("Processing %s, %s", args.in_dir, sample) out_fn = os.path.join(args.out_dir, sample) raw_fns = natsorted(glob.glob( os.path.join(args.in_dir, "BBBC010_v2_images", "*_" + sample + "_*.tif"))) # print(raw_fns) raw_gfp = load_array(raw_fns[0]).astype(np.float32) # print(raw_fns[0], np.min(raw_gfp), np.max(raw_gfp)) raw_gfp = preprocess(args, raw_gfp, sample) raw_gfp = normalize(args, raw_gfp, sample, args.raw_gfp_min, args.raw_gfp_max) raw_gfp = pad(args, raw_gfp, 'constant') raw_bf = load_array(raw_fns[1]).astype(np.float32) # print(raw_fns[1], np.min(raw_bf), np.max(raw_bf)) raw_bf = preprocess(args, raw_bf, sample) raw_bf = normalize(args, raw_bf, sample, args.raw_bf_min, args.raw_bf_max) raw_bf = pad(args, raw_bf, 'constant') files = natsorted( glob.glob(os.path.join(args.in_dir, "BBBC010_v1_foreground_eachworm", sample + "*" + "_ground_truth.png"))) # print(files) logger.info("number files: %s", len(files)) gt_labels = np.zeros((len(files),) + raw_gfp.shape, dtype=np.uint16) gt_instances = np.zeros(raw_gfp.shape, dtype=np.uint16) for idx, f in enumerate(files): gt_label = load_array(f).astype(np.uint16) gt_labels[idx, ...] = 1*(gt_label!=0) gt_instances[gt_label != 0] = idx+1 gt_numinst = np.sum(gt_labels, axis=0) tmp = np.sum(gt_labels, axis=0, keepdims=True) tmp[tmp==0] = 1 gt_labels_norm = (gt_labels.astype(np.float32) / tmp.astype(np.float32)).astype(np.float32) gt_fgbg = np.zeros(gt_labels.shape[1:], dtype=np.uint8) gt_fgbg[np.sum(gt_labels, axis=0) > 0] = 1 # is slightly displaced +1,+0 # gt_fgbg2 = load_array( # os.path.join(args.in_dir, "BBBC010_v1_foreground", # sample + "_binary.png")) # gt_fgbg2[0:-1,1:] = gt_fgbg2[1:,1:] # print(np.count_nonzero(gt_fgbg2 != gt_fgbg)) # gt_fgbg2 = pad(args, gt_fgbg2, 'constant') gt_labels = pad(args, gt_labels, 'constant') gt_instances = pad(args, gt_instances, 'constant') gt_labels_norm = pad(args, gt_labels_norm, 'constant') gt_numinst = pad(args, gt_numinst, 'constant') gt_fgbg = pad(args, gt_fgbg, 'constant') if raw_gfp.shape[0] != 1: raw_gfp = np.expand_dims(raw_gfp, 0) if raw_bf.shape[0] != 1: raw_bf = np.expand_dims(raw_bf, 0) raw = np.concatenate((raw_gfp, raw_bf), axis=0) if gt_instances.shape[0] != 1: gt_instances = np.expand_dims(gt_instances, 0) gt_numinst = np.expand_dims(gt_numinst, 0) gt_fgbg = np.expand_dims(gt_fgbg, 0) gt_labels_padded = np.pad(gt_labels, (0, 20-len(files)), 'constant') gt_labels_norm_padded = np.pad(gt_labels_norm, (0, 20-len(files)), 'constant') if args
raw = normalize_min_max(raw, mn, mx)
conditional_block
consolidate_data.py
[..., :3], [1.0/3.0, 1.0/3.0, 1.0/3.0]), mi, mx) # https://github.com/CSBDeep/CSBDeep/blob/master/csbdeep/utils/utils.py def normalize_percentile(x, pmin=3, pmax=99.8, axis=None, clip=False, eps=1e-20, dtype=np.float32): mi = np.percentile(x, pmin, axis=axis, keepdims=True) ma = np.percentile(x, pmax, axis=axis, keepdims=True) logger.debug("min/max %s %s %s %s", mi, ma, np.min(x), np.max(x)) return normalize_min_max(x, mi, ma, clip=clip, eps=eps, dtype=dtype) def normalize_min_max(x, mi, ma, clip=False, eps=1e-20, dtype=np.float32): if mi is None: mi = np.min(x) if ma is None: ma = np.max(x) if dtype is not None: x = x.astype(dtype, copy=False) mi = dtype(mi) if np.isscalar(mi) else mi.astype(dtype, copy=False) ma = dtype(ma) if np.isscalar(ma) else ma.astype(dtype, copy=False) eps = dtype(eps) try: import numexpr x = numexpr.evaluate("(x - mi) / ( ma - mi + eps )") except ImportError: x = (x - mi) / ( ma - mi + eps ) if clip: x = np.clip(x, 0, 1) return x def normalize(args, raw, sample, mn, mx): logger.debug( "%s before norm %s: min %s, max %s, mean %s, std %s, median %s", sample, args.normalize, np.min(raw), np.max(raw), np.mean(raw), np.std(raw), np.median(raw)) if args.normalize == "minmax": raw = normalize_min_max(raw, mn, mx) elif args.normalize == "percentile": raw = normalize_percentile(raw, mn, mx) logger.debug( "%s after norm %s: min %s, max %s, mean %s, std %s, median %s", sample, args.normalize, np.min(raw), np.max(raw), np.mean(raw), np.std(raw), np.median(raw)) return raw def preprocess(args, raw, sample): logger.debug("%s before preproc %s: skew %s", sample, args.preprocess, scipy.stats.skew(raw.ravel())) if args.preprocess is None or args.preprocess == "no": pass elif args.preprocess == "square": raw = np.square(raw) elif args.preprocess == "cuberoot": raw = np.cbrt(raw) logger.debug("%s after preproc %s: skew %s", sample, args.preprocess, scipy.stats.skew(raw.ravel())) return raw def
(args, array, mode): if args.padding != 0: array = np.pad(array, ((args.padding, args.padding), (args.padding, args.padding), (args.padding, args.padding)), mode) return array def get_arguments(): parser = argparse.ArgumentParser() parser.add_argument('-i', '--in-dir', dest='in_dir', required=True, help='location of input files') parser.add_argument('-o', '--out-dir', dest='out_dir', required=True, help='where to place output files') parser.add_argument('--out-format', dest='out_format', default="hdf", help='format of output files') parser.add_argument('-p', '--parallel', default=1, type=int) parser.add_argument('--raw-gfp-min', dest='raw_gfp_min', type=int) parser.add_argument('--raw-gfp-max', dest='raw_gfp_max', type=int) parser.add_argument('--raw-bf-min', dest='raw_bf_min', type=int) parser.add_argument('--raw-bf-max', dest='raw_bf_max', type=int) parser.add_argument('--normalize', default='minmax', choices=['minmax', 'percentile', 'meanstd']) parser.add_argument('--preprocess', default='no', choices=['no', 'square', 'cuberoot']) parser.add_argument('--padding', default=0, type=int) args = parser.parse_args() os.makedirs(args.out_dir, exist_ok=True) return args def main(): logging.basicConfig(level='INFO') args = get_arguments() files = map( lambda fn: fn.split("/")[-1].split(".")[0].split("_")[6], glob.glob(os.path.join(args.in_dir, 'BBBC010_v2_images/*.tif'))) files = sorted(list(set(files))) print(files) if args.parallel > 1: Parallel(n_jobs=args.parallel, verbose=1, backend='multiprocessing') \ (delayed(work)(args, f) for f in files) else: for f in files: work(args, f) def work(args, sample): logger.info("Processing %s, %s", args.in_dir, sample) out_fn = os.path.join(args.out_dir, sample) raw_fns = natsorted(glob.glob( os.path.join(args.in_dir, "BBBC010_v2_images", "*_" + sample + "_*.tif"))) # print(raw_fns) raw_gfp = load_array(raw_fns[0]).astype(np.float32) # print(raw_fns[0], np.min(raw_gfp), np.max(raw_gfp)) raw_gfp = preprocess(args, raw_gfp, sample) raw_gfp = normalize(args, raw_gfp, sample, args.raw_gfp_min, args.raw_gfp_max) raw_gfp = pad(args, raw_gfp, 'constant') raw_bf = load_array(raw_fns[1]).astype(np.float32) # print(raw_fns[1], np.min(raw_bf), np.max(raw_bf)) raw_bf = preprocess(args, raw_bf, sample) raw_bf = normalize(args, raw_bf, sample, args.raw_bf_min, args.raw_bf_max) raw_bf = pad(args, raw_bf, 'constant') files = natsorted( glob.glob(os.path.join(args.in_dir, "BBBC010_v1_foreground_eachworm", sample + "*" + "_ground_truth.png"))) # print(files) logger.info("number files: %s", len(files)) gt_labels = np.zeros((len(files),) + raw_gfp.shape, dtype=np.uint16) gt_instances = np.zeros(raw_gfp.shape, dtype=np.uint16) for idx, f in enumerate(files): gt_label = load_array(f).astype(np.uint16) gt_labels[idx, ...] = 1*(gt_label!=0) gt_instances[gt_label != 0] = idx+1 gt_numinst = np.sum(gt_labels, axis=0) tmp = np.sum(gt_labels, axis=0, keepdims=True) tmp[tmp==0] = 1 gt_labels_norm = (gt_labels.astype(np.float32) / tmp.astype(np.float32)).astype(np.float32) gt_fgbg = np.zeros(gt_labels.shape[1:], dtype=np.uint8) gt_fgbg[np.sum(gt_labels, axis=0) > 0] = 1 # is slightly displaced +1,+0 # gt_fgbg2 = load_array( # os.path.join(args.in_dir, "BBBC010_v1_foreground", # sample + "_binary.png")) # gt_fgbg2[0:-1,1:] = gt_fgbg2[1:,1:] # print(np.count_nonzero(gt_fgbg2 != gt_fgbg)) # gt_fgbg2 = pad(args, gt_fgbg2, 'constant') gt_labels = pad(args, gt_labels, 'constant') gt_instances = pad(args, gt_instances, 'constant') gt_labels_norm = pad(args, gt_labels_norm, 'constant') gt_numinst = pad(args, gt_numinst, 'constant') gt_fgbg = pad(args, gt_fgbg, 'constant') if raw_gfp.shape[0] != 1: raw_gfp = np.expand_dims(raw_gfp, 0) if raw_bf.shape[0] != 1: raw_bf = np.expand_dims(raw_bf, 0) raw = np.concatenate((raw_gfp, raw_bf), axis=0) if gt_instances.shape[0] != 1: gt_instances = np.expand_dims(gt_instances, 0) gt_numinst = np.expand_dims(gt_numinst, 0) gt_fgbg = np.expand_dims(gt_fgbg, 0) gt_labels_padded = np.pad(gt_labels, (0, 20-len(files)), 'constant') gt_labels_norm_padded = np.pad(gt_labels_norm, (0, 20-len(files)), 'constant') if args
pad
identifier_name
consolidate_data.py
) # https://github.com/CSBDeep/CSBDeep/blob/master/csbdeep/utils/utils.py def normalize_percentile(x, pmin=3, pmax=99.8, axis=None, clip=False, eps=1e-20, dtype=np.float32): mi = np.percentile(x, pmin, axis=axis, keepdims=True) ma = np.percentile(x, pmax, axis=axis, keepdims=True) logger.debug("min/max %s %s %s %s", mi, ma, np.min(x), np.max(x)) return normalize_min_max(x, mi, ma, clip=clip, eps=eps, dtype=dtype) def normalize_min_max(x, mi, ma, clip=False, eps=1e-20, dtype=np.float32): if mi is None: mi = np.min(x) if ma is None: ma = np.max(x) if dtype is not None: x = x.astype(dtype, copy=False) mi = dtype(mi) if np.isscalar(mi) else mi.astype(dtype, copy=False) ma = dtype(ma) if np.isscalar(ma) else ma.astype(dtype, copy=False) eps = dtype(eps) try: import numexpr x = numexpr.evaluate("(x - mi) / ( ma - mi + eps )") except ImportError: x = (x - mi) / ( ma - mi + eps ) if clip: x = np.clip(x, 0, 1) return x def normalize(args, raw, sample, mn, mx): logger.debug( "%s before norm %s: min %s, max %s, mean %s, std %s, median %s", sample, args.normalize, np.min(raw), np.max(raw), np.mean(raw), np.std(raw), np.median(raw)) if args.normalize == "minmax": raw = normalize_min_max(raw, mn, mx) elif args.normalize == "percentile": raw = normalize_percentile(raw, mn, mx) logger.debug( "%s after norm %s: min %s, max %s, mean %s, std %s, median %s", sample, args.normalize, np.min(raw), np.max(raw), np.mean(raw), np.std(raw), np.median(raw)) return raw def preprocess(args, raw, sample): logger.debug("%s before preproc %s: skew %s", sample, args.preprocess, scipy.stats.skew(raw.ravel())) if args.preprocess is None or args.preprocess == "no": pass elif args.preprocess == "square": raw = np.square(raw) elif args.preprocess == "cuberoot": raw = np.cbrt(raw) logger.debug("%s after preproc %s: skew %s", sample, args.preprocess, scipy.stats.skew(raw.ravel())) return raw def pad(args, array, mode): if args.padding != 0: array = np.pad(array, ((args.padding, args.padding), (args.padding, args.padding), (args.padding, args.padding)), mode) return array def get_arguments(): parser = argparse.ArgumentParser() parser.add_argument('-i', '--in-dir', dest='in_dir', required=True, help='location of input files') parser.add_argument('-o', '--out-dir', dest='out_dir', required=True, help='where to place output files') parser.add_argument('--out-format', dest='out_format', default="hdf", help='format of output files') parser.add_argument('-p', '--parallel', default=1, type=int) parser.add_argument('--raw-gfp-min', dest='raw_gfp_min', type=int) parser.add_argument('--raw-gfp-max', dest='raw_gfp_max', type=int) parser.add_argument('--raw-bf-min', dest='raw_bf_min', type=int) parser.add_argument('--raw-bf-max', dest='raw_bf_max', type=int) parser.add_argument('--normalize', default='minmax', choices=['minmax', 'percentile', 'meanstd']) parser.add_argument('--preprocess', default='no', choices=['no', 'square', 'cuberoot']) parser.add_argument('--padding', default=0, type=int) args = parser.parse_args() os.makedirs(args.out_dir, exist_ok=True) return args def main(): logging.basicConfig(level='INFO') args = get_arguments() files = map( lambda fn: fn.split("/")[-1].split(".")[0].split("_")[6], glob.glob(os.path.join(args.in_dir, 'BBBC010_v2_images/*.tif'))) files = sorted(list(set(files))) print(files) if args.parallel > 1: Parallel(n_jobs=args.parallel, verbose=1, backend='multiprocessing') \ (delayed(work)(args, f) for f in files) else: for f in files: work(args, f) def work(args, sample): logger.info("Processing %s, %s", args.in_dir, sample) out_fn = os.path.join(args.out_dir, sample) raw_fns = natsorted(glob.glob( os.path.join(args.in_dir, "BBBC010_v2_images", "*_" + sample + "_*.tif"))) # print(raw_fns) raw_gfp = load_array(raw_fns[0]).astype(np.float32) # print(raw_fns[0], np.min(raw_gfp), np.max(raw_gfp)) raw_gfp = preprocess(args, raw_gfp, sample) raw_gfp = normalize(args, raw_gfp, sample, args.raw_gfp_min, args.raw_gfp_max) raw_gfp = pad(args, raw_gfp, 'constant') raw_bf = load_array(raw_fns[1]).astype(np.float32) # print(raw_fns[1], np.min(raw_bf), np.max(raw_bf)) raw_bf = preprocess(args, raw_bf, sample) raw_bf = normalize(args, raw_bf, sample, args.raw_bf_min, args.raw_bf_max) raw_bf = pad(args, raw_bf, 'constant') files = natsorted( glob.glob(os.path.join(args.in_dir, "BBBC010_v1_foreground_eachworm", sample + "*" + "_ground_truth.png"))) # print(files) logger.info("number files: %s", len(files)) gt_labels = np.zeros((len(files),) + raw_gfp.shape, dtype=np.uint16) gt_instances = np.zeros(raw_gfp.shape, dtype=np.uint16) for idx, f in enumerate(files): gt_label = load_array(f).astype(np.uint16) gt_labels[idx, ...] = 1*(gt_label!=0) gt_instances[gt_label != 0] = idx+1 gt_numinst = np.sum(gt_labels, axis=0) tmp = np.sum(gt_labels, axis=0, keepdims=True) tmp[tmp==0] = 1 gt_labels_norm = (gt_labels.astype(np.float32) / tmp.astype(np.float32)).astype(np.float32) gt_fgbg = np.zeros(gt_labels.shape[1:], dtype=np.uint8) gt_fgbg[np.sum(gt_labels, axis=0) > 0] = 1 # is slightly displaced +1,+0 # gt_fgbg2 = load_array( # os.path.join(args.in_dir, "BBBC010_v1_foreground", # sample + "_binary.png")) # gt_fgbg2[0:-1,1:] = gt_fgbg2[1:,1:] # print(np.count_nonzero(gt_fgbg2 != gt_fgbg)) # gt_fgbg2 = pad(args, gt_fgbg2, 'constant') gt_labels = pad(args, gt_labels, 'constant') gt_instances = pad(args, gt_instances, 'constant') gt_labels_norm = pad(args, gt_labels_norm, 'constant') gt_numinst = pad(args, gt_numinst, 'constant') gt_fgbg = pad(args, gt_fgbg, 'constant') if raw_gfp.shape[0] != 1: raw_gfp = np.expand_dims(raw_gfp, 0) if raw_bf.shape[0] != 1: raw_bf = np.expand_dims(raw_bf, 0) raw = np.concatenate((raw_gfp, raw_bf), axis=0) if gt_instances.shape[0] != 1: gt_instances = np.expand_dims(gt_instances, 0) gt_numinst = np.expand_dims(gt_numinst, 0) gt_fgbg = np.expand_dims(gt_fgbg, 0) gt_labels_padded = np.pad(gt_labels, (0, 20-len(files)), 'constant') gt_labels_norm_padded = np.pad(gt_labels_norm, (0, 20-len(files)), 'constant')
if args.out_format == "hdf": f = h5py.File(out_fn + '.hdf', 'w') elif args.out_format == "zarr":
random_line_split
consolidate_data.py
[..., :3], [1.0/3.0, 1.0/3.0, 1.0/3.0]), mi, mx) # https://github.com/CSBDeep/CSBDeep/blob/master/csbdeep/utils/utils.py def normalize_percentile(x, pmin=3, pmax=99.8, axis=None, clip=False, eps=1e-20, dtype=np.float32): mi = np.percentile(x, pmin, axis=axis, keepdims=True) ma = np.percentile(x, pmax, axis=axis, keepdims=True) logger.debug("min/max %s %s %s %s", mi, ma, np.min(x), np.max(x)) return normalize_min_max(x, mi, ma, clip=clip, eps=eps, dtype=dtype) def normalize_min_max(x, mi, ma, clip=False, eps=1e-20, dtype=np.float32): if mi is None: mi = np.min(x) if ma is None: ma = np.max(x) if dtype is not None: x = x.astype(dtype, copy=False) mi = dtype(mi) if np.isscalar(mi) else mi.astype(dtype, copy=False) ma = dtype(ma) if np.isscalar(ma) else ma.astype(dtype, copy=False) eps = dtype(eps) try: import numexpr x = numexpr.evaluate("(x - mi) / ( ma - mi + eps )") except ImportError: x = (x - mi) / ( ma - mi + eps ) if clip: x = np.clip(x, 0, 1) return x def normalize(args, raw, sample, mn, mx): logger.debug( "%s before norm %s: min %s, max %s, mean %s, std %s, median %s", sample, args.normalize, np.min(raw), np.max(raw), np.mean(raw), np.std(raw), np.median(raw)) if args.normalize == "minmax": raw = normalize_min_max(raw, mn, mx) elif args.normalize == "percentile": raw = normalize_percentile(raw, mn, mx) logger.debug( "%s after norm %s: min %s, max %s, mean %s, std %s, median %s", sample, args.normalize, np.min(raw), np.max(raw), np.mean(raw), np.std(raw), np.median(raw)) return raw def preprocess(args, raw, sample): logger.debug("%s before preproc %s: skew %s", sample, args.preprocess, scipy.stats.skew(raw.ravel())) if args.preprocess is None or args.preprocess == "no": pass elif args.preprocess == "square": raw = np.square(raw) elif args.preprocess == "cuberoot": raw = np.cbrt(raw) logger.debug("%s after preproc %s: skew %s", sample, args.preprocess, scipy.stats.skew(raw.ravel())) return raw def pad(args, array, mode): if args.padding != 0: array = np.pad(array, ((args.padding, args.padding), (args.padding, args.padding), (args.padding, args.padding)), mode) return array def get_arguments(): parser = argparse.ArgumentParser() parser.add_argument('-i', '--in-dir', dest='in_dir', required=True, help='location of input files') parser.add_argument('-o', '--out-dir', dest='out_dir', required=True, help='where to place output files') parser.add_argument('--out-format', dest='out_format', default="hdf", help='format of output files') parser.add_argument('-p', '--parallel', default=1, type=int) parser.add_argument('--raw-gfp-min', dest='raw_gfp_min', type=int) parser.add_argument('--raw-gfp-max', dest='raw_gfp_max', type=int) parser.add_argument('--raw-bf-min', dest='raw_bf_min', type=int) parser.add_argument('--raw-bf-max', dest='raw_bf_max', type=int) parser.add_argument('--normalize', default='minmax', choices=['minmax', 'percentile', 'meanstd']) parser.add_argument('--preprocess', default='no', choices=['no', 'square', 'cuberoot']) parser.add_argument('--padding', default=0, type=int) args = parser.parse_args() os.makedirs(args.out_dir, exist_ok=True) return args def main():
def work(args, sample): logger.info("Processing %s, %s", args.in_dir, sample) out_fn = os.path.join(args.out_dir, sample) raw_fns = natsorted(glob.glob( os.path.join(args.in_dir, "BBBC010_v2_images", "*_" + sample + "_*.tif"))) # print(raw_fns) raw_gfp = load_array(raw_fns[0]).astype(np.float32) # print(raw_fns[0], np.min(raw_gfp), np.max(raw_gfp)) raw_gfp = preprocess(args, raw_gfp, sample) raw_gfp = normalize(args, raw_gfp, sample, args.raw_gfp_min, args.raw_gfp_max) raw_gfp = pad(args, raw_gfp, 'constant') raw_bf = load_array(raw_fns[1]).astype(np.float32) # print(raw_fns[1], np.min(raw_bf), np.max(raw_bf)) raw_bf = preprocess(args, raw_bf, sample) raw_bf = normalize(args, raw_bf, sample, args.raw_bf_min, args.raw_bf_max) raw_bf = pad(args, raw_bf, 'constant') files = natsorted( glob.glob(os.path.join(args.in_dir, "BBBC010_v1_foreground_eachworm", sample + "*" + "_ground_truth.png"))) # print(files) logger.info("number files: %s", len(files)) gt_labels = np.zeros((len(files),) + raw_gfp.shape, dtype=np.uint16) gt_instances = np.zeros(raw_gfp.shape, dtype=np.uint16) for idx, f in enumerate(files): gt_label = load_array(f).astype(np.uint16) gt_labels[idx, ...] = 1*(gt_label!=0) gt_instances[gt_label != 0] = idx+1 gt_numinst = np.sum(gt_labels, axis=0) tmp = np.sum(gt_labels, axis=0, keepdims=True) tmp[tmp==0] = 1 gt_labels_norm = (gt_labels.astype(np.float32) / tmp.astype(np.float32)).astype(np.float32) gt_fgbg = np.zeros(gt_labels.shape[1:], dtype=np.uint8) gt_fgbg[np.sum(gt_labels, axis=0) > 0] = 1 # is slightly displaced +1,+0 # gt_fgbg2 = load_array( # os.path.join(args.in_dir, "BBBC010_v1_foreground", # sample + "_binary.png")) # gt_fgbg2[0:-1,1:] = gt_fgbg2[1:,1:] # print(np.count_nonzero(gt_fgbg2 != gt_fgbg)) # gt_fgbg2 = pad(args, gt_fgbg2, 'constant') gt_labels = pad(args, gt_labels, 'constant') gt_instances = pad(args, gt_instances, 'constant') gt_labels_norm = pad(args, gt_labels_norm, 'constant') gt_numinst = pad(args, gt_numinst, 'constant') gt_fgbg = pad(args, gt_fgbg, 'constant') if raw_gfp.shape[0] != 1: raw_gfp = np.expand_dims(raw_gfp, 0) if raw_bf.shape[0] != 1: raw_bf = np.expand_dims(raw_bf, 0) raw = np.concatenate((raw_gfp, raw_bf), axis=0) if gt_instances.shape[0] != 1: gt_instances = np.expand_dims(gt_instances, 0) gt_numinst = np.expand_dims(gt_numinst, 0) gt_fgbg = np.expand_dims(gt_fgbg, 0) gt_labels_padded = np.pad(gt_labels, (0, 20-len(files)), 'constant') gt_labels_norm_padded = np.pad(gt_labels_norm, (0, 20-len(files)), 'constant') if args
logging.basicConfig(level='INFO') args = get_arguments() files = map( lambda fn: fn.split("/")[-1].split(".")[0].split("_")[6], glob.glob(os.path.join(args.in_dir, 'BBBC010_v2_images/*.tif'))) files = sorted(list(set(files))) print(files) if args.parallel > 1: Parallel(n_jobs=args.parallel, verbose=1, backend='multiprocessing') \ (delayed(work)(args, f) for f in files) else: for f in files: work(args, f)
identifier_body
lib.rs
_fn(name: &syn::Ident, fn_name: &str) -> syn::Ident { syn::Ident::new(&format!("{}{}_{}", SwigTag::SwigInject, fn_name, name), Span::call_site()) } fn swig_free(name: &syn::Ident) -> syn::Ident { swig_fn(name, "free") } impl ToSwig for syn::DeriveInput { fn to_swig(&self) -> String { /// Generate the SWIG wrapper code as a string. /// Basically, a class for the Rust struct `Foo` is just a wrapper /// class called `Foo` which contains a pointer to the actual Rust /// object. // prefix with tag let mut swigged = SwigTag::CodeStart.to_string(); let mut swigged_h = SwigTag::HdrStart.to_string(); let name = &self.ident; match &self.data { syn::Data::Struct(ref _ds) => { // simple wrapper definition to wrap opaque pointer. // methods get added elsewhere swigged.push_str(&format!("\ // Wrapper for Rust class {name} class {name} {{ public: ffi::{name} *self; {name}(ffi::{name} *ptr) {{ self = ptr; }}; ~{name}(){{ ffi::{free_name}(self); self = NULL; }}; ", name=name, free_name=swig_free(&name)) ); swigged_h.push_str(&format!("\ // Wrapper for Rust class {name} class {name} {{ ffi::{name} *self; public: ~{name}(); ", name=name) ); // pull out any derive implementations we want to wrap // TODO: do this in a less ad-hoc way get_derives(&self.attrs).iter().for_each(|w| { match w.as_str() { "Default" => { swigged.push_str(&format!( "{name}() {{ self = {def_name}(); }};\n", name=name, def_name=swig_fn(&name, "default") )); swigged_h.push_str(&format!("{}();\n",name)); }, _ => (), } }); swigged.push_str("};\n"); swigged_h.push_str("};\n"); }, _ => unimplemented!(), } swigged.push_str(&SwigTag::CodeEnd.to_str()); swigged_h.push_str(&SwigTag::HdrEnd.to_str()); swigged.push_str(&swigged_h); swigged } } impl<'a> ToSwig for InternalFn<'a> { fn to_swig(&self) -> String { // Generate SWIG wrapper for methods. // Main complication is making sure that namespaces are correct since // we are basically overwriting names. // Also a bit of magic to take an impl method, and add it back into // being a class method. // prefix with tag let mut swigged = SwigTag::CodeStart.to_string(); let mut swigged_h = SwigTag::HdrStart.to_string(); let name = &self.fn_def.ident; let cb_fn = cbindgen::ir::Function::load(name.to_string(), &self.fn_def.decl, true, &[], &None).unwrap(); let mut args = String::new(); let mut caller = String::new(); // Convert function arguments cb_fn.args.iter().for_each(|arg| { if args.len() > 0 { args += ", "; } if caller.len() > 0 { caller += ", "; } if arg.0 == "self" { caller += "$self->self"; } else { args += &cbindgen_write(arg); caller += &arg.0; } }); // Convert return type let mut out = cbindgen_write(&cb_fn.ret); if out == "Self" { out = self.base.clone().expect("Cannot convert `Self` return type without provided base name. Try: `#[swiggen(Foo)]` in macro").to_string(); } else if out == "String" { out = "char *".to_string() } let mut ret_out = out.clone(); // Convert function name. let name = if name.to_string() == "new" { // Custom format for new functions ret_out = "".to_string(); out = "new PKG_NAME::".to_string() + &out; self.base.clone().expect("Cannot convert `Self` return type without provided base name. Try: `#[swiggen(Foo)]` in macro").to_string() } else { name.to_string() }; // Get the mangled name exported by Rust let ext_name = swig_fn(&self.fn_def.ident, "ffi"); // The following code generates the function definitions and the header // Code needed for SWIG to generate bindings. if self.base.is_none() { swigged.push_str(&format!("\ {ret_out} {name}({args}) {{ return ({out})(ffi::{ext_name}({caller})); }}" , name=name, ext_name=ext_name, out=out, ret_out=ret_out, args=args, caller=caller)); } if let Some(base) = self.base { // Note the %extend is used by SWIG to make this a class method for // `base`. swigged_h.push_str(&format!(" %extend {base_name} {{ {ret_out} {name}({args}) {{ return ({out})(ffi::{ext_name}({caller})); }} }};\n" ,name=name, ext_name=ext_name, base_name=base, ret_out=ret_out, out=out, args=args, caller=caller)); } else { swigged_h.push_str(&format!("\ {out} {name}({args});" , name=name, out=out, args=args)); } swigged.push_str(&SwigTag::CodeEnd.to_str()); swigged_h.push_str(&SwigTag::HdrEnd.to_str()); swigged.push_str(&swigged_h); swigged } } /// Generate extern and SWIG code for a `#[derive(Swig)]` annotated item. pub fn impl_extern_it(ast: &syn::DeriveInput) -> TokenStream { let comment = ast.to_swig(); let comment = format!("#[doc=\"{}\"] #[allow(non_camel_case_types)] struct {}{};", comment, SwigTag::SwigInject, ast.ident); let doc_comment: syn::ItemStruct = syn::parse_str(&comment).expect("failed to generate SWIG code correctly"); let mut tokens: TokenStream = doc_comment.into_token_stream(); tokens.append_all(ast.as_extern().into_iter()); tokens } /// Generate extern and SWIG code for a `#[swiggen]` annotated method. pub fn impl_extern_fn(base_name: &Option<syn::Ident>, ast: &syn::ItemFn) -> TokenStream { let ifn = InternalFn { base: base_name, fn_def: ast, }; let tok = ifn.as_extern(); let comment = ifn.to_swig(); let hidden = swig_fn(&ast.ident, "hidden_ffi"); quote! { #[allow(non_snake_case)] #[doc=#comment] fn #hidden(){} #tok } } /// Write the swig code (injected via doc comments) into `swig.i`. /// This parses expanded Rust code, and writes the SWIG code to a file. pub fn gen_swig(pkg_name: &str, src: &str) { let mut tmp_file = File::create("swig.i").unwrap(); tmp_file.write_all(format!("\ %module {name} #define PKG_NAME {name} %include <std_vector.i> %include <stdint.i> %include <std_string.i> %typemap(newfree) char * \"free_string($1);\"; %{{ namespace ffi {{ #include \"bindings.h\" }} using namespace ffi; namespace {name} {{ ", name=pkg_name).as_bytes()).unwrap(); let syntax = syn::parse_file(&src).expect("Unable to parse file"); trace!("Syntax: {:#?}", syntax); let mut hdr = String::new(); // SWIG code is inside doc comments: // #[doc = "<swig code here>"] // struct __SWIG_INJECT_Foo; // // So we extract this out. syntax.items.iter().flat_map(|i| { // Extract out all of the attributes which are attached to structs/functions // starting with "__SWIG_INJECT" match i { syn::Item::Impl(ii) => { ii.items.iter().fold(Vec::new(), |mut acc, ref ii| { match ii { syn::ImplItem::Method(iim) => {
debug!("{:#?}", iim); if iim.sig.ident.to_string().starts_with(SwigTag::SwigInject.to_str()) { acc.extend_from_slice(&iim.attrs[..]);
random_line_split
lib.rs
need to do some pointer/box stuff if ac.ty.clone().into_token_stream().to_string().ends_with("str") { caller_ref.push(quote!{@str #id}); } else if let syn::Type::Reference(_) = ac.ty { caller_ref.push(quote!{@ref #id}); } else { caller_ref.push(quote!{@prim #id}); } }, _ => () } }); let base = self.base; let out = convert_ret_type(&self.fn_def.decl.output, self.base); // Similar to the above, this later calls the appropriate macro function // as to whether we need to do some pointer/box stuff let res_ref = if let syn::ReturnType::Type(_, ref ty) = self.fn_def.decl.output { if ty.clone().into_token_stream().to_string() == "String" { quote!{@str res} } else if needs_ref(&ty) { quote!{res} } else { quote!{@prim res} } } else { quote!{@prim res} }; /// Generate the function. We also inject some macro /// definitions to help with converting pointers into types and types /// into pointers. let tokens = quote! { #[allow(non_snake_case)] #[no_mangle] pub extern "C" fn #ext_name(#(#args),*) #out { #(ffi_ref!(#caller_ref);)* let res = #base::#name(#(#caller),*); box_ptr!(#res_ref) } }; tokens } } /// Helper function to define the exported/mangled names. fn swig_fn(name: &syn::Ident, fn_name: &str) -> syn::Ident { syn::Ident::new(&format!("{}{}_{}", SwigTag::SwigInject, fn_name, name), Span::call_site()) } fn swig_free(name: &syn::Ident) -> syn::Ident { swig_fn(name, "free") } impl ToSwig for syn::DeriveInput { fn to_swig(&self) -> String { /// Generate the SWIG wrapper code as a string. /// Basically, a class for the Rust struct `Foo` is just a wrapper /// class called `Foo` which contains a pointer to the actual Rust /// object. // prefix with tag let mut swigged = SwigTag::CodeStart.to_string(); let mut swigged_h = SwigTag::HdrStart.to_string(); let name = &self.ident; match &self.data { syn::Data::Struct(ref _ds) => { // simple wrapper definition to wrap opaque pointer. // methods get added elsewhere swigged.push_str(&format!("\ // Wrapper for Rust class {name} class {name} {{ public: ffi::{name} *self; {name}(ffi::{name} *ptr) {{ self = ptr; }}; ~{name}(){{ ffi::{free_name}(self); self = NULL; }}; ", name=name, free_name=swig_free(&name)) ); swigged_h.push_str(&format!("\ // Wrapper for Rust class {name} class {name} {{ ffi::{name} *self; public: ~{name}(); ", name=name) ); // pull out any derive implementations we want to wrap // TODO: do this in a less ad-hoc way get_derives(&self.attrs).iter().for_each(|w| { match w.as_str() { "Default" => { swigged.push_str(&format!( "{name}() {{ self = {def_name}(); }};\n", name=name, def_name=swig_fn(&name, "default") )); swigged_h.push_str(&format!("{}();\n",name)); }, _ => (), } }); swigged.push_str("};\n"); swigged_h.push_str("};\n"); }, _ => unimplemented!(), } swigged.push_str(&SwigTag::CodeEnd.to_str()); swigged_h.push_str(&SwigTag::HdrEnd.to_str()); swigged.push_str(&swigged_h); swigged } } impl<'a> ToSwig for InternalFn<'a> { fn to_swig(&self) -> String { // Generate SWIG wrapper for methods. // Main complication is making sure that namespaces are correct since // we are basically overwriting names. // Also a bit of magic to take an impl method, and add it back into // being a class method. // prefix with tag let mut swigged = SwigTag::CodeStart.to_string(); let mut swigged_h = SwigTag::HdrStart.to_string(); let name = &self.fn_def.ident; let cb_fn = cbindgen::ir::Function::load(name.to_string(), &self.fn_def.decl, true, &[], &None).unwrap(); let mut args = String::new(); let mut caller = String::new(); // Convert function arguments cb_fn.args.iter().for_each(|arg| { if args.len() > 0 { args += ", "; } if caller.len() > 0 { caller += ", "; } if arg.0 == "self" { caller += "$self->self"; } else { args += &cbindgen_write(arg); caller += &arg.0; } }); // Convert return type let mut out = cbindgen_write(&cb_fn.ret); if out == "Self" { out = self.base.clone().expect("Cannot convert `Self` return type without provided base name. Try: `#[swiggen(Foo)]` in macro").to_string(); } else if out == "String" { out = "char *".to_string() } let mut ret_out = out.clone(); // Convert function name. let name = if name.to_string() == "new" { // Custom format for new functions ret_out = "".to_string(); out = "new PKG_NAME::".to_string() + &out; self.base.clone().expect("Cannot convert `Self` return type without provided base name. Try: `#[swiggen(Foo)]` in macro").to_string() } else { name.to_string() }; // Get the mangled name exported by Rust let ext_name = swig_fn(&self.fn_def.ident, "ffi"); // The following code generates the function definitions and the header // Code needed for SWIG to generate bindings. if self.base.is_none() { swigged.push_str(&format!("\ {ret_out} {name}({args}) {{ return ({out})(ffi::{ext_name}({caller})); }}" , name=name, ext_name=ext_name, out=out, ret_out=ret_out, args=args, caller=caller)); } if let Some(base) = self.base { // Note the %extend is used by SWIG to make this a class method for // `base`. swigged_h.push_str(&format!(" %extend {base_name} {{ {ret_out} {name}({args}) {{ return ({out})(ffi::{ext_name}({caller})); }} }};\n" ,name=name, ext_name=ext_name, base_name=base, ret_out=ret_out, out=out, args=args, caller=caller)); } else { swigged_h.push_str(&format!("\ {out} {name}({args});" , name=name, out=out, args=args)); } swigged.push_str(&SwigTag::CodeEnd.to_str()); swigged_h.push_str(&SwigTag::HdrEnd.to_str()); swigged.push_str(&swigged_h); swigged } } /// Generate extern and SWIG code for a `#[derive(Swig)]` annotated item. pub fn impl_extern_it(ast: &syn::DeriveInput) -> TokenStream { let comment = ast.to_swig(); let comment = format!("#[doc=\"{}\"] #[allow(non_camel_case_types)] struct {}{};", comment, SwigTag::SwigInject, ast.ident); let doc_comment: syn::ItemStruct = syn::parse_str(&comment).expect("failed to generate SWIG code correctly"); let mut tokens: TokenStream = doc_comment.into_token_stream(); tokens.append_all(ast.as_extern().into_iter()); tokens } /// Generate extern and SWIG code for a `#[swiggen]` annotated method. pub fn impl_extern_fn(base_name: &Option<syn::Ident>, ast: &syn::ItemFn) -> TokenStream
{ let ifn = InternalFn { base: base_name, fn_def: ast, }; let tok = ifn.as_extern(); let comment = ifn.to_swig(); let hidden = swig_fn(&ast.ident, "hidden_ffi"); quote! { #[allow(non_snake_case)] #[doc=#comment] fn #hidden(){} #tok } }
identifier_body
lib.rs
igTag::CodeStart.to_string(); let mut swigged_h = SwigTag::HdrStart.to_string(); let name = &self.fn_def.ident; let cb_fn = cbindgen::ir::Function::load(name.to_string(), &self.fn_def.decl, true, &[], &None).unwrap(); let mut args = String::new(); let mut caller = String::new(); // Convert function arguments cb_fn.args.iter().for_each(|arg| { if args.len() > 0 { args += ", "; } if caller.len() > 0 { caller += ", "; } if arg.0 == "self" { caller += "$self->self"; } else { args += &cbindgen_write(arg); caller += &arg.0; } }); // Convert return type let mut out = cbindgen_write(&cb_fn.ret); if out == "Self" { out = self.base.clone().expect("Cannot convert `Self` return type without provided base name. Try: `#[swiggen(Foo)]` in macro").to_string(); } else if out == "String" { out = "char *".to_string() } let mut ret_out = out.clone(); // Convert function name. let name = if name.to_string() == "new" { // Custom format for new functions ret_out = "".to_string(); out = "new PKG_NAME::".to_string() + &out; self.base.clone().expect("Cannot convert `Self` return type without provided base name. Try: `#[swiggen(Foo)]` in macro").to_string() } else { name.to_string() }; // Get the mangled name exported by Rust let ext_name = swig_fn(&self.fn_def.ident, "ffi"); // The following code generates the function definitions and the header // Code needed for SWIG to generate bindings. if self.base.is_none() { swigged.push_str(&format!("\ {ret_out} {name}({args}) {{ return ({out})(ffi::{ext_name}({caller})); }}" , name=name, ext_name=ext_name, out=out, ret_out=ret_out, args=args, caller=caller)); } if let Some(base) = self.base { // Note the %extend is used by SWIG to make this a class method for // `base`. swigged_h.push_str(&format!(" %extend {base_name} {{ {ret_out} {name}({args}) {{ return ({out})(ffi::{ext_name}({caller})); }} }};\n" ,name=name, ext_name=ext_name, base_name=base, ret_out=ret_out, out=out, args=args, caller=caller)); } else { swigged_h.push_str(&format!("\ {out} {name}({args});" , name=name, out=out, args=args)); } swigged.push_str(&SwigTag::CodeEnd.to_str()); swigged_h.push_str(&SwigTag::HdrEnd.to_str()); swigged.push_str(&swigged_h); swigged } } /// Generate extern and SWIG code for a `#[derive(Swig)]` annotated item. pub fn impl_extern_it(ast: &syn::DeriveInput) -> TokenStream { let comment = ast.to_swig(); let comment = format!("#[doc=\"{}\"] #[allow(non_camel_case_types)] struct {}{};", comment, SwigTag::SwigInject, ast.ident); let doc_comment: syn::ItemStruct = syn::parse_str(&comment).expect("failed to generate SWIG code correctly"); let mut tokens: TokenStream = doc_comment.into_token_stream(); tokens.append_all(ast.as_extern().into_iter()); tokens } /// Generate extern and SWIG code for a `#[swiggen]` annotated method. pub fn impl_extern_fn(base_name: &Option<syn::Ident>, ast: &syn::ItemFn) -> TokenStream { let ifn = InternalFn { base: base_name, fn_def: ast, }; let tok = ifn.as_extern(); let comment = ifn.to_swig(); let hidden = swig_fn(&ast.ident, "hidden_ffi"); quote! { #[allow(non_snake_case)] #[doc=#comment] fn #hidden(){} #tok } } /// Write the swig code (injected via doc comments) into `swig.i`. /// This parses expanded Rust code, and writes the SWIG code to a file. pub fn gen_swig(pkg_name: &str, src: &str) { let mut tmp_file = File::create("swig.i").unwrap(); tmp_file.write_all(format!("\ %module {name} #define PKG_NAME {name} %include <std_vector.i> %include <stdint.i> %include <std_string.i> %typemap(newfree) char * \"free_string($1);\"; %{{ namespace ffi {{ #include \"bindings.h\" }} using namespace ffi; namespace {name} {{ ", name=pkg_name).as_bytes()).unwrap(); let syntax = syn::parse_file(&src).expect("Unable to parse file"); trace!("Syntax: {:#?}", syntax); let mut hdr = String::new(); // SWIG code is inside doc comments: // #[doc = "<swig code here>"] // struct __SWIG_INJECT_Foo; // // So we extract this out. syntax.items.iter().flat_map(|i| { // Extract out all of the attributes which are attached to structs/functions // starting with "__SWIG_INJECT" match i { syn::Item::Impl(ii) => { ii.items.iter().fold(Vec::new(), |mut acc, ref ii| { match ii { syn::ImplItem::Method(iim) => { debug!("{:#?}", iim); if iim.sig.ident.to_string().starts_with(SwigTag::SwigInject.to_str()) { acc.extend_from_slice(&iim.attrs[..]); } acc }, _ => Vec::new(), } }) }, syn::Item::Struct(syn::ItemStruct { attrs, ident, .. }) | syn::Item::Fn(syn::ItemFn { attrs, ident, ..}) => { if ident.to_string().starts_with(SwigTag::SwigInject.to_str()) { debug!("{:#?}", attrs); attrs.clone() } else { Vec::new() } }, _ => Vec::new() } }).for_each(|ref attr| { match attr.interpret_meta() { Some(syn::Meta::NameValue(ref mnv)) if &mnv.ident.to_string() == "doc" => { // Extract out the doc comment for these attributes if let syn::Lit::Str(ref ls) = mnv.lit { let mut swig_class = ls.value().replace("\\n", "\n"); let prefix_offset = swig_class.find(SwigTag::CodeStart.to_str()).expect("no code prefix") + SwigTag::CodeStart.len(); let suffix_offset = swig_class.find(SwigTag::CodeEnd.to_str()).expect("no code suffix"); let final_class = &swig_class[prefix_offset..suffix_offset]; let prefix_offset = swig_class.find(SwigTag::HdrStart.to_str()).expect("no header prefix") + SwigTag::HdrStart.len(); let suffix_offset = swig_class.find(SwigTag::HdrEnd.to_str()).expect("no header suffix"); let final_hdr = &swig_class[prefix_offset..suffix_offset]; tmp_file.write_all(&final_class.replace("\\n", "\n").as_bytes()).unwrap(); hdr += &final_hdr.replace("\\n", "\n"); debug!("{}", final_hdr); debug!("{}", final_class); } }, _ => () } }); tmp_file.write_all(format!("\ }} %}} namespace {name} {{ {header} }} %ignore {inject}; %include \"bindings.h\"; ", name=pkg_name, header=hdr, inject=SwigTag::SwigInject).as_bytes()).unwrap(); } /// Extract out any `derive(Foo)` attributes. fn get_derives(attrs: &[syn::Attribute]) -> Vec<String> { attrs.iter().filter_map(|a| a.interpret_meta()) .filter_map(|a| { if let syn::Meta::List(ml) = a { Some(ml) } else { None } }).filter(|ml| ml.ident.to_string() == "swig_derive") .flat_map(|ml| ml.nested) .filter_map(|nm| { if let syn::NestedMeta::Meta(m) = nm { if let syn::Meta::Word(w) = m { Some(w.to_string()) } else { None } } else { None } }).collect() } /// Parse a Rust file to extract any extern "C" functions or /// `#[swiggen]`-annotated methods and move these out of the impl block. pub fn
split_out_externs
identifier_name
park.rs
the `lazy_static!` macro. /// /// This type uses the blocking synchronization mechanism provided by the /// underlying operating system. /// /// For the API of this type alias, see the API of the generic /// [`Lazy`](crate::doc::Lazy) type. /// /// # Examples /// /// ``` /// use std::sync::Mutex; /// /// # #[cfg(feature = "std")] /// use conquer_once::Lazy; /// # #[cfg(not(feature = "std"))] /// # use conquer_once::spin::Lazy; /// /// static MUTEX: Lazy<Mutex<Vec<i32>>> = Lazy::new(Mutex::default); /// /// let mut lock = MUTEX.lock().unwrap(); /// /// lock.push(1); /// lock.push(2); /// lock.push(3); /// /// assert_eq!(lock.as_slice(), &[1, 2, 3]); /// ``` /// /// The associated [`new`](crate::lazy::Lazy::new) function can be used with any /// function or closure that implements `Fn() -> T`. /// /// ``` /// use std::collections::HashMap; /// /// # #[cfg(feature = "std")] /// use conquer_once::Lazy; /// # #[cfg(not(feature = "std"))] /// # use conquer_once::spin::Lazy; /// /// static CAPITALS: Lazy<HashMap<&str, &str>> = Lazy::new(|| { /// let mut map = HashMap::new(); /// map.insert("Norway", "Oslo"); /// map.insert("Belgium", "Brussels"); /// map.insert("Latvia", "Riga"); /// map /// }); /// /// assert_eq!(CAPITALS.get(&"Norway"), Some(&"Oslo")); /// assert_eq!(CAPITALS.get(&"Belgium"), Some(&"Brussels")); /// assert_eq!(CAPITALS.get(&"Latvia"), Some(&"Riga")); /// ``` pub type Lazy<T, F = fn() -> T> = crate::lazy::Lazy<T, ParkThread, F>; #[cfg(any(test, feature = "std"))] /// An interior mutability cell type which allows synchronized one-time /// initialization and read-only access exclusively after initialization. /// /// This type uses the blocking synchronization mechanism provided by the /// underlying operating system. /// /// For the API of this type alias, see the generic /// [`OnceCell`](crate::doc::OnceCell) type. /// /// # Examples /// /// ``` /// # #[cfg(feature = "std")] /// use conquer_once::OnceCell; /// # #[cfg(not(feature = "std"))] /// # use conquer_once::spin::OnceCell; /// /// #[derive(Copy, Clone)] /// struct Configuration { /// mode: i32, /// threshold: u64, /// msg: &'static str, /// } /// /// static CONFIG: OnceCell<Configuration> = OnceCell::uninit(); /// /// // producer thread /// CONFIG.init_once(|| Configuration { /// mode: 2, /// threshold: 128, /// msg: "..." /// }); /// /// // consumer thread /// let res = CONFIG.get().copied(); /// if let Some(config) = res { /// assert_eq!(config.mode, 2); /// assert_eq!(config.threshold, 128); /// } /// ``` pub type OnceCell<T> = crate::cell::OnceCell<T, ParkThread>; #[cfg(any(test, feature = "std"))] /// A synchronization primitive which can be used to run a one-time global /// initialization. /// /// This type uses the blocking synchronization mechanism provided by the /// underlying operating system. /// /// For the API of this type alias, see the generic /// [`OnceCell`](crate::doc::OnceCell) type. /// This is a specialization with `T = ()`. /// /// # Examples /// /// ``` /// # #[cfg(feature = "std")] /// use conquer_once::Once; /// # #[cfg(not(feature = "std"))] /// # use conquer_once::spin::Once; /// /// static mut GLOBAL: usize = 0; /// static INIT: Once = Once::uninit(); /// /// fn get_global() -> usize { /// // SAFETY: this is safe because the `Once` ensures the `static mut` is /// // assigned by only one thread and without data races. /// unsafe { /// INIT.init_once(|| { /// GLOBAL = expensive_computation(); /// }); /// # assert_eq!(GLOBAL, 1); /// GLOBAL /// } /// } /// /// fn expensive_computation() -> usize { /// // ... /// # 1 /// } /// ``` pub type Once = OnceCell<()>; mod internal { /// Blocking strategy using low-level and OS-reliant parking and un-parking /// mechanisms. #[derive(Copy, Clone, Debug, Default, Eq, Ord, PartialEq, PartialOrd)] pub struct ParkThread; } impl ParkThread { #[inline] pub(crate) fn try_block_spinning( state: &AtomicOnceState, back_off: &BackOff, ) -> Result<(), BlockedState> { loop { // (wait:1) this acquire load syncs-with the release swaps (guard:2) // and the acq-rel CAS (wait:2) match state.load(Ordering::Acquire).expect(POISON_PANIC_MSG) { Ready => return Ok(()), WouldBlock(blocked) if back_off.advise_yield() => { back_off.reset(); return Err(blocked); } _ => {} } back_off.spin(); } } } impl Unblock for ParkThread { /// Unblocks all blocked waiting threads. #[inline] unsafe fn on_unblock(state: BlockedState) { let mut curr = state.as_ptr() as *const StackWaiter; while !curr.is_null() { let thread = { // SAFETY: no mutable references to a stack waiter can exist // and the waiter struct is ensured to live while its thread is // parked, so the pointer can be safely dereferenced #[allow(unused_unsafe)] let waiter = unsafe { &*curr }; curr = waiter.next.get(); // there can be now data race when mutating the thread-cell as only the unblocking // thread will access it, the stack waiter can dropped as soon as the following // store becomes visible, so the thread MUST be taken out first let thread = waiter.thread.take().unwrap(); // (ready:2) this release store syncs-with the acquire load (ready:1) waiter.ready.store(true, Ordering::Release); thread }; thread.unpark(); } } } unsafe impl Block for ParkThread { /// Blocks (parks) the current thread until it is woken up by the thread /// with permission to initialize the `OnceCell`. #[inline] fn
(state: &AtomicOnceState) { // spin a little before parking the thread in case the state is // quickly unlocked again let back_off = BackOff::new(); let blocked = match Self::try_block_spinning(state, &back_off) { Ok(_) => return, Err(blocked) => blocked, }; // create a linked list node on the current thread's stack, which is // guaranteed to stay alive while the thread is parked. let waiter = StackWaiter { ready: AtomicBool::new(false), thread: Cell::new(Some(thread::current())), next: Cell::new(blocked.as_ptr() as *const StackWaiter), }; let mut curr = blocked; let head = BlockedState::from(&waiter as *const _); // SAFETY: `head` is a valid pointer to a `StackWaiter` that will live // for the duration of this function, which in turn will only return // when no other thread can still observe any pointer to it // (wait:2) this acq-rel CAS syncs-with itself and the acq load (wait:1) while let Err(err) = unsafe { state.try_enqueue_waiter(curr, head, Ordering::AcqRel) } { match err { // another parked thread succeeded in placing itself at the queue's front WouldBlock(queue) => { // the waiter hasn't been shared yet, so it's still safe to // mutate the next pointer curr = queue; waiter.next.set(queue.as_ptr() as *const StackWaiter); back_off.spin(); } // acquire-release is required here to enforce acquire ordering in the failure case, // which guarantees that any (non-atomic) stores to the cell's inner state preceding // (guard:2) have become visible, if the function returns; // (alternatively an explicit acquire fence could be placed into this path) Ready => return, Uninit => unreachable!("cell state can not become `UNINIT again`"), } } // park the thread until it is woken up by the thread that first set the state to blocked. // the loop guards against spurious wake ups // (ready:1) this acquire load syncs-with the release store (ready:2) while !waiter.ready.load(Ordering::Acquire) { thread::park(); } // SAFETY: propagates poisoning as required by the trait // (wait:3) this acquire load syncs-with the acq-rel swap (guard:2) assert_eq!(state.load(Ordering::Acquire).expect(
block
identifier_name
park.rs
the `lazy_static!` macro. /// /// This type uses the blocking synchronization mechanism provided by the /// underlying operating system. /// /// For the API of this type alias, see the API of the generic /// [`Lazy`](crate::doc::Lazy) type. /// /// # Examples /// /// ``` /// use std::sync::Mutex; /// /// # #[cfg(feature = "std")] /// use conquer_once::Lazy; /// # #[cfg(not(feature = "std"))] /// # use conquer_once::spin::Lazy; /// /// static MUTEX: Lazy<Mutex<Vec<i32>>> = Lazy::new(Mutex::default); /// /// let mut lock = MUTEX.lock().unwrap(); /// /// lock.push(1); /// lock.push(2); /// lock.push(3); /// /// assert_eq!(lock.as_slice(), &[1, 2, 3]); /// ``` /// /// The associated [`new`](crate::lazy::Lazy::new) function can be used with any /// function or closure that implements `Fn() -> T`. /// /// ``` /// use std::collections::HashMap; /// /// # #[cfg(feature = "std")] /// use conquer_once::Lazy; /// # #[cfg(not(feature = "std"))] /// # use conquer_once::spin::Lazy; /// /// static CAPITALS: Lazy<HashMap<&str, &str>> = Lazy::new(|| { /// let mut map = HashMap::new(); /// map.insert("Norway", "Oslo"); /// map.insert("Belgium", "Brussels"); /// map.insert("Latvia", "Riga"); /// map /// }); /// /// assert_eq!(CAPITALS.get(&"Norway"), Some(&"Oslo")); /// assert_eq!(CAPITALS.get(&"Belgium"), Some(&"Brussels")); /// assert_eq!(CAPITALS.get(&"Latvia"), Some(&"Riga")); /// ``` pub type Lazy<T, F = fn() -> T> = crate::lazy::Lazy<T, ParkThread, F>; #[cfg(any(test, feature = "std"))] /// An interior mutability cell type which allows synchronized one-time /// initialization and read-only access exclusively after initialization. /// /// This type uses the blocking synchronization mechanism provided by the /// underlying operating system. /// /// For the API of this type alias, see the generic /// [`OnceCell`](crate::doc::OnceCell) type. /// /// # Examples /// /// ``` /// # #[cfg(feature = "std")] /// use conquer_once::OnceCell; /// # #[cfg(not(feature = "std"))] /// # use conquer_once::spin::OnceCell; /// /// #[derive(Copy, Clone)] /// struct Configuration { /// mode: i32, /// threshold: u64, /// msg: &'static str, /// } /// /// static CONFIG: OnceCell<Configuration> = OnceCell::uninit(); /// /// // producer thread /// CONFIG.init_once(|| Configuration { /// mode: 2, /// threshold: 128, /// msg: "..." /// }); /// /// // consumer thread /// let res = CONFIG.get().copied(); /// if let Some(config) = res { /// assert_eq!(config.mode, 2); /// assert_eq!(config.threshold, 128); /// } /// ``` pub type OnceCell<T> = crate::cell::OnceCell<T, ParkThread>; #[cfg(any(test, feature = "std"))] /// A synchronization primitive which can be used to run a one-time global /// initialization. /// /// This type uses the blocking synchronization mechanism provided by the /// underlying operating system. /// /// For the API of this type alias, see the generic /// [`OnceCell`](crate::doc::OnceCell) type. /// This is a specialization with `T = ()`. /// /// # Examples /// /// ``` /// # #[cfg(feature = "std")] /// use conquer_once::Once; /// # #[cfg(not(feature = "std"))] /// # use conquer_once::spin::Once; /// /// static mut GLOBAL: usize = 0; /// static INIT: Once = Once::uninit(); /// /// fn get_global() -> usize { /// // SAFETY: this is safe because the `Once` ensures the `static mut` is /// // assigned by only one thread and without data races. /// unsafe { /// INIT.init_once(|| { /// GLOBAL = expensive_computation(); /// }); /// # assert_eq!(GLOBAL, 1); /// GLOBAL /// } /// } /// /// fn expensive_computation() -> usize { /// // ... /// # 1 /// } /// ``` pub type Once = OnceCell<()>; mod internal { /// Blocking strategy using low-level and OS-reliant parking and un-parking /// mechanisms. #[derive(Copy, Clone, Debug, Default, Eq, Ord, PartialEq, PartialOrd)] pub struct ParkThread; } impl ParkThread { #[inline] pub(crate) fn try_block_spinning( state: &AtomicOnceState, back_off: &BackOff, ) -> Result<(), BlockedState> { loop { // (wait:1) this acquire load syncs-with the release swaps (guard:2) // and the acq-rel CAS (wait:2) match state.load(Ordering::Acquire).expect(POISON_PANIC_MSG) { Ready => return Ok(()), WouldBlock(blocked) if back_off.advise_yield() => { back_off.reset(); return Err(blocked); } _ => {} } back_off.spin(); } } } impl Unblock for ParkThread { /// Unblocks all blocked waiting threads. #[inline] unsafe fn on_unblock(state: BlockedState) { let mut curr = state.as_ptr() as *const StackWaiter; while !curr.is_null() { let thread = { // SAFETY: no mutable references to a stack waiter can exist // and the waiter struct is ensured to live while its thread is // parked, so the pointer can be safely dereferenced #[allow(unused_unsafe)] let waiter = unsafe { &*curr }; curr = waiter.next.get(); // there can be now data race when mutating the thread-cell as only the unblocking // thread will access it, the stack waiter can dropped as soon as the following // store becomes visible, so the thread MUST be taken out first let thread = waiter.thread.take().unwrap(); // (ready:2) this release store syncs-with the acquire load (ready:1) waiter.ready.store(true, Ordering::Release); thread }; thread.unpark(); } } } unsafe impl Block for ParkThread { /// Blocks (parks) the current thread until it is woken up by the thread /// with permission to initialize the `OnceCell`. #[inline] fn block(state: &AtomicOnceState)
// SAFETY: `head` is a valid pointer to a `StackWaiter` that will live // for the duration of this function, which in turn will only return // when no other thread can still observe any pointer to it // (wait:2) this acq-rel CAS syncs-with itself and the acq load (wait:1) while let Err(err) = unsafe { state.try_enqueue_waiter(curr, head, Ordering::AcqRel) } { match err { // another parked thread succeeded in placing itself at the queue's front WouldBlock(queue) => { // the waiter hasn't been shared yet, so it's still safe to // mutate the next pointer curr = queue; waiter.next.set(queue.as_ptr() as *const StackWaiter); back_off.spin(); } // acquire-release is required here to enforce acquire ordering in the failure case, // which guarantees that any (non-atomic) stores to the cell's inner state preceding // (guard:2) have become visible, if the function returns; // (alternatively an explicit acquire fence could be placed into this path) Ready => return, Uninit => unreachable!("cell state can not become `UNINIT again`"), } } // park the thread until it is woken up by the thread that first set the state to blocked. // the loop guards against spurious wake ups // (ready:1) this acquire load syncs-with the release store (ready:2) while !waiter.ready.load(Ordering::Acquire) { thread::park(); } // SAFETY: propagates poisoning as required by the trait // (wait:3) this acquire load syncs-with the acq-rel swap (guard:2) assert_eq!(state.load(Ordering::Acquire).expect(
{ // spin a little before parking the thread in case the state is // quickly unlocked again let back_off = BackOff::new(); let blocked = match Self::try_block_spinning(state, &back_off) { Ok(_) => return, Err(blocked) => blocked, }; // create a linked list node on the current thread's stack, which is // guaranteed to stay alive while the thread is parked. let waiter = StackWaiter { ready: AtomicBool::new(false), thread: Cell::new(Some(thread::current())), next: Cell::new(blocked.as_ptr() as *const StackWaiter), }; let mut curr = blocked; let head = BlockedState::from(&waiter as *const _);
identifier_body
park.rs
the `lazy_static!` macro. /// /// This type uses the blocking synchronization mechanism provided by the /// underlying operating system. /// /// For the API of this type alias, see the API of the generic /// [`Lazy`](crate::doc::Lazy) type. /// /// # Examples /// /// ``` /// use std::sync::Mutex; /// /// # #[cfg(feature = "std")] /// use conquer_once::Lazy; /// # #[cfg(not(feature = "std"))] /// # use conquer_once::spin::Lazy; /// /// static MUTEX: Lazy<Mutex<Vec<i32>>> = Lazy::new(Mutex::default); /// /// let mut lock = MUTEX.lock().unwrap(); /// /// lock.push(1); /// lock.push(2); /// lock.push(3); /// /// assert_eq!(lock.as_slice(), &[1, 2, 3]); /// ``` /// /// The associated [`new`](crate::lazy::Lazy::new) function can be used with any /// function or closure that implements `Fn() -> T`. /// /// ``` /// use std::collections::HashMap; /// /// # #[cfg(feature = "std")] /// use conquer_once::Lazy; /// # #[cfg(not(feature = "std"))] /// # use conquer_once::spin::Lazy; /// /// static CAPITALS: Lazy<HashMap<&str, &str>> = Lazy::new(|| { /// let mut map = HashMap::new(); /// map.insert("Norway", "Oslo"); /// map.insert("Belgium", "Brussels"); /// map.insert("Latvia", "Riga"); /// map /// }); /// /// assert_eq!(CAPITALS.get(&"Norway"), Some(&"Oslo")); /// assert_eq!(CAPITALS.get(&"Belgium"), Some(&"Brussels")); /// assert_eq!(CAPITALS.get(&"Latvia"), Some(&"Riga")); /// ``` pub type Lazy<T, F = fn() -> T> = crate::lazy::Lazy<T, ParkThread, F>; #[cfg(any(test, feature = "std"))] /// An interior mutability cell type which allows synchronized one-time /// initialization and read-only access exclusively after initialization. /// /// This type uses the blocking synchronization mechanism provided by the /// underlying operating system. /// /// For the API of this type alias, see the generic /// [`OnceCell`](crate::doc::OnceCell) type. /// /// # Examples /// /// ``` /// # #[cfg(feature = "std")] /// use conquer_once::OnceCell; /// # #[cfg(not(feature = "std"))] /// # use conquer_once::spin::OnceCell; /// /// #[derive(Copy, Clone)] /// struct Configuration { /// mode: i32, /// threshold: u64, /// msg: &'static str, /// } /// /// static CONFIG: OnceCell<Configuration> = OnceCell::uninit(); /// /// // producer thread /// CONFIG.init_once(|| Configuration { /// mode: 2, /// threshold: 128, /// msg: "..." /// }); /// /// // consumer thread /// let res = CONFIG.get().copied(); /// if let Some(config) = res { /// assert_eq!(config.mode, 2); /// assert_eq!(config.threshold, 128); /// } /// ``` pub type OnceCell<T> = crate::cell::OnceCell<T, ParkThread>; #[cfg(any(test, feature = "std"))] /// A synchronization primitive which can be used to run a one-time global /// initialization. /// /// This type uses the blocking synchronization mechanism provided by the /// underlying operating system. /// /// For the API of this type alias, see the generic /// [`OnceCell`](crate::doc::OnceCell) type. /// This is a specialization with `T = ()`. /// /// # Examples /// /// ``` /// # #[cfg(feature = "std")] /// use conquer_once::Once; /// # #[cfg(not(feature = "std"))] /// # use conquer_once::spin::Once; /// /// static mut GLOBAL: usize = 0; /// static INIT: Once = Once::uninit(); /// /// fn get_global() -> usize { /// // SAFETY: this is safe because the `Once` ensures the `static mut` is /// // assigned by only one thread and without data races. /// unsafe { /// INIT.init_once(|| { /// GLOBAL = expensive_computation(); /// }); /// # assert_eq!(GLOBAL, 1); /// GLOBAL /// } /// } /// /// fn expensive_computation() -> usize { /// // ... /// # 1 /// } /// ``` pub type Once = OnceCell<()>; mod internal { /// Blocking strategy using low-level and OS-reliant parking and un-parking /// mechanisms. #[derive(Copy, Clone, Debug, Default, Eq, Ord, PartialEq, PartialOrd)] pub struct ParkThread; } impl ParkThread { #[inline] pub(crate) fn try_block_spinning( state: &AtomicOnceState, back_off: &BackOff, ) -> Result<(), BlockedState> { loop { // (wait:1) this acquire load syncs-with the release swaps (guard:2) // and the acq-rel CAS (wait:2) match state.load(Ordering::Acquire).expect(POISON_PANIC_MSG) { Ready => return Ok(()), WouldBlock(blocked) if back_off.advise_yield() => { back_off.reset(); return Err(blocked); } _ => {} }
} } } impl Unblock for ParkThread { /// Unblocks all blocked waiting threads. #[inline] unsafe fn on_unblock(state: BlockedState) { let mut curr = state.as_ptr() as *const StackWaiter; while !curr.is_null() { let thread = { // SAFETY: no mutable references to a stack waiter can exist // and the waiter struct is ensured to live while its thread is // parked, so the pointer can be safely dereferenced #[allow(unused_unsafe)] let waiter = unsafe { &*curr }; curr = waiter.next.get(); // there can be now data race when mutating the thread-cell as only the unblocking // thread will access it, the stack waiter can dropped as soon as the following // store becomes visible, so the thread MUST be taken out first let thread = waiter.thread.take().unwrap(); // (ready:2) this release store syncs-with the acquire load (ready:1) waiter.ready.store(true, Ordering::Release); thread }; thread.unpark(); } } } unsafe impl Block for ParkThread { /// Blocks (parks) the current thread until it is woken up by the thread /// with permission to initialize the `OnceCell`. #[inline] fn block(state: &AtomicOnceState) { // spin a little before parking the thread in case the state is // quickly unlocked again let back_off = BackOff::new(); let blocked = match Self::try_block_spinning(state, &back_off) { Ok(_) => return, Err(blocked) => blocked, }; // create a linked list node on the current thread's stack, which is // guaranteed to stay alive while the thread is parked. let waiter = StackWaiter { ready: AtomicBool::new(false), thread: Cell::new(Some(thread::current())), next: Cell::new(blocked.as_ptr() as *const StackWaiter), }; let mut curr = blocked; let head = BlockedState::from(&waiter as *const _); // SAFETY: `head` is a valid pointer to a `StackWaiter` that will live // for the duration of this function, which in turn will only return // when no other thread can still observe any pointer to it // (wait:2) this acq-rel CAS syncs-with itself and the acq load (wait:1) while let Err(err) = unsafe { state.try_enqueue_waiter(curr, head, Ordering::AcqRel) } { match err { // another parked thread succeeded in placing itself at the queue's front WouldBlock(queue) => { // the waiter hasn't been shared yet, so it's still safe to // mutate the next pointer curr = queue; waiter.next.set(queue.as_ptr() as *const StackWaiter); back_off.spin(); } // acquire-release is required here to enforce acquire ordering in the failure case, // which guarantees that any (non-atomic) stores to the cell's inner state preceding // (guard:2) have become visible, if the function returns; // (alternatively an explicit acquire fence could be placed into this path) Ready => return, Uninit => unreachable!("cell state can not become `UNINIT again`"), } } // park the thread until it is woken up by the thread that first set the state to blocked. // the loop guards against spurious wake ups // (ready:1) this acquire load syncs-with the release store (ready:2) while !waiter.ready.load(Ordering::Acquire) { thread::park(); } // SAFETY: propagates poisoning as required by the trait // (wait:3) this acquire load syncs-with the acq-rel swap (guard:2) assert_eq!(state.load(Ordering::Acquire).expect(PO
back_off.spin();
random_line_split
zapis.py
(2,), wiec odwoluje sie bezposrednio do zmiennej ustawienia if ustawienia.shape == (2,): parametry[re.sub('"','',ustawienia[0])] = ustawienia[1] # jak mamy wiecej parametrow odwoluje sie do kolejnych linijek macierzy # ustawienia else: for l in ustawienia: parametry[re.sub('"','',l[0])] = l[1] # zamieniamy napisy na odpowiednie wartosci - kontroluje te parametry, wiec # robie to recznie try: parametry['tryb'] = parametry['tryb'].strip() #tryb # jak nie podano danego parametru to idz dalej, nie wyrzucaj bledu except KeyError: print("Podaj tryb odczytu!") try: parametry['bpm'] = int(parametry['bpm']) # tempo # jak nie podano danego parametru to idz dalej, nie wyrzucaj bledu except KeyError: pass try: parametry['freq'] = int(parametry['freq']) # frekwencja wyjsciowego wav # jak nie podano danego parametru to idz dalej, nie wyrzucaj bledu except KeyError: pass try: parametry['loud'] = float(parametry['loud'] ) # glosnosc # jak nie podano danego parametru to idz dalej, nie wyrzucaj bledu except KeyError: pass try: # lista wag dla sampli parametry['wages'] = [float(s) for s in parametry['wages'].split(",")] # jak nie podano danego parametru to idz dalej, nie wyrzucaj bledu except KeyError: pass return parametry #b = wczytywanie_ustawien("defs.txt") #zglasnianie utworu def z
utwor, procent = 0): """ zmienia glosnosc utworu (jego amplitudy) arg: numpy.ndarray (numpy.int16): utwor - dzwiek, ktory ma byc zglosniony lub zciszony float: procent - liczba obrazujaca zmiane glosnosci utworu, osiaga wartosci od -1 do 1, dla 0 brak zmian, dla 1 - "100% glosniej", dla -1 "100% ciszej" wyjscie: numpy.ndarray (numpy.int16): glosniejszy -sciszony lub zglosniony utwor """ if(-1 <= procent <= 1): #ile razy mamy pomnozyc amplitude naszego dzwieku mnoznik = 0 if( procent < 0 ): mnoznik = 1 + procent else: # obliczamy najwyzsza amplitude w danym utworze i ona bedzie # wyznaczac jak bardzo mozemy podglosnic maks_ampli = 0 maks_ampli = max(abs(utwor)) mnoznik = 32767/maks_ampli # maksymalny mnoznik # mnoznik minimalnie moze osiagnac wartosc 1, to co powyzej # (mnoznik-1) mnozymy o procent zglosnienia # i dodajemy do podstawy (czyli 1) mnoznik = 1 + (mnoznik - 1)*procent glosniej = mnoznik * utwor #glosniej = np.array(glosniej, dtype=np.int16) glosniej = glosniej.astype(np.int16) return glosniej else: print("Podaj procent z zakresu -1 do 1") #wierszyk1 = zmiana_glosnosci(wierszyk, b['loud']) #wierszyk1 def tworzenie_piosenki(macierz_piosenki, czy_pelna = True, bpm = 120, \ freq = 44100, wages = None, loud = 0): """ glowna funkcja generujaca cala piosenke arg: numpy.ndarray (str: U2): macierz_piosenki - macierz zawierajaca definicje kolejnych cwiercnut (co ma byc grane w danej cwiercnucie) bool: czy_pelna - zmienna sprawdzajaca czy macierz_piosenki jest zapisana (nie jest, gdy tracki mialy nieodpowiednia liczbe wierszy lub kolumn) int: bpm - tempo piosenki w jednostce bpm int: freq - ilosc probek w jednej sekundzie list (float): wages - wagi kolejnych sampli (jakie znaczenie ma miec 1 probka, 2 etc.) float: loud - procent glosnosci, 0 - tak jak oryginalne probki, 1 - na maxa, -1 - sciszamy na maxa wyjscie: numpy.ndarray (numpy.int16): gotowy utwór """ # macierz piosenki byla pusta, piosenka nie zostala utworzona if(czy_pelna == False): print("Nie utworzono piosenki") return None else: import numpy as np import scipy.io.wavfile t_cwiercnuty = 60 / bpm # czas trwania jednej cwiercnuty (zalezy od #tempa) ile_cwiercnut = macierz_piosenki.shape[0] # ilosc cwiercnut kanaly = macierz_piosenki.shape[1] # ilosc uzywanych sampli frekw = freq czas_utworu = ile_cwiercnut*t_cwiercnuty # ile elementow bedzie w nowym utworze ilosc_probek = int(frekw*czas_utworu) # bedziemy tylko raz wczytywac zawartosc sampleXY.wav, wiec potrzebuje # unikalne numery sampli rozne_sample = np.unique(macierz_piosenki) # bierze lacznie z "--" # w slownikach zapiszemy parametry tych sampli # slownik z wartosciami danego sampla (tj. macierze numpy-owe z # amplitudami) sample_co = {} sample_frekw = {} # slownik z ich frekwencjami sample_dl = {} # slownik z ich dlugosciami #wczytujemy te sample # w iteratorze bierzemy napisy "01" "02" "--" itd. stringi!!! for ktory_sampel in rozne_sample: if(ktory_sampel != '--'): # tworzymy napis z nazwa pliku sampla, np. "sample01.wav" plik = ''.join(['sample',ktory_sampel,'.wav']) # wczytujemy zawartosc i frekwencje danego sampla do # odpowiednio nazwanego elementu w slowniku sample_co i # sample_frekw sample_frekw[ktory_sampel], sample_co[ktory_sampel] = \ scipy.io.wavfile.read(plik) # tworzymy mono z naszego sampla sample_co[ktory_sampel] = np.mean(sample_co[ktory_sampel],\ axis=1)/32767 # normalizujemy te wartosci sample_co[ktory_sampel] = np.int16(sample_co[ktory_sampel]/ \ max(np.abs(sample_co[ktory_sampel])) * 32767) # zapisujemy dlugosc sampli, czyli ilosc probek # ( = czas_trwania*frekwencja) sample_dl[ktory_sampel] = sample_co[ktory_sampel].shape[0] else: # to samo robimy dla "--" recznie ustawiamy #
miana_glosnosci(
identifier_name
zapis.py
znik = 1 + procent else: # obliczamy najwyzsza amplitude w danym utworze i ona bedzie # wyznaczac jak bardzo mozemy podglosnic maks_ampli = 0 maks_ampli = max(abs(utwor)) mnoznik = 32767/maks_ampli # maksymalny mnoznik # mnoznik minimalnie moze osiagnac wartosc 1, to co powyzej # (mnoznik-1) mnozymy o procent zglosnienia # i dodajemy do podstawy (czyli 1) mnoznik = 1 + (mnoznik - 1)*procent glosniej = mnoznik * utwor #glosniej = np.array(glosniej, dtype=np.int16) glosniej = glosniej.astype(np.int16) return glosniej else: print("Podaj procent z zakresu -1 do 1") #wierszyk1 = zmiana_glosnosci(wierszyk, b['loud']) #wierszyk1 def tworzenie_piosenki(macierz_piosenki, czy_pelna = True, bpm = 120, \ freq = 44100, wages = None, loud = 0): """ glowna funkcja generujaca cala piosenke arg: numpy.ndarray (str: U2): macierz_piosenki - macierz zawierajaca definicje kolejnych cwiercnut (co ma byc grane w danej cwiercnucie) bool: czy_pelna - zmienna sprawdzajaca czy macierz_piosenki jest zapisana (nie jest, gdy tracki mialy nieodpowiednia liczbe wierszy lub kolumn) int: bpm - tempo piosenki w jednostce bpm int: freq - ilosc probek w jednej sekundzie list (float): wages - wagi kolejnych sampli (jakie znaczenie ma miec 1 probka, 2 etc.) float: loud - procent glosnosci, 0 - tak jak oryginalne probki, 1 - na maxa, -1 - sciszamy na maxa wyjscie: numpy.ndarray (numpy.int16): gotowy utwór """ # macierz piosenki byla pusta, piosenka nie zostala utworzona if(czy_pelna == False): print("Nie utworzono piosenki") return None else: import numpy as np import scipy.io.wavfile t_cwiercnuty = 60 / bpm # czas trwania jednej cwiercnuty (zalezy od #tempa) ile_cwiercnut = macierz_piosenki.shape[0] # ilosc cwiercnut kanaly = macierz_piosenki.shape[1] # ilosc uzywanych sampli frekw = freq czas_utworu = ile_cwiercnut*t_cwiercnuty # ile elementow bedzie w nowym utworze ilosc_probek = int(frekw*czas_utworu) # bedziemy tylko raz wczytywac zawartosc sampleXY.wav, wiec potrzebuje # unikalne numery sampli rozne_sample = np.unique(macierz_piosenki) # bierze lacznie z "--" # w slownikach zapiszemy parametry tych sampli # slownik z wartosciami danego sampla (tj. macierze numpy-owe z # amplitudami) sample_co = {} sample_frekw = {} # slownik z ich frekwencjami sample_dl = {} # slownik z ich dlugosciami #wczytujemy te sample # w iteratorze bierzemy napisy "01" "02" "--" itd. stringi!!! for ktory_sampel in rozne_sample: if(ktory_sampel != '--'): # tworzymy napis z nazwa pliku sampla, np. "sample01.wav" plik = ''.join(['sample',ktory_sampel,'.wav']) # wczytujemy zawartosc i frekwencje danego sampla do # odpowiednio nazwanego elementu w slowniku sample_co i # sample_frekw sample_frekw[ktory_sampel], sample_co[ktory_sampel] = \ scipy.io.wavfile.read(plik) # tworzymy mono z naszego sampla sample_co[ktory_sampel] = np.mean(sample_co[ktory_sampel],\ axis=1)/32767 # normalizujemy te wartosci sample_co[ktory_sampel] = np.int16(sample_co[ktory_sampel]/ \ max(np.abs(sample_co[ktory_sampel])) * 32767) # zapisujemy dlugosc sampli, czyli ilosc probek # ( = czas_trwania*frekwencja) sample_dl[ktory_sampel] = sample_co[ktory_sampel].shape[0] else: # to samo robimy dla "--" recznie ustawiamy # robimy cisze, gdy -- sample_co[ktory_sampel] = np.zeros((1,), dtype=np.int16) sample_frekw[ktory_sampel] = frekw # taka sama jak domyslna sample_dl[ktory_sampel] = 0 # zakladamy czas 0 sekund if wages is None: wages = np.ones((1,kanaly)) else: # zeby mialo wymiar (1,kanaly), a nie (kanaly,) wages = np.array(wages).reshape(1,kanaly) # definicja nowego utworu T = np.linspace(0, czas_utworu, ilosc_probek) for wiersz in range(0, ile_cwiercnut): sample = [] # wczytamy sample z danej cwiecnuty dlugosci = [] # tu zapiszemy ich dlugosci w tej cwiercnucie for i in range(0, kanaly): sampus = macierz_piosenki[wiersz,i] sample.append(sample_co[sampus]) dlugosci.append(sample_dl[sampus]) # bierzemy najdluzszy sample i w calosci bedziemy go odtwarzac; # reszte zatem tez w calosci odtworzymy, a gdy sie skoncza damy # cisze (zera) maksik = max(dlugosci) # mamy tutaj macierz 4 na max dlugosc, przygotowana do zlaczenia # potem tych dzwiekow w jeden pusty = np.int16(np.zeros((len(sample), maksik))) # dodajemy nasze dzwieki do tej pustej for k in range(0, kanaly): pusty[k][0:dlugosci[k]] = sample[k] # mnozymy kolejne elementy wektora pusty (czyli sample) przez # wagi i sumujemy cwiercnuta = np.dot(wages, pusty) #otrzymamy wymiar (1, x), a chcemy (x,), wiec bierzemy pierwszy # element cwiercnuta = cwiercnuta[0] # poczatek biezacej cwiercnuty poczatek_cwiercnuty = int(wiersz*t_cwiercnuty*frekw) # jesli dodanie ostatnich cwiercnut bedzie wiazalo sie z # przekroczeniem dlugosci tworzonego utworu, obcinamy ostatnie # dzwieki, tak by zmiescic sie w tej dlugosci if (poczatek_cwiercnuty + maksik) > ilosc_probek: T[
poczatek_cwiercnuty:(poczatek_cwiercnuty + maksik)]=\ cwiercnuta[0:len(T[poczatek_cwiercnuty:(poczatek_cwiercnuty +\ maksik)])]
conditional_block
zapis.py
noznik-1) mnozymy o procent zglosnienia # i dodajemy do podstawy (czyli 1) mnoznik = 1 + (mnoznik - 1)*procent glosniej = mnoznik * utwor #glosniej = np.array(glosniej, dtype=np.int16) glosniej = glosniej.astype(np.int16) return glosniej else: print("Podaj procent z zakresu -1 do 1") #wierszyk1 = zmiana_glosnosci(wierszyk, b['loud']) #wierszyk1 def tworzenie_piosenki(macierz_piosenki, czy_pelna = True, bpm = 120, \ freq = 44100, wages = None, loud = 0): """ glowna funkcja generujaca cala piosenke arg: numpy.ndarray (str: U2): macierz_piosenki - macierz zawierajaca definicje kolejnych cwiercnut (co ma byc grane w danej cwiercnucie) bool: czy_pelna - zmienna sprawdzajaca czy macierz_piosenki jest zapisana (nie jest, gdy tracki mialy nieodpowiednia liczbe wierszy lub kolumn) int: bpm - tempo piosenki w jednostce bpm int: freq - ilosc probek w jednej sekundzie list (float): wages - wagi kolejnych sampli (jakie znaczenie ma miec 1 probka, 2 etc.) float: loud - procent glosnosci, 0 - tak jak oryginalne probki, 1 - na maxa, -1 - sciszamy na maxa wyjscie: numpy.ndarray (numpy.int16): gotowy utwór """ # macierz piosenki byla pusta, piosenka nie zostala utworzona if(czy_pelna == False): print("Nie utworzono piosenki") return None else: import numpy as np import scipy.io.wavfile t_cwiercnuty = 60 / bpm # czas trwania jednej cwiercnuty (zalezy od #tempa) ile_cwiercnut = macierz_piosenki.shape[0] # ilosc cwiercnut kanaly = macierz_piosenki.shape[1] # ilosc uzywanych sampli frekw = freq czas_utworu = ile_cwiercnut*t_cwiercnuty # ile elementow bedzie w nowym utworze ilosc_probek = int(frekw*czas_utworu) # bedziemy tylko raz wczytywac zawartosc sampleXY.wav, wiec potrzebuje # unikalne numery sampli rozne_sample = np.unique(macierz_piosenki) # bierze lacznie z "--" # w slownikach zapiszemy parametry tych sampli # slownik z wartosciami danego sampla (tj. macierze numpy-owe z # amplitudami) sample_co = {} sample_frekw = {} # slownik z ich frekwencjami sample_dl = {} # slownik z ich dlugosciami #wczytujemy te sample # w iteratorze bierzemy napisy "01" "02" "--" itd. stringi!!! for ktory_sampel in rozne_sample: if(ktory_sampel != '--'): # tworzymy napis z nazwa pliku sampla, np. "sample01.wav" plik = ''.join(['sample',ktory_sampel,'.wav']) # wczytujemy zawartosc i frekwencje danego sampla do # odpowiednio nazwanego elementu w slowniku sample_co i # sample_frekw sample_frekw[ktory_sampel], sample_co[ktory_sampel] = \ scipy.io.wavfile.read(plik) # tworzymy mono z naszego sampla sample_co[ktory_sampel] = np.mean(sample_co[ktory_sampel],\ axis=1)/32767 # normalizujemy te wartosci sample_co[ktory_sampel] = np.int16(sample_co[ktory_sampel]/ \ max(np.abs(sample_co[ktory_sampel])) * 32767) # zapisujemy dlugosc sampli, czyli ilosc probek # ( = czas_trwania*frekwencja) sample_dl[ktory_sampel] = sample_co[ktory_sampel].shape[0] else: # to samo robimy dla "--" recznie ustawiamy # robimy cisze, gdy -- sample_co[ktory_sampel] = np.zeros((1,), dtype=np.int16) sample_frekw[ktory_sampel] = frekw # taka sama jak domyslna sample_dl[ktory_sampel] = 0 # zakladamy czas 0 sekund if wages is None: wages = np.ones((1,kanaly)) else: # zeby mialo wymiar (1,kanaly), a nie (kanaly,) wages = np.array(wages).reshape(1,kanaly) # definicja nowego utworu T = np.linspace(0, czas_utworu, ilosc_probek) for wiersz in range(0, ile_cwiercnut): sample = [] # wczytamy sample z danej cwiecnuty dlugosci = [] # tu zapiszemy ich dlugosci w tej cwiercnucie for i in range(0, kanaly): sampus = macierz_piosenki[wiersz,i] sample.append(sample_co[sampus]) dlugosci.append(sample_dl[sampus]) # bierzemy najdluzszy sample i w calosci bedziemy go odtwarzac; # reszte zatem tez w calosci odtworzymy, a gdy sie skoncza damy # cisze (zera) maksik = max(dlugosci) # mamy tutaj macierz 4 na max dlugosc, przygotowana do zlaczenia # potem tych dzwiekow w jeden pusty = np.int16(np.zeros((len(sample), maksik))) # dodajemy nasze dzwieki do tej pustej for k in range(0, kanaly): pusty[k][0:dlugosci[k]] = sample[k] # mnozymy kolejne elementy wektora pusty (czyli sample) przez # wagi i sumujemy cwiercnuta = np.dot(wages, pusty) #otrzymamy wymiar (1, x), a chcemy (x,), wiec bierzemy pierwszy # element cwiercnuta = cwiercnuta[0] # poczatek biezacej cwiercnuty poczatek_cwiercnuty = int(wiersz*t_cwiercnuty*frekw) # jesli dodanie ostatnich cwiercnut bedzie wiazalo sie z # przekroczeniem dlugosci tworzonego utworu, obcinamy ostatnie # dzwieki, tak by zmiescic sie w tej dlugosci if (poczatek_cwiercnuty + maksik) > ilosc_probek: T[poczatek_cwiercnuty:(poczatek_cwiercnuty + maksik)]=\ cwiercnuta[0:len(T[poczatek_cwiercnuty:(poczatek_cwiercnuty +\ maksik)])] else: T[poczatek_cwiercnuty:(poczatek_cwiercnuty + maksik)] += \ cwiercnuta T= np.array(T, dtype=np.int16) #ustalamy glosnosc utworu T = zmiana_glosnosci(T, loud) return T
#pios, k = wczytywanie_sciezek(a) #wierszyk = tworzenie_piosenki(pios, k, bpm = b['bpm'], freq = b['freq'], \
random_line_split
zapis.py
2,), wiec odwoluje sie bezposrednio do zmiennej ustawienia if ustawienia.shape == (2,): parametry[re.sub('"','',ustawienia[0])] = ustawienia[1] # jak mamy wiecej parametrow odwoluje sie do kolejnych linijek macierzy # ustawienia else: for l in ustawienia: parametry[re.sub('"','',l[0])] = l[1] # zamieniamy napisy na odpowiednie wartosci - kontroluje te parametry, wiec # robie to recznie try: parametry['tryb'] = parametry['tryb'].strip() #tryb # jak nie podano danego parametru to idz dalej, nie wyrzucaj bledu except KeyError: print("Podaj tryb odczytu!") try: parametry['bpm'] = int(parametry['bpm']) # tempo # jak nie podano danego parametru to idz dalej, nie wyrzucaj bledu except KeyError: pass try: parametry['freq'] = int(parametry['freq']) # frekwencja wyjsciowego wav # jak nie podano danego parametru to idz dalej, nie wyrzucaj bledu except KeyError: pass try: parametry['loud'] = float(parametry['loud'] ) # glosnosc # jak nie podano danego parametru to idz dalej, nie wyrzucaj bledu except KeyError: pass try: # lista wag dla sampli parametry['wages'] = [float(s) for s in parametry['wages'].split(",")] # jak nie podano danego parametru to idz dalej, nie wyrzucaj bledu except KeyError: pass return parametry #b = wczytywanie_ustawien("defs.txt") #zglasnianie utworu def zmiana_glosnosci(utwor, procent = 0): """ zmienia glosnosc utworu (jego amplitudy) arg: numpy.ndarray (numpy.int16): utwor - dzwiek, ktory ma byc zglosniony lub zciszony float: procent - liczba obrazujaca zmiane glosnosci utworu, osiaga wartosci od -1 do 1, dla 0 brak zmian, dla 1 - "100% glosniej", dla -1 "100% ciszej" wyjscie: numpy.ndarray (numpy.int16): glosniejszy -sciszony lub zglosniony utwor """ if(-1 <= procent <= 1): #ile razy mamy pomnozyc amplitude naszego dzwieku mnoznik = 0 if( procent < 0 ): mnoznik = 1 + procent else: # obliczamy najwyzsza amplitude w danym utworze i ona bedzie # wyznaczac jak bardzo mozemy podglosnic maks_ampli = 0 maks_ampli = max(abs(utwor)) mnoznik = 32767/maks_ampli # maksymalny mnoznik # mnoznik minimalnie moze osiagnac wartosc 1, to co powyzej # (mnoznik-1) mnozymy o procent zglosnienia # i dodajemy do podstawy (czyli 1) mnoznik = 1 + (mnoznik - 1)*procent glosniej = mnoznik * utwor #glosniej = np.array(glosniej, dtype=np.int16) glosniej = glosniej.astype(np.int16) return glosniej else: print("Podaj procent z zakresu -1 do 1") #wierszyk1 = zmiana_glosnosci(wierszyk, b['loud']) #wierszyk1 def tworzenie_piosenki(macierz_piosenki, czy_pelna = True, bpm = 120, \ freq = 44100, wages = None, loud = 0): "
maxa, -1 - sciszamy na maxa wyjscie: numpy.ndarray (numpy.int16): gotowy utwór """ # macierz piosenki byla pusta, piosenka nie zostala utworzona if(czy_pelna == False): print("Nie utworzono piosenki") return None else: import numpy as np import scipy.io.wavfile t_cwiercnuty = 60 / bpm # czas trwania jednej cwiercnuty (zalezy od #tempa) ile_cwiercnut = macierz_piosenki.shape[0] # ilosc cwiercnut kanaly = macierz_piosenki.shape[1] # ilosc uzywanych sampli frekw = freq czas_utworu = ile_cwiercnut*t_cwiercnuty # ile elementow bedzie w nowym utworze ilosc_probek = int(frekw*czas_utworu) # bedziemy tylko raz wczytywac zawartosc sampleXY.wav, wiec potrzebuje # unikalne numery sampli rozne_sample = np.unique(macierz_piosenki) # bierze lacznie z "--" # w slownikach zapiszemy parametry tych sampli # slownik z wartosciami danego sampla (tj. macierze numpy-owe z # amplitudami) sample_co = {} sample_frekw = {} # slownik z ich frekwencjami sample_dl = {} # slownik z ich dlugosciami #wczytujemy te sample # w iteratorze bierzemy napisy "01" "02" "--" itd. stringi!!! for ktory_sampel in rozne_sample: if(ktory_sampel != '--'): # tworzymy napis z nazwa pliku sampla, np. "sample01.wav" plik = ''.join(['sample',ktory_sampel,'.wav']) # wczytujemy zawartosc i frekwencje danego sampla do # odpowiednio nazwanego elementu w slowniku sample_co i # sample_frekw sample_frekw[ktory_sampel], sample_co[ktory_sampel] = \ scipy.io.wavfile.read(plik) # tworzymy mono z naszego sampla sample_co[ktory_sampel] = np.mean(sample_co[ktory_sampel],\ axis=1)/32767 # normalizujemy te wartosci sample_co[ktory_sampel] = np.int16(sample_co[ktory_sampel]/ \ max(np.abs(sample_co[ktory_sampel])) * 32767) # zapisujemy dlugosc sampli, czyli ilosc probek # ( = czas_trwania*frekwencja) sample_dl[ktory_sampel] = sample_co[ktory_sampel].shape[0] else: # to samo robimy dla "--" recznie ustawiamy
"" glowna funkcja generujaca cala piosenke arg: numpy.ndarray (str: U2): macierz_piosenki - macierz zawierajaca definicje kolejnych cwiercnut (co ma byc grane w danej cwiercnucie) bool: czy_pelna - zmienna sprawdzajaca czy macierz_piosenki jest zapisana (nie jest, gdy tracki mialy nieodpowiednia liczbe wierszy lub kolumn) int: bpm - tempo piosenki w jednostce bpm int: freq - ilosc probek w jednej sekundzie list (float): wages - wagi kolejnych sampli (jakie znaczenie ma miec 1 probka, 2 etc.) float: loud - procent glosnosci, 0 - tak jak oryginalne probki, 1 - na
identifier_body
proof.rs
the `Proof` pub lemma: Lemma, /// The value concerned by this `Proof` pub value: T, } #[cfg(feature = "serialization-serde")] mod algorithm_serde { use ring::digest::{self, Algorithm}; use serde::de::Error; use serde::{Deserialize, Deserializer, Serialize, Serializer}; pub fn serialize<S: Serializer>( algorithm: &'static Algorithm, se: S, ) -> Result<S::Ok, S::Error> { // The `Debug` implementation of `Algorithm` prints its ID. format!("{:?}", algorithm).serialize(se) } pub fn deserialize<'de, D: Deserializer<'de>>(de: D) -> Result<&'static Algorithm, D::Error> { let alg_str: String = Deserialize::deserialize(de)?; match &*alg_str { "SHA1" => Ok(&digest::SHA1_FOR_LEGACY_USE_ONLY), "SHA256" => Ok(&digest::SHA256), "SHA384" => Ok(&digest::SHA384), "SHA512" => Ok(&digest::SHA512), "SHA512_256" => Ok(&digest::SHA512_256), _ => Err(D::Error::custom("unknown hash algorithm")), } } #[cfg(test)] mod test { use super::*; use ring::digest::{ SHA1_FOR_LEGACY_USE_ONLY as sha1, SHA256 as sha256, SHA384 as sha384, SHA512 as sha512, SHA512_256 as sha512_256, }; static SHA1: &Algorithm = &sha1; static SHA256: &Algorithm = &sha256; static SHA384: &Algorithm = &sha384; static SHA512: &Algorithm = &sha512; static SHA512_256: &Algorithm = &sha512_256; #[test] fn test_serialize_known_algorithms() { extern crate serde_json; for alg in &[SHA1, SHA256, SHA384, SHA512, SHA512_256] { let mut serializer = serde_json::Serializer::with_formatter( vec![], serde_json::ser::PrettyFormatter::new(), ); serialize(alg, &mut serializer).unwrap_or_else(|_| panic!("{:?}", alg)); let alg_ = deserialize(&mut serde_json::Deserializer::from_slice( &serializer.into_inner()[..], )) .unwrap_or_else(|_| panic!("{:?}", alg)); assert_eq!(*alg, alg_); } } #[test] #[should_panic(expected = "unknown hash algorithm")] fn test_serialize_unknown_algorithm() { extern crate serde_json; { let alg_str = "\"BLAKE2b\""; let mut deserializer = serde_json::Deserializer::from_str(alg_str); let _ = deserialize(&mut deserializer) .unwrap_or_else(|_| panic!("unknown hash algorithm {:?}", alg_str)); } } } } impl<T: PartialEq> PartialEq for Proof<T> { fn eq(&self, other: &Proof<T>) -> bool { self.root_hash == other.root_hash && self.lemma == other.lemma && self.value == other.value } } impl<T: Eq> Eq for Proof<T> {} impl<T: Ord> PartialOrd for Proof<T> { fn partial_cmp(&self, other: &Proof<T>) -> Option<Ordering>
} impl<T: Ord> Ord for Proof<T> { fn cmp(&self, other: &Proof<T>) -> Ordering { self.root_hash .cmp(&other.root_hash) .then(self.value.cmp(&other.value)) .then_with(|| self.lemma.cmp(&other.lemma)) } } impl<T: Hash> Hash for Proof<T> { fn hash<H: Hasher>(&self, state: &mut H) { self.root_hash.hash(state); self.lemma.hash(state); self.value.hash(state); } } impl<T> Proof<T> { /// Constructs a new `Proof` pub fn new(algorithm: &'static Algorithm, root_hash: Vec<u8>, lemma: Lemma, value: T) -> Self { Proof { algorithm, root_hash, lemma, value, } } /// Checks whether this inclusion proof is well-formed, /// and whether its root hash matches the given `root_hash`. pub fn validate(&self, root_hash: &[u8]) -> bool { if self.root_hash != root_hash || self.lemma.node_hash != root_hash { return false; } self.lemma.validate(self.algorithm) } /// Returns the index of this proof's value, given the total number of items in the tree. /// /// # Panics /// /// Panics if the proof is malformed. Call `validate` first. pub fn index(&self, count: usize) -> usize { self.lemma.index(count) } } /// A `Lemma` holds the hash of a node, the hash of its sibling node, /// and a sub lemma, whose `node_hash`, when combined with this `sibling_hash` /// must be equal to this `node_hash`. #[cfg_attr(feature = "serialization-serde", derive(Serialize, Deserialize))] #[derive(Clone, Debug, PartialEq, Eq, PartialOrd, Ord, Hash)] pub struct Lemma { pub node_hash: Vec<u8>, pub sibling_hash: Option<Positioned<Vec<u8>>>, pub sub_lemma: Option<Box<Lemma>>, } impl Lemma { /// Attempts to generate a proof that the a value with hash `needle` is a /// member of the given `tree`. pub fn new<T>(tree: &Tree<T>, needle: &[u8]) -> Option<Lemma> { match *tree { Tree::Empty { .. } => None, Tree::Leaf { ref hash, .. } => Lemma::new_leaf_proof(hash, needle), Tree::Node { ref hash, ref left, ref right, } => Lemma::new_tree_proof(hash, needle, left, right), } } /// Attempts to generate a proof that the `idx`-th leaf is a member of /// the given tree. The `count` must equal the number of leaves in the /// `tree`. If `idx >= count`, `None` is returned. Otherwise it returns /// the new `Lemma` and the `idx`-th value. pub fn new_by_index<T>(tree: &Tree<T>, idx: usize, count: usize) -> Option<(Lemma, &T)> { if idx >= count { return None; } match *tree { Tree::Empty { .. } => None, Tree::Leaf { ref hash, ref value, .. } => { if count != 1 { return None; } let lemma = Lemma { node_hash: hash.clone(), sibling_hash: None, sub_lemma: None, }; Some((lemma, value)) } Tree::Node { ref hash, ref left, ref right, } => { let left_count = count.next_power_of_two() / 2; let (sub_lem_val, sibling_hash); if idx < left_count { sub_lem_val = Lemma::new_by_index(left, idx, left_count); sibling_hash = Positioned::Right(right.hash().clone()); } else { sub_lem_val = Lemma::new_by_index(right, idx - left_count, count - left_count); sibling_hash = Positioned::Left(left.hash().clone()); } sub_lem_val.map(|(sub_lemma, value)| { let lemma = Lemma { node_hash: hash.clone(), sibling_hash: Some(sibling_hash), sub_lemma: Some(Box::new(sub_lemma)), }; (lemma, value) }) } } } /// Returns the index of this lemma's value, given the total number of items in the tree. /// /// # Panics /// /// Panics if the lemma is malformed. Call `validate_lemma` first. pub fn index(&self, count: usize) -> usize { let left_count = count.next_power_of_two() / 2; match (self.sub_lemma.as_ref(), self.sibling_hash.as_ref()) { (None, None) => 0, (Some(l), Some(&Positioned::Left(_))) => left_count + l.index(count - left_count), (Some(l), Some(&Positioned::Right(_))) => l.index(left_count), (None, Some(_)) | (Some(_), None) => panic!("malformed lemma"), } } fn new_leaf_proof(hash: &[u8], needle: &[u8]) -> Option<Lemma> { if *hash == *needle { Some(Lemma { node_hash: hash.into(), sibling_hash: None, sub_lemma: None, }) } else { None } } fn new_tree_proof<T>( hash: &[u8], needle: &[u8],
{ Some(self.cmp(other)) }
identifier_body
proof.rs
the `Proof` pub lemma: Lemma, /// The value concerned by this `Proof` pub value: T, } #[cfg(feature = "serialization-serde")] mod algorithm_serde { use ring::digest::{self, Algorithm}; use serde::de::Error; use serde::{Deserialize, Deserializer, Serialize, Serializer}; pub fn serialize<S: Serializer>( algorithm: &'static Algorithm, se: S, ) -> Result<S::Ok, S::Error> { // The `Debug` implementation of `Algorithm` prints its ID. format!("{:?}", algorithm).serialize(se) } pub fn deserialize<'de, D: Deserializer<'de>>(de: D) -> Result<&'static Algorithm, D::Error> { let alg_str: String = Deserialize::deserialize(de)?; match &*alg_str { "SHA1" => Ok(&digest::SHA1_FOR_LEGACY_USE_ONLY), "SHA256" => Ok(&digest::SHA256), "SHA384" => Ok(&digest::SHA384), "SHA512" => Ok(&digest::SHA512), "SHA512_256" => Ok(&digest::SHA512_256), _ => Err(D::Error::custom("unknown hash algorithm")), } } #[cfg(test)] mod test { use super::*; use ring::digest::{ SHA1_FOR_LEGACY_USE_ONLY as sha1, SHA256 as sha256, SHA384 as sha384, SHA512 as sha512, SHA512_256 as sha512_256, }; static SHA1: &Algorithm = &sha1; static SHA256: &Algorithm = &sha256; static SHA384: &Algorithm = &sha384; static SHA512: &Algorithm = &sha512; static SHA512_256: &Algorithm = &sha512_256; #[test] fn test_serialize_known_algorithms() { extern crate serde_json; for alg in &[SHA1, SHA256, SHA384, SHA512, SHA512_256] { let mut serializer = serde_json::Serializer::with_formatter( vec![], serde_json::ser::PrettyFormatter::new(), ); serialize(alg, &mut serializer).unwrap_or_else(|_| panic!("{:?}", alg)); let alg_ = deserialize(&mut serde_json::Deserializer::from_slice( &serializer.into_inner()[..], )) .unwrap_or_else(|_| panic!("{:?}", alg)); assert_eq!(*alg, alg_); } } #[test] #[should_panic(expected = "unknown hash algorithm")] fn test_serialize_unknown_algorithm() { extern crate serde_json; { let alg_str = "\"BLAKE2b\""; let mut deserializer = serde_json::Deserializer::from_str(alg_str); let _ = deserialize(&mut deserializer) .unwrap_or_else(|_| panic!("unknown hash algorithm {:?}", alg_str)); } } } } impl<T: PartialEq> PartialEq for Proof<T> { fn eq(&self, other: &Proof<T>) -> bool { self.root_hash == other.root_hash && self.lemma == other.lemma && self.value == other.value } } impl<T: Eq> Eq for Proof<T> {} impl<T: Ord> PartialOrd for Proof<T> { fn partial_cmp(&self, other: &Proof<T>) -> Option<Ordering> { Some(self.cmp(other)) } } impl<T: Ord> Ord for Proof<T> { fn cmp(&self, other: &Proof<T>) -> Ordering { self.root_hash .cmp(&other.root_hash) .then(self.value.cmp(&other.value)) .then_with(|| self.lemma.cmp(&other.lemma)) } } impl<T: Hash> Hash for Proof<T> { fn hash<H: Hasher>(&self, state: &mut H) { self.root_hash.hash(state); self.lemma.hash(state); self.value.hash(state); } } impl<T> Proof<T> { /// Constructs a new `Proof` pub fn new(algorithm: &'static Algorithm, root_hash: Vec<u8>, lemma: Lemma, value: T) -> Self { Proof { algorithm, root_hash, lemma, value, } } /// Checks whether this inclusion proof is well-formed, /// and whether its root hash matches the given `root_hash`. pub fn validate(&self, root_hash: &[u8]) -> bool { if self.root_hash != root_hash || self.lemma.node_hash != root_hash { return false; } self.lemma.validate(self.algorithm) } /// Returns the index of this proof's value, given the total number of items in the tree. /// /// # Panics /// /// Panics if the proof is malformed. Call `validate` first. pub fn index(&self, count: usize) -> usize { self.lemma.index(count) } } /// A `Lemma` holds the hash of a node, the hash of its sibling node, /// and a sub lemma, whose `node_hash`, when combined with this `sibling_hash` /// must be equal to this `node_hash`. #[cfg_attr(feature = "serialization-serde", derive(Serialize, Deserialize))] #[derive(Clone, Debug, PartialEq, Eq, PartialOrd, Ord, Hash)] pub struct Lemma { pub node_hash: Vec<u8>, pub sibling_hash: Option<Positioned<Vec<u8>>>, pub sub_lemma: Option<Box<Lemma>>, } impl Lemma { /// Attempts to generate a proof that the a value with hash `needle` is a /// member of the given `tree`. pub fn new<T>(tree: &Tree<T>, needle: &[u8]) -> Option<Lemma> { match *tree { Tree::Empty { .. } => None, Tree::Leaf { ref hash, .. } => Lemma::new_leaf_proof(hash, needle), Tree::Node { ref hash, ref left, ref right, } => Lemma::new_tree_proof(hash, needle, left, right), } } /// Attempts to generate a proof that the `idx`-th leaf is a member of /// the given tree. The `count` must equal the number of leaves in the /// `tree`. If `idx >= count`, `None` is returned. Otherwise it returns /// the new `Lemma` and the `idx`-th value. pub fn new_by_index<T>(tree: &Tree<T>, idx: usize, count: usize) -> Option<(Lemma, &T)> { if idx >= count { return None; } match *tree { Tree::Empty { .. } => None, Tree::Leaf { ref hash, ref value, .. } => { if count != 1 { return None; } let lemma = Lemma { node_hash: hash.clone(), sibling_hash: None, sub_lemma: None, }; Some((lemma, value)) } Tree::Node { ref hash, ref left, ref right, } => { let left_count = count.next_power_of_two() / 2; let (sub_lem_val, sibling_hash); if idx < left_count { sub_lem_val = Lemma::new_by_index(left, idx, left_count); sibling_hash = Positioned::Right(right.hash().clone()); } else { sub_lem_val = Lemma::new_by_index(right, idx - left_count, count - left_count); sibling_hash = Positioned::Left(left.hash().clone()); } sub_lem_val.map(|(sub_lemma, value)| { let lemma = Lemma { node_hash: hash.clone(), sibling_hash: Some(sibling_hash), sub_lemma: Some(Box::new(sub_lemma)), }; (lemma, value) }) } } } /// Returns the index of this lemma's value, given the total number of items in the tree. /// /// # Panics /// /// Panics if the lemma is malformed. Call `validate_lemma` first. pub fn index(&self, count: usize) -> usize { let left_count = count.next_power_of_two() / 2; match (self.sub_lemma.as_ref(), self.sibling_hash.as_ref()) { (None, None) => 0, (Some(l), Some(&Positioned::Left(_))) => left_count + l.index(count - left_count), (Some(l), Some(&Positioned::Right(_))) => l.index(left_count), (None, Some(_)) | (Some(_), None) => panic!("malformed lemma"), } } fn new_leaf_proof(hash: &[u8], needle: &[u8]) -> Option<Lemma> { if *hash == *needle
else { None } } fn new_tree_proof<T>( hash: &[u8], needle: &[u8],
{ Some(Lemma { node_hash: hash.into(), sibling_hash: None, sub_lemma: None, }) }
conditional_block
proof.rs
the `Proof` pub lemma: Lemma, /// The value concerned by this `Proof` pub value: T, } #[cfg(feature = "serialization-serde")] mod algorithm_serde { use ring::digest::{self, Algorithm}; use serde::de::Error; use serde::{Deserialize, Deserializer, Serialize, Serializer}; pub fn serialize<S: Serializer>( algorithm: &'static Algorithm, se: S, ) -> Result<S::Ok, S::Error> { // The `Debug` implementation of `Algorithm` prints its ID. format!("{:?}", algorithm).serialize(se) } pub fn deserialize<'de, D: Deserializer<'de>>(de: D) -> Result<&'static Algorithm, D::Error> { let alg_str: String = Deserialize::deserialize(de)?; match &*alg_str { "SHA1" => Ok(&digest::SHA1_FOR_LEGACY_USE_ONLY), "SHA256" => Ok(&digest::SHA256), "SHA384" => Ok(&digest::SHA384), "SHA512" => Ok(&digest::SHA512), "SHA512_256" => Ok(&digest::SHA512_256), _ => Err(D::Error::custom("unknown hash algorithm")), } } #[cfg(test)] mod test { use super::*; use ring::digest::{ SHA1_FOR_LEGACY_USE_ONLY as sha1, SHA256 as sha256, SHA384 as sha384, SHA512 as sha512, SHA512_256 as sha512_256, }; static SHA1: &Algorithm = &sha1; static SHA256: &Algorithm = &sha256; static SHA384: &Algorithm = &sha384; static SHA512: &Algorithm = &sha512; static SHA512_256: &Algorithm = &sha512_256; #[test] fn test_serialize_known_algorithms() { extern crate serde_json; for alg in &[SHA1, SHA256, SHA384, SHA512, SHA512_256] { let mut serializer = serde_json::Serializer::with_formatter( vec![], serde_json::ser::PrettyFormatter::new(), ); serialize(alg, &mut serializer).unwrap_or_else(|_| panic!("{:?}", alg)); let alg_ = deserialize(&mut serde_json::Deserializer::from_slice( &serializer.into_inner()[..], )) .unwrap_or_else(|_| panic!("{:?}", alg)); assert_eq!(*alg, alg_); } } #[test] #[should_panic(expected = "unknown hash algorithm")] fn test_serialize_unknown_algorithm() { extern crate serde_json; { let alg_str = "\"BLAKE2b\""; let mut deserializer = serde_json::Deserializer::from_str(alg_str); let _ = deserialize(&mut deserializer) .unwrap_or_else(|_| panic!("unknown hash algorithm {:?}", alg_str)); } } } } impl<T: PartialEq> PartialEq for Proof<T> { fn eq(&self, other: &Proof<T>) -> bool { self.root_hash == other.root_hash && self.lemma == other.lemma && self.value == other.value } } impl<T: Eq> Eq for Proof<T> {} impl<T: Ord> PartialOrd for Proof<T> { fn partial_cmp(&self, other: &Proof<T>) -> Option<Ordering> { Some(self.cmp(other)) } } impl<T: Ord> Ord for Proof<T> { fn cmp(&self, other: &Proof<T>) -> Ordering { self.root_hash .cmp(&other.root_hash) .then(self.value.cmp(&other.value)) .then_with(|| self.lemma.cmp(&other.lemma)) } } impl<T: Hash> Hash for Proof<T> { fn hash<H: Hasher>(&self, state: &mut H) { self.root_hash.hash(state); self.lemma.hash(state); self.value.hash(state); } } impl<T> Proof<T> { /// Constructs a new `Proof` pub fn new(algorithm: &'static Algorithm, root_hash: Vec<u8>, lemma: Lemma, value: T) -> Self { Proof { algorithm, root_hash, lemma, value, } } /// Checks whether this inclusion proof is well-formed, /// and whether its root hash matches the given `root_hash`. pub fn validate(&self, root_hash: &[u8]) -> bool { if self.root_hash != root_hash || self.lemma.node_hash != root_hash { return false; } self.lemma.validate(self.algorithm) } /// Returns the index of this proof's value, given the total number of items in the tree. /// /// # Panics /// /// Panics if the proof is malformed. Call `validate` first. pub fn index(&self, count: usize) -> usize { self.lemma.index(count) } } /// A `Lemma` holds the hash of a node, the hash of its sibling node, /// and a sub lemma, whose `node_hash`, when combined with this `sibling_hash` /// must be equal to this `node_hash`. #[cfg_attr(feature = "serialization-serde", derive(Serialize, Deserialize))] #[derive(Clone, Debug, PartialEq, Eq, PartialOrd, Ord, Hash)] pub struct Lemma { pub node_hash: Vec<u8>, pub sibling_hash: Option<Positioned<Vec<u8>>>, pub sub_lemma: Option<Box<Lemma>>, } impl Lemma { /// Attempts to generate a proof that the a value with hash `needle` is a /// member of the given `tree`. pub fn new<T>(tree: &Tree<T>, needle: &[u8]) -> Option<Lemma> { match *tree { Tree::Empty { .. } => None, Tree::Leaf { ref hash, .. } => Lemma::new_leaf_proof(hash, needle), Tree::Node { ref hash, ref left, ref right, } => Lemma::new_tree_proof(hash, needle, left, right), } } /// Attempts to generate a proof that the `idx`-th leaf is a member of /// the given tree. The `count` must equal the number of leaves in the /// `tree`. If `idx >= count`, `None` is returned. Otherwise it returns /// the new `Lemma` and the `idx`-th value. pub fn new_by_index<T>(tree: &Tree<T>, idx: usize, count: usize) -> Option<(Lemma, &T)> { if idx >= count { return None; } match *tree { Tree::Empty { .. } => None, Tree::Leaf { ref hash, ref value, .. } => { if count != 1 { return None; } let lemma = Lemma { node_hash: hash.clone(), sibling_hash: None, sub_lemma: None, }; Some((lemma, value)) } Tree::Node { ref hash, ref left, ref right, } => { let left_count = count.next_power_of_two() / 2; let (sub_lem_val, sibling_hash); if idx < left_count { sub_lem_val = Lemma::new_by_index(left, idx, left_count); sibling_hash = Positioned::Right(right.hash().clone()); } else { sub_lem_val = Lemma::new_by_index(right, idx - left_count, count - left_count); sibling_hash = Positioned::Left(left.hash().clone()); } sub_lem_val.map(|(sub_lemma, value)| { let lemma = Lemma { node_hash: hash.clone(), sibling_hash: Some(sibling_hash), sub_lemma: Some(Box::new(sub_lemma)), }; (lemma, value) }) } } } /// Returns the index of this lemma's value, given the total number of items in the tree. /// /// # Panics /// /// Panics if the lemma is malformed. Call `validate_lemma` first. pub fn index(&self, count: usize) -> usize { let left_count = count.next_power_of_two() / 2; match (self.sub_lemma.as_ref(), self.sibling_hash.as_ref()) { (None, None) => 0, (Some(l), Some(&Positioned::Left(_))) => left_count + l.index(count - left_count), (Some(l), Some(&Positioned::Right(_))) => l.index(left_count), (None, Some(_)) | (Some(_), None) => panic!("malformed lemma"), } } fn new_leaf_proof(hash: &[u8], needle: &[u8]) -> Option<Lemma> { if *hash == *needle { Some(Lemma { node_hash: hash.into(), sibling_hash: None, sub_lemma: None, }) } else { None } } fn
<T>( hash: &[u8], needle: &[u8],
new_tree_proof
identifier_name
proof.rs
of the `Proof` pub lemma: Lemma, /// The value concerned by this `Proof` pub value: T, } #[cfg(feature = "serialization-serde")] mod algorithm_serde { use ring::digest::{self, Algorithm}; use serde::de::Error; use serde::{Deserialize, Deserializer, Serialize, Serializer}; pub fn serialize<S: Serializer>( algorithm: &'static Algorithm, se: S, ) -> Result<S::Ok, S::Error> { // The `Debug` implementation of `Algorithm` prints its ID. format!("{:?}", algorithm).serialize(se) } pub fn deserialize<'de, D: Deserializer<'de>>(de: D) -> Result<&'static Algorithm, D::Error> { let alg_str: String = Deserialize::deserialize(de)?; match &*alg_str { "SHA1" => Ok(&digest::SHA1_FOR_LEGACY_USE_ONLY), "SHA256" => Ok(&digest::SHA256), "SHA384" => Ok(&digest::SHA384), "SHA512" => Ok(&digest::SHA512), "SHA512_256" => Ok(&digest::SHA512_256), _ => Err(D::Error::custom("unknown hash algorithm")), } } #[cfg(test)] mod test { use super::*; use ring::digest::{ SHA1_FOR_LEGACY_USE_ONLY as sha1, SHA256 as sha256, SHA384 as sha384, SHA512 as sha512, SHA512_256 as sha512_256, }; static SHA1: &Algorithm = &sha1; static SHA256: &Algorithm = &sha256; static SHA384: &Algorithm = &sha384; static SHA512: &Algorithm = &sha512; static SHA512_256: &Algorithm = &sha512_256; #[test] fn test_serialize_known_algorithms() { extern crate serde_json; for alg in &[SHA1, SHA256, SHA384, SHA512, SHA512_256] { let mut serializer = serde_json::Serializer::with_formatter( vec![], serde_json::ser::PrettyFormatter::new(), ); serialize(alg, &mut serializer).unwrap_or_else(|_| panic!("{:?}", alg)); let alg_ = deserialize(&mut serde_json::Deserializer::from_slice( &serializer.into_inner()[..], )) .unwrap_or_else(|_| panic!("{:?}", alg)); assert_eq!(*alg, alg_); } } #[test] #[should_panic(expected = "unknown hash algorithm")] fn test_serialize_unknown_algorithm() { extern crate serde_json; { let alg_str = "\"BLAKE2b\""; let mut deserializer = serde_json::Deserializer::from_str(alg_str); let _ = deserialize(&mut deserializer) .unwrap_or_else(|_| panic!("unknown hash algorithm {:?}", alg_str)); } } } } impl<T: PartialEq> PartialEq for Proof<T> { fn eq(&self, other: &Proof<T>) -> bool { self.root_hash == other.root_hash && self.lemma == other.lemma && self.value == other.value } } impl<T: Eq> Eq for Proof<T> {} impl<T: Ord> PartialOrd for Proof<T> { fn partial_cmp(&self, other: &Proof<T>) -> Option<Ordering> { Some(self.cmp(other)) } } impl<T: Ord> Ord for Proof<T> { fn cmp(&self, other: &Proof<T>) -> Ordering { self.root_hash .cmp(&other.root_hash) .then(self.value.cmp(&other.value)) .then_with(|| self.lemma.cmp(&other.lemma)) } } impl<T: Hash> Hash for Proof<T> { fn hash<H: Hasher>(&self, state: &mut H) { self.root_hash.hash(state); self.lemma.hash(state); self.value.hash(state); } } impl<T> Proof<T> { /// Constructs a new `Proof` pub fn new(algorithm: &'static Algorithm, root_hash: Vec<u8>, lemma: Lemma, value: T) -> Self { Proof { algorithm, root_hash, lemma, value, } } /// Checks whether this inclusion proof is well-formed, /// and whether its root hash matches the given `root_hash`. pub fn validate(&self, root_hash: &[u8]) -> bool { if self.root_hash != root_hash || self.lemma.node_hash != root_hash { return false; } self.lemma.validate(self.algorithm) } /// Returns the index of this proof's value, given the total number of items in the tree. /// /// # Panics /// /// Panics if the proof is malformed. Call `validate` first. pub fn index(&self, count: usize) -> usize { self.lemma.index(count) } } /// A `Lemma` holds the hash of a node, the hash of its sibling node, /// and a sub lemma, whose `node_hash`, when combined with this `sibling_hash` /// must be equal to this `node_hash`. #[cfg_attr(feature = "serialization-serde", derive(Serialize, Deserialize))] #[derive(Clone, Debug, PartialEq, Eq, PartialOrd, Ord, Hash)] pub struct Lemma { pub node_hash: Vec<u8>, pub sibling_hash: Option<Positioned<Vec<u8>>>, pub sub_lemma: Option<Box<Lemma>>, } impl Lemma { /// Attempts to generate a proof that the a value with hash `needle` is a /// member of the given `tree`. pub fn new<T>(tree: &Tree<T>, needle: &[u8]) -> Option<Lemma> { match *tree { Tree::Empty { .. } => None, Tree::Leaf { ref hash, .. } => Lemma::new_leaf_proof(hash, needle), Tree::Node { ref hash, ref left, ref right, } => Lemma::new_tree_proof(hash, needle, left, right), } } /// Attempts to generate a proof that the `idx`-th leaf is a member of /// the given tree. The `count` must equal the number of leaves in the /// `tree`. If `idx >= count`, `None` is returned. Otherwise it returns /// the new `Lemma` and the `idx`-th value. pub fn new_by_index<T>(tree: &Tree<T>, idx: usize, count: usize) -> Option<(Lemma, &T)> { if idx >= count { return None; } match *tree { Tree::Empty { .. } => None, Tree::Leaf {
ref hash, ref value, .. } => { if count != 1 { return None; } let lemma = Lemma { node_hash: hash.clone(), sibling_hash: None, sub_lemma: None, }; Some((lemma, value)) } Tree::Node { ref hash, ref left, ref right, } => { let left_count = count.next_power_of_two() / 2; let (sub_lem_val, sibling_hash); if idx < left_count { sub_lem_val = Lemma::new_by_index(left, idx, left_count); sibling_hash = Positioned::Right(right.hash().clone()); } else { sub_lem_val = Lemma::new_by_index(right, idx - left_count, count - left_count); sibling_hash = Positioned::Left(left.hash().clone()); } sub_lem_val.map(|(sub_lemma, value)| { let lemma = Lemma { node_hash: hash.clone(), sibling_hash: Some(sibling_hash), sub_lemma: Some(Box::new(sub_lemma)), }; (lemma, value) }) } } } /// Returns the index of this lemma's value, given the total number of items in the tree. /// /// # Panics /// /// Panics if the lemma is malformed. Call `validate_lemma` first. pub fn index(&self, count: usize) -> usize { let left_count = count.next_power_of_two() / 2; match (self.sub_lemma.as_ref(), self.sibling_hash.as_ref()) { (None, None) => 0, (Some(l), Some(&Positioned::Left(_))) => left_count + l.index(count - left_count), (Some(l), Some(&Positioned::Right(_))) => l.index(left_count), (None, Some(_)) | (Some(_), None) => panic!("malformed lemma"), } } fn new_leaf_proof(hash: &[u8], needle: &[u8]) -> Option<Lemma> { if *hash == *needle { Some(Lemma { node_hash: hash.into(), sibling_hash: None, sub_lemma: None, }) } else { None } } fn new_tree_proof<T>( hash: &[u8], needle: &[u8],
random_line_split
main.rs
Session}, }; fn run( display_index: usize, output_path: &str, bit_rate: u32, frame_rate: u32, resolution: Resolution, encoder_index: usize, verbose: bool, wait_for_debugger: bool, console_mode: bool, ) -> Result<()> { unsafe { RoInitialize(RO_INIT_MULTITHREADED)?; } unsafe { MFStartup(MF_VERSION, MFSTARTUP_FULL)? } if wait_for_debugger { let pid = unsafe { GetCurrentProcessId() }; println!("Waiting for a debugger to attach (PID: {})...", pid); loop { if unsafe { IsDebuggerPresent().into() } { break; } std::thread::sleep(Duration::from_secs(1)); } unsafe { DebugBreak(); } } // Check to make sure Windows.Graphics.Capture is available
if verbose { println!( "Using index \"{}\" and path \"{}\".", display_index, output_path ); } // Get the display handle using the provided index let display_handle = get_display_handle_from_index(display_index) .expect("The provided display index was out of bounds!"); let item = create_capture_item_for_monitor(display_handle)?; // Resolve encoding settings let resolution = if let Some(resolution) = resolution.get_size() { resolution } else { item.Size()? }; let bit_rate = bit_rate * 1000000; let encoder_devices = VideoEncoderDevice::enumerate()?; if encoder_devices.is_empty() { exit_with_error("No hardware H264 encoders found!"); } if verbose { println!("Encoders ({}):", encoder_devices.len()); for encoder_device in &encoder_devices { println!(" {}", encoder_device.display_name()); } } let encoder_device = if let Some(encoder_device) = encoder_devices.get(encoder_index) { encoder_device } else { exit_with_error("Encoder index is out of bounds!"); }; if verbose { println!("Using: {}", encoder_device.display_name()); } // Create our file let path = unsafe { let mut output_path: Vec<u16> = output_path.encode_utf16().collect(); output_path.push(0); let mut new_path = vec![0u16; MAX_PATH as usize]; let length = GetFullPathNameW( PWSTR(output_path.as_mut_ptr()), new_path.len() as u32, PWSTR(new_path.as_mut_ptr()), std::ptr::null_mut(), ); new_path.resize(length as usize, 0); String::from_utf16(&new_path).unwrap() }; let path = Path::new(&path); let parent_folder_path = path.parent().unwrap(); let parent_folder = StorageFolder::GetFolderFromPathAsync(parent_folder_path.as_os_str().to_str().unwrap())? .get()?; let file_name = path.file_name().unwrap(); let file = parent_folder .CreateFileAsync( file_name.to_str().unwrap(), CreationCollisionOption::ReplaceExisting, )? .get()?; // Start the recording { let stream = file.OpenAsync(FileAccessMode::ReadWrite)?.get()?; let d3d_device = create_d3d_device()?; let mut session = create_encoding_session( d3d_device, item, encoder_device, resolution, bit_rate, frame_rate, stream, )?; if !console_mode { let mut is_recording = false; pump_messages(|| -> Result<bool> { Ok(if !is_recording { is_recording = true; println!("Starting recording..."); session.start()?; false } else { true }) })?; println!("Stopping recording..."); } else { session.start()?; pause(); } session.stop()?; } Ok(()) } fn main() { let mut app = App::new(env!("CARGO_PKG_NAME")) .version(env!("CARGO_PKG_VERSION")) .author(env!("CARGO_PKG_AUTHORS")) .about(env!("CARGO_PKG_DESCRIPTION")) .arg( Arg::with_name("display") .short("d") .long("display") .value_name("display index") .help("The index of the display you'd like to record.") .takes_value(true) .default_value("0") .required(false), ) .arg( Arg::with_name("bitRate") .short("b") .long("bitRate") .value_name("bit rate (in Mbps)") .help("The bit rate you would like to encode at (in Mbps).") .takes_value(true) .default_value("18") .required(false), ) .arg( Arg::with_name("frameRate") .short("f") .long("frameRate") .value_name("frame rate") .help("The frame rate you would like to encode at.") .takes_value(true) .default_value("60") .required(false), ) .arg( Arg::with_name("resolution") .short("r") .long("resolution") .value_name("resolution enum") .help("The resolution you would like to encode at: native, 720p, 1080p, 2160p, or 4320p.") .takes_value(true) .default_value("native") .required(false), ) .arg( Arg::with_name("encoder") .short("e") .long("encoder") .value_name("encoder index") .help("The index of the encoder you'd like to use to record (use enum-encoders command for a list of encoders and their indices).") .takes_value(true) .default_value("0") .required(false), ) .arg( Arg::with_name("verbose") .short("v") .help("Enables verbose (debug) output.") .required(false), ) .arg( Arg::with_name("waitForDebugger") .long("waitForDebugger") .help("The program will wait for a debugger to attach before starting.") .required(false), ) .arg( Arg::with_name("consoleMode") .long("consoleMode") .help("Recording immediately starts. End the recording through console input.") .required(false), ) .arg( Arg::with_name("OUTPUT FILE") .help("The output file that will contain the recording.") .default_value("recording.mp4") .required(false), ) .subcommand( SubCommand::with_name("enum-encoders") .about("Lists the available hardware H264 encoders.") ); // Handle /? let args: Vec<_> = std::env::args().collect(); if args.contains(&"/?".to_owned()) { app.print_help().unwrap(); std::process::exit(0); } let matches = app.get_matches(); if let Some(name) = matches.subcommand_name() { if name == "enum-encoders" { enum_encoders().unwrap(); return; } } let monitor_index: usize = matches .value_of("display") .unwrap() .parse() .expect("Invalid diplay index value!"); let output_path = matches.value_of("OUTPUT FILE").unwrap(); let verbose = matches.is_present("verbose"); let wait_for_debugger = matches.is_present("waitForDebugger"); let console_mode = matches.is_present("consoleMode"); let bit_rate: u32 = matches .value_of("bitRate") .unwrap() .parse() .expect("Invalid bit rate value!"); let frame_rate: u32 = matches .value_of("frameRate") .unwrap() .parse() .expect("Invalid frame rate value!"); let resolution: Resolution = matches .value_of("resolution") .unwrap() .parse() .expect("Invalid resolution value! Expecting: native, 720p, 1080p, 2160p, or 4320p."); let encoder_index: usize = matches .value_of("encoder") .unwrap() .parse() .expect("Invalid encoder index value!"); // Validate some of the params if !validate_path(output_path) { exit_with_error("Invalid path specified!"); } let result = run( monitor_index, &output_path, bit_rate, frame_rate, resolution, encoder_index, verbose | wait_for_debugger, wait_for_debugger, console_mode, ); // We do this for nicer HRESULT printing when errors occur. if let Err(error) = result { error.code().unwrap(); } } fn pause() { println!("Press ENTER to stop recording..."); std::io::Read::read
if !required_capture_features_supported()? { exit_with_error("The required screen capture features are not supported on this device for this release of Windows!\nPlease update your operating system (minimum: Windows 10 Version 1903, Build 18362)."); }
random_line_split
main.rs
let encoder_devices = VideoEncoderDevice::enumerate()?; if encoder_devices.is_empty() { exit_with_error("No hardware H264 encoders found!"); } if verbose { println!("Encoders ({}):", encoder_devices.len()); for encoder_device in &encoder_devices { println!(" {}", encoder_device.display_name()); } } let encoder_device = if let Some(encoder_device) = encoder_devices.get(encoder_index) { encoder_device } else { exit_with_error("Encoder index is out of bounds!"); }; if verbose { println!("Using: {}", encoder_device.display_name()); } // Create our file let path = unsafe { let mut output_path: Vec<u16> = output_path.encode_utf16().collect(); output_path.push(0); let mut new_path = vec![0u16; MAX_PATH as usize]; let length = GetFullPathNameW( PWSTR(output_path.as_mut_ptr()), new_path.len() as u32, PWSTR(new_path.as_mut_ptr()), std::ptr::null_mut(), ); new_path.resize(length as usize, 0); String::from_utf16(&new_path).unwrap() }; let path = Path::new(&path); let parent_folder_path = path.parent().unwrap(); let parent_folder = StorageFolder::GetFolderFromPathAsync(parent_folder_path.as_os_str().to_str().unwrap())? .get()?; let file_name = path.file_name().unwrap(); let file = parent_folder .CreateFileAsync( file_name.to_str().unwrap(), CreationCollisionOption::ReplaceExisting, )? .get()?; // Start the recording { let stream = file.OpenAsync(FileAccessMode::ReadWrite)?.get()?; let d3d_device = create_d3d_device()?; let mut session = create_encoding_session( d3d_device, item, encoder_device, resolution, bit_rate, frame_rate, stream, )?; if !console_mode { let mut is_recording = false; pump_messages(|| -> Result<bool> { Ok(if !is_recording { is_recording = true; println!("Starting recording..."); session.start()?; false } else { true }) })?; println!("Stopping recording..."); } else { session.start()?; pause(); } session.stop()?; } Ok(()) } fn main() { let mut app = App::new(env!("CARGO_PKG_NAME")) .version(env!("CARGO_PKG_VERSION")) .author(env!("CARGO_PKG_AUTHORS")) .about(env!("CARGO_PKG_DESCRIPTION")) .arg( Arg::with_name("display") .short("d") .long("display") .value_name("display index") .help("The index of the display you'd like to record.") .takes_value(true) .default_value("0") .required(false), ) .arg( Arg::with_name("bitRate") .short("b") .long("bitRate") .value_name("bit rate (in Mbps)") .help("The bit rate you would like to encode at (in Mbps).") .takes_value(true) .default_value("18") .required(false), ) .arg( Arg::with_name("frameRate") .short("f") .long("frameRate") .value_name("frame rate") .help("The frame rate you would like to encode at.") .takes_value(true) .default_value("60") .required(false), ) .arg( Arg::with_name("resolution") .short("r") .long("resolution") .value_name("resolution enum") .help("The resolution you would like to encode at: native, 720p, 1080p, 2160p, or 4320p.") .takes_value(true) .default_value("native") .required(false), ) .arg( Arg::with_name("encoder") .short("e") .long("encoder") .value_name("encoder index") .help("The index of the encoder you'd like to use to record (use enum-encoders command for a list of encoders and their indices).") .takes_value(true) .default_value("0") .required(false), ) .arg( Arg::with_name("verbose") .short("v") .help("Enables verbose (debug) output.") .required(false), ) .arg( Arg::with_name("waitForDebugger") .long("waitForDebugger") .help("The program will wait for a debugger to attach before starting.") .required(false), ) .arg( Arg::with_name("consoleMode") .long("consoleMode") .help("Recording immediately starts. End the recording through console input.") .required(false), ) .arg( Arg::with_name("OUTPUT FILE") .help("The output file that will contain the recording.") .default_value("recording.mp4") .required(false), ) .subcommand( SubCommand::with_name("enum-encoders") .about("Lists the available hardware H264 encoders.") ); // Handle /? let args: Vec<_> = std::env::args().collect(); if args.contains(&"/?".to_owned()) { app.print_help().unwrap(); std::process::exit(0); } let matches = app.get_matches(); if let Some(name) = matches.subcommand_name() { if name == "enum-encoders" { enum_encoders().unwrap(); return; } } let monitor_index: usize = matches .value_of("display") .unwrap() .parse() .expect("Invalid diplay index value!"); let output_path = matches.value_of("OUTPUT FILE").unwrap(); let verbose = matches.is_present("verbose"); let wait_for_debugger = matches.is_present("waitForDebugger"); let console_mode = matches.is_present("consoleMode"); let bit_rate: u32 = matches .value_of("bitRate") .unwrap() .parse() .expect("Invalid bit rate value!"); let frame_rate: u32 = matches .value_of("frameRate") .unwrap() .parse() .expect("Invalid frame rate value!"); let resolution: Resolution = matches .value_of("resolution") .unwrap() .parse() .expect("Invalid resolution value! Expecting: native, 720p, 1080p, 2160p, or 4320p."); let encoder_index: usize = matches .value_of("encoder") .unwrap() .parse() .expect("Invalid encoder index value!"); // Validate some of the params if !validate_path(output_path) { exit_with_error("Invalid path specified!"); } let result = run( monitor_index, &output_path, bit_rate, frame_rate, resolution, encoder_index, verbose | wait_for_debugger, wait_for_debugger, console_mode, ); // We do this for nicer HRESULT printing when errors occur. if let Err(error) = result { error.code().unwrap(); } } fn pause() { println!("Press ENTER to stop recording..."); std::io::Read::read(&mut std::io::stdin(), &mut [0]).unwrap(); } fn enum_encoders() -> Result<()> { let encoder_devices = VideoEncoderDevice::enumerate()?; if encoder_devices.is_empty() { exit_with_error("No hardware H264 encoders found!"); } println!("Encoders ({}):", encoder_devices.len()); for (i, encoder_device) in encoder_devices.iter().enumerate() { println!(" {} - {}", i, encoder_device.display_name()); } Ok(()) } fn create_encoding_session( d3d_device: ID3D11Device, item: GraphicsCaptureItem, encoder_device: &VideoEncoderDevice, resolution: SizeInt32, bit_rate: u32, frame_rate: u32, stream: IRandomAccessStream, ) -> Result<VideoEncodingSession> { let result = VideoEncodingSession::new( d3d_device, item, encoder_device, resolution, bit_rate, frame_rate, stream, ); if result.is_err() { println!("Error during encoder setup, try another set of encoding settings."); } result } fn validate_path<P: AsRef<Path>>(path: P) -> bool { let path = path.as_ref(); let mut valid = true; if let Some(extension) = path.extension() { if extension != "mp4" { valid = false; } } else { valid = false; } valid } fn exit_with_error(message: &str) -> ! { println!("{}", message); std::process::exit(1); } fn win32_programmatic_capture_supported() -> Result<bool>
{ ApiInformation::IsApiContractPresentByMajor("Windows.Foundation.UniversalApiContract", 8) }
identifier_body
main.rs
}, }; fn run( display_index: usize, output_path: &str, bit_rate: u32, frame_rate: u32, resolution: Resolution, encoder_index: usize, verbose: bool, wait_for_debugger: bool, console_mode: bool, ) -> Result<()> { unsafe { RoInitialize(RO_INIT_MULTITHREADED)?; } unsafe { MFStartup(MF_VERSION, MFSTARTUP_FULL)? } if wait_for_debugger { let pid = unsafe { GetCurrentProcessId() }; println!("Waiting for a debugger to attach (PID: {})...", pid); loop { if unsafe { IsDebuggerPresent().into() } { break; } std::thread::sleep(Duration::from_secs(1)); } unsafe { DebugBreak(); } } // Check to make sure Windows.Graphics.Capture is available if !required_capture_features_supported()?
if verbose { println!( "Using index \"{}\" and path \"{}\".", display_index, output_path ); } // Get the display handle using the provided index let display_handle = get_display_handle_from_index(display_index) .expect("The provided display index was out of bounds!"); let item = create_capture_item_for_monitor(display_handle)?; // Resolve encoding settings let resolution = if let Some(resolution) = resolution.get_size() { resolution } else { item.Size()? }; let bit_rate = bit_rate * 1000000; let encoder_devices = VideoEncoderDevice::enumerate()?; if encoder_devices.is_empty() { exit_with_error("No hardware H264 encoders found!"); } if verbose { println!("Encoders ({}):", encoder_devices.len()); for encoder_device in &encoder_devices { println!(" {}", encoder_device.display_name()); } } let encoder_device = if let Some(encoder_device) = encoder_devices.get(encoder_index) { encoder_device } else { exit_with_error("Encoder index is out of bounds!"); }; if verbose { println!("Using: {}", encoder_device.display_name()); } // Create our file let path = unsafe { let mut output_path: Vec<u16> = output_path.encode_utf16().collect(); output_path.push(0); let mut new_path = vec![0u16; MAX_PATH as usize]; let length = GetFullPathNameW( PWSTR(output_path.as_mut_ptr()), new_path.len() as u32, PWSTR(new_path.as_mut_ptr()), std::ptr::null_mut(), ); new_path.resize(length as usize, 0); String::from_utf16(&new_path).unwrap() }; let path = Path::new(&path); let parent_folder_path = path.parent().unwrap(); let parent_folder = StorageFolder::GetFolderFromPathAsync(parent_folder_path.as_os_str().to_str().unwrap())? .get()?; let file_name = path.file_name().unwrap(); let file = parent_folder .CreateFileAsync( file_name.to_str().unwrap(), CreationCollisionOption::ReplaceExisting, )? .get()?; // Start the recording { let stream = file.OpenAsync(FileAccessMode::ReadWrite)?.get()?; let d3d_device = create_d3d_device()?; let mut session = create_encoding_session( d3d_device, item, encoder_device, resolution, bit_rate, frame_rate, stream, )?; if !console_mode { let mut is_recording = false; pump_messages(|| -> Result<bool> { Ok(if !is_recording { is_recording = true; println!("Starting recording..."); session.start()?; false } else { true }) })?; println!("Stopping recording..."); } else { session.start()?; pause(); } session.stop()?; } Ok(()) } fn main() { let mut app = App::new(env!("CARGO_PKG_NAME")) .version(env!("CARGO_PKG_VERSION")) .author(env!("CARGO_PKG_AUTHORS")) .about(env!("CARGO_PKG_DESCRIPTION")) .arg( Arg::with_name("display") .short("d") .long("display") .value_name("display index") .help("The index of the display you'd like to record.") .takes_value(true) .default_value("0") .required(false), ) .arg( Arg::with_name("bitRate") .short("b") .long("bitRate") .value_name("bit rate (in Mbps)") .help("The bit rate you would like to encode at (in Mbps).") .takes_value(true) .default_value("18") .required(false), ) .arg( Arg::with_name("frameRate") .short("f") .long("frameRate") .value_name("frame rate") .help("The frame rate you would like to encode at.") .takes_value(true) .default_value("60") .required(false), ) .arg( Arg::with_name("resolution") .short("r") .long("resolution") .value_name("resolution enum") .help("The resolution you would like to encode at: native, 720p, 1080p, 2160p, or 4320p.") .takes_value(true) .default_value("native") .required(false), ) .arg( Arg::with_name("encoder") .short("e") .long("encoder") .value_name("encoder index") .help("The index of the encoder you'd like to use to record (use enum-encoders command for a list of encoders and their indices).") .takes_value(true) .default_value("0") .required(false), ) .arg( Arg::with_name("verbose") .short("v") .help("Enables verbose (debug) output.") .required(false), ) .arg( Arg::with_name("waitForDebugger") .long("waitForDebugger") .help("The program will wait for a debugger to attach before starting.") .required(false), ) .arg( Arg::with_name("consoleMode") .long("consoleMode") .help("Recording immediately starts. End the recording through console input.") .required(false), ) .arg( Arg::with_name("OUTPUT FILE") .help("The output file that will contain the recording.") .default_value("recording.mp4") .required(false), ) .subcommand( SubCommand::with_name("enum-encoders") .about("Lists the available hardware H264 encoders.") ); // Handle /? let args: Vec<_> = std::env::args().collect(); if args.contains(&"/?".to_owned()) { app.print_help().unwrap(); std::process::exit(0); } let matches = app.get_matches(); if let Some(name) = matches.subcommand_name() { if name == "enum-encoders" { enum_encoders().unwrap(); return; } } let monitor_index: usize = matches .value_of("display") .unwrap() .parse() .expect("Invalid diplay index value!"); let output_path = matches.value_of("OUTPUT FILE").unwrap(); let verbose = matches.is_present("verbose"); let wait_for_debugger = matches.is_present("waitForDebugger"); let console_mode = matches.is_present("consoleMode"); let bit_rate: u32 = matches .value_of("bitRate") .unwrap() .parse() .expect("Invalid bit rate value!"); let frame_rate: u32 = matches .value_of("frameRate") .unwrap() .parse() .expect("Invalid frame rate value!"); let resolution: Resolution = matches .value_of("resolution") .unwrap() .parse() .expect("Invalid resolution value! Expecting: native, 720p, 1080p, 2160p, or 4320p."); let encoder_index: usize = matches .value_of("encoder") .unwrap() .parse() .expect("Invalid encoder index value!"); // Validate some of the params if !validate_path(output_path) { exit_with_error("Invalid path specified!"); } let result = run( monitor_index, &output_path, bit_rate, frame_rate, resolution, encoder_index, verbose | wait_for_debugger, wait_for_debugger, console_mode, ); // We do this for nicer HRESULT printing when errors occur. if let Err(error) = result { error.code().unwrap(); } } fn pause() { println!("Press ENTER to stop recording..."); std::io::Read::
{ exit_with_error("The required screen capture features are not supported on this device for this release of Windows!\nPlease update your operating system (minimum: Windows 10 Version 1903, Build 18362)."); }
conditional_block
main.rs
Session}, }; fn
( display_index: usize, output_path: &str, bit_rate: u32, frame_rate: u32, resolution: Resolution, encoder_index: usize, verbose: bool, wait_for_debugger: bool, console_mode: bool, ) -> Result<()> { unsafe { RoInitialize(RO_INIT_MULTITHREADED)?; } unsafe { MFStartup(MF_VERSION, MFSTARTUP_FULL)? } if wait_for_debugger { let pid = unsafe { GetCurrentProcessId() }; println!("Waiting for a debugger to attach (PID: {})...", pid); loop { if unsafe { IsDebuggerPresent().into() } { break; } std::thread::sleep(Duration::from_secs(1)); } unsafe { DebugBreak(); } } // Check to make sure Windows.Graphics.Capture is available if !required_capture_features_supported()? { exit_with_error("The required screen capture features are not supported on this device for this release of Windows!\nPlease update your operating system (minimum: Windows 10 Version 1903, Build 18362)."); } if verbose { println!( "Using index \"{}\" and path \"{}\".", display_index, output_path ); } // Get the display handle using the provided index let display_handle = get_display_handle_from_index(display_index) .expect("The provided display index was out of bounds!"); let item = create_capture_item_for_monitor(display_handle)?; // Resolve encoding settings let resolution = if let Some(resolution) = resolution.get_size() { resolution } else { item.Size()? }; let bit_rate = bit_rate * 1000000; let encoder_devices = VideoEncoderDevice::enumerate()?; if encoder_devices.is_empty() { exit_with_error("No hardware H264 encoders found!"); } if verbose { println!("Encoders ({}):", encoder_devices.len()); for encoder_device in &encoder_devices { println!(" {}", encoder_device.display_name()); } } let encoder_device = if let Some(encoder_device) = encoder_devices.get(encoder_index) { encoder_device } else { exit_with_error("Encoder index is out of bounds!"); }; if verbose { println!("Using: {}", encoder_device.display_name()); } // Create our file let path = unsafe { let mut output_path: Vec<u16> = output_path.encode_utf16().collect(); output_path.push(0); let mut new_path = vec![0u16; MAX_PATH as usize]; let length = GetFullPathNameW( PWSTR(output_path.as_mut_ptr()), new_path.len() as u32, PWSTR(new_path.as_mut_ptr()), std::ptr::null_mut(), ); new_path.resize(length as usize, 0); String::from_utf16(&new_path).unwrap() }; let path = Path::new(&path); let parent_folder_path = path.parent().unwrap(); let parent_folder = StorageFolder::GetFolderFromPathAsync(parent_folder_path.as_os_str().to_str().unwrap())? .get()?; let file_name = path.file_name().unwrap(); let file = parent_folder .CreateFileAsync( file_name.to_str().unwrap(), CreationCollisionOption::ReplaceExisting, )? .get()?; // Start the recording { let stream = file.OpenAsync(FileAccessMode::ReadWrite)?.get()?; let d3d_device = create_d3d_device()?; let mut session = create_encoding_session( d3d_device, item, encoder_device, resolution, bit_rate, frame_rate, stream, )?; if !console_mode { let mut is_recording = false; pump_messages(|| -> Result<bool> { Ok(if !is_recording { is_recording = true; println!("Starting recording..."); session.start()?; false } else { true }) })?; println!("Stopping recording..."); } else { session.start()?; pause(); } session.stop()?; } Ok(()) } fn main() { let mut app = App::new(env!("CARGO_PKG_NAME")) .version(env!("CARGO_PKG_VERSION")) .author(env!("CARGO_PKG_AUTHORS")) .about(env!("CARGO_PKG_DESCRIPTION")) .arg( Arg::with_name("display") .short("d") .long("display") .value_name("display index") .help("The index of the display you'd like to record.") .takes_value(true) .default_value("0") .required(false), ) .arg( Arg::with_name("bitRate") .short("b") .long("bitRate") .value_name("bit rate (in Mbps)") .help("The bit rate you would like to encode at (in Mbps).") .takes_value(true) .default_value("18") .required(false), ) .arg( Arg::with_name("frameRate") .short("f") .long("frameRate") .value_name("frame rate") .help("The frame rate you would like to encode at.") .takes_value(true) .default_value("60") .required(false), ) .arg( Arg::with_name("resolution") .short("r") .long("resolution") .value_name("resolution enum") .help("The resolution you would like to encode at: native, 720p, 1080p, 2160p, or 4320p.") .takes_value(true) .default_value("native") .required(false), ) .arg( Arg::with_name("encoder") .short("e") .long("encoder") .value_name("encoder index") .help("The index of the encoder you'd like to use to record (use enum-encoders command for a list of encoders and their indices).") .takes_value(true) .default_value("0") .required(false), ) .arg( Arg::with_name("verbose") .short("v") .help("Enables verbose (debug) output.") .required(false), ) .arg( Arg::with_name("waitForDebugger") .long("waitForDebugger") .help("The program will wait for a debugger to attach before starting.") .required(false), ) .arg( Arg::with_name("consoleMode") .long("consoleMode") .help("Recording immediately starts. End the recording through console input.") .required(false), ) .arg( Arg::with_name("OUTPUT FILE") .help("The output file that will contain the recording.") .default_value("recording.mp4") .required(false), ) .subcommand( SubCommand::with_name("enum-encoders") .about("Lists the available hardware H264 encoders.") ); // Handle /? let args: Vec<_> = std::env::args().collect(); if args.contains(&"/?".to_owned()) { app.print_help().unwrap(); std::process::exit(0); } let matches = app.get_matches(); if let Some(name) = matches.subcommand_name() { if name == "enum-encoders" { enum_encoders().unwrap(); return; } } let monitor_index: usize = matches .value_of("display") .unwrap() .parse() .expect("Invalid diplay index value!"); let output_path = matches.value_of("OUTPUT FILE").unwrap(); let verbose = matches.is_present("verbose"); let wait_for_debugger = matches.is_present("waitForDebugger"); let console_mode = matches.is_present("consoleMode"); let bit_rate: u32 = matches .value_of("bitRate") .unwrap() .parse() .expect("Invalid bit rate value!"); let frame_rate: u32 = matches .value_of("frameRate") .unwrap() .parse() .expect("Invalid frame rate value!"); let resolution: Resolution = matches .value_of("resolution") .unwrap() .parse() .expect("Invalid resolution value! Expecting: native, 720p, 1080p, 2160p, or 4320p."); let encoder_index: usize = matches .value_of("encoder") .unwrap() .parse() .expect("Invalid encoder index value!"); // Validate some of the params if !validate_path(output_path) { exit_with_error("Invalid path specified!"); } let result = run( monitor_index, &output_path, bit_rate, frame_rate, resolution, encoder_index, verbose | wait_for_debugger, wait_for_debugger, console_mode, ); // We do this for nicer HRESULT printing when errors occur. if let Err(error) = result { error.code().unwrap(); } } fn pause() { println!("Press ENTER to stop recording..."); std::io::Read::
run
identifier_name
utils.py
def create_table_descriptive_row_from_analysis( attribute_name: Text, base_analysis: Analysis, additional_analysis: Analysis, figure_base_path: Text ) -> Text: # pylint: disable-msg=too-many-locals """Create makrdown formatted descriptive analysis result Args:
figure_base_path: (string), the folder for holding figures Returns: string, markdown formatted content """ row_template = template.TABLE_DESCRIPTIVE_ROW_TEMPLATE stats_template = template.TABLE_DESCRIPTIVE_STATS_TEMPLATE metrics = base_analysis.smetrics attribute_type = base_analysis.features[0].type # Make sure the display order of each attribute is consistent common_order = query_constants.COMMON_ORDER if attribute_type == Attribute.NUMERICAL: detail_order = query_constants.NUMERICAL_ORDER else: detail_order = query_constants.CATEGORICAL_ORDER # Use a OrderedDict to store the result result_holder = OrderedDict( [(item, 0) for item in common_order + detail_order]) for item in metrics: name = ScalarMetric.Name.Name(item.name) value = formatting.numeric_formatting(item.value) result_holder[name] = value # Construct the markdown formatted row row_stats_contents = [] for item in result_holder: row_stats_contents.append(stats_template.format( metric=item, value=result_holder[item] )) figure_path = visualization.plot_bar_chart(additional_analysis, figure_base_path) return row_template.format( name=attribute_name, type=Attribute.Type.Name(attribute_type), stats=' <br/> '.join(row_stats_contents), url=figure_path, alt_text=attribute_name, ) def create_table_from_table_metric(table_metric: TableMetric) -> Text: """Create a table for a TableMetric object. Currently, this function is used for Contingency_Table and TABLE_DESCRIPTIVE Examples: &#x200B;|Cash|Credit Card|No Charge|Unknown|Mobile|Prcard :-----:|:-----:|:-----:|:-----:|:-----:|:-----:|:-----: frequency|108114952.0|74475448.0|797730.0|369844.0|255082.0|192063.0 Args: table_metric: (analysis_entity_pb2.TableMetric) Returns: string """ supported_metric = { TableMetric.CONTINGENCY_TABLE, TableMetric.TABLE_DESCRIPTIVE } assert table_metric.name in supported_metric table_template = template.TABLE_TEMPLATE headers = ['&#x200B;'] + list(table_metric.column_indexes) header_string = "|".join(headers) header_separator = "|".join([":-----:" for i in range(len(headers))]) table_content = [] for row in table_metric.rows: # row header is in BOLD row_header = template.BOLD.format( content=str(row.row_index).strip()) row_values = [row_header] + [formatting.numeric_formatting(item.value) for item in row.cells] table_content.append("|".join(row_values)) table_content_string = "\n".join(table_content) return table_template.format( header=header_string, header_separator=header_separator, table_content=table_content_string ) def create_pairwise_metric_table( row_list: Set[Text], column_list: Set[Text], name_value_map: Dict[Text, float], same_match_value ) -> Text: """Construct table for pair-wise computed metrics, e.g., PEARSON_CORRELATION, ANOVA, CHI_SQUARE, INFORMATION_GAIN Examples: &#x200B;|tips|tolls|trip_total :-----:|:-----:|:-----:|:-----: tips|1|0.0001942405360750854|0.1952170878648758 tolls|0.0001942405360750854|1|0.22858665883541107 trip_total|0.1952170878648758|0.22858665883541107|1 Args: row_list: (List[str]), list of attribute names for table header column_list: (List[str]), list of attribute names for table row name name_value_map: (Dict[str, float]), map of name -> value same_match_value: value if the column and row name are the same. This could be either float or 'NA' depends on whether the computation of A-v.s.-A makes sense Returns: string """ table_template = template.TABLE_TEMPLATE headers = ['&#x200B;'] + list(column_list) header_string = "|".join(headers) header_separator = "|".join([":-----:" for i in range(len(headers))]) table_content = [] for row_name in row_list: # row header is in BOLD row_values = [template.BOLD.format(content=row_name.strip())] for col_name in column_list: # same_match_value is used when row_name == column_name if row_name == col_name: value = same_match_value else: value = name_value_map[row_name + '-' + col_name] # if the same_match_value is string, simply append it if isinstance(value, str): row_values.append(same_match_value) else: row_values.append(formatting.numeric_formatting(value)) table_content.append("|".join(row_values)) table_content_string = "\n".join(table_content) return table_template.format( header=header_string, header_separator=header_separator, table_content=table_content_string ) def create_no_order_pair_metric_section( analysis_list: List[Analysis], same_match_value: Union[Text, float], figure_base_path: Text, table_name: Text = "NA" ) -> Text: """Create metric table for pairwise comparison Args: analysis_list: (List[analysis_entity_pb2.Analysis]) same_match_value: (Union[str, float]) figure_base_path: (string), the folder for holding figures table_name: (str) Returns: string """ attribute_list = set() # a dictionary with {attributeone-attributetwo: metric_value} analysis_name_value_map = {} for item in analysis_list: value = item.smetrics[0].value name_list = [att.name for att in item.features] attribute_list.update(name_list) analysis_name_value_map['-'.join(name_list)] = value analysis_name_value_map['-'.join(reversed(name_list))] = value table_content = create_pairwise_metric_table( row_list=attribute_list, column_list=attribute_list, name_value_map=analysis_name_value_map, same_match_value=same_match_value) if table_name != "NA": figure_path = visualization.plot_heat_map_for_metric_table( heat_map_name=table_name, row_list=attribute_list, column_list=attribute_list, name_value_map=analysis_name_value_map, same_match_value=same_match_value, figure_base_path=figure_base_path) figure_content = template.IMAGE_TEMPLATE.format( url=figure_path, alt_text=table_name ) else: figure_content = "" return table_content + figure_content def create_order_pair_metric_section( analysis_list: List[Analysis], same_match_value: Union[Text, float] ) -> Text: """Create metric table for pairwise comparison Args: analysis_list: (List[analysis_entity_pb2.Analysis]) same_match_value: (Union[str, float]) Returns: string """ row_list = set() column_list = set() # a dictionary with {attributeone-attributetwo: metric_value} analysis_name_value_map = {} for item in analysis_list: value = item.smetrics[0].value name_list = [att.name for att in item.features] row_list.add(name_list[0]) column_list.add(name_list[1]) analysis_name_value_map['-'.join(name_list)] = value return create_pairwise_metric_table( row_list=row_list, column_list=column_list, name_value_map=analysis_name_value_map, same_match_value=same_match_value) def create_target_metrics_highlight( target_name: Text, metric_name_list: List[Text], metric_analysis_list: List[List[Analysis]] ) -> Text: # pylint: disable-msg=too-many-locals """Create the content for highlight section regarding a target attribute
attribute_name: (string), name of the attribute base_analysis: (analysis_entity_pb2.Analysis), analysis holding all the metrics additional_analysis: (analysis_entity_pb2.Analysis), histogram for numerical attribute, value_counts for categorical attributes
random_line_split
utils.py
def create_table_descriptive_row_from_analysis( attribute_name: Text, base_analysis: Analysis, additional_analysis: Analysis, figure_base_path: Text ) -> Text: # pylint: disable-msg=too-many-locals """Create makrdown formatted descriptive analysis result Args: attribute_name: (string), name of the attribute base_analysis: (analysis_entity_pb2.Analysis), analysis holding all the metrics additional_analysis: (analysis_entity_pb2.Analysis), histogram for numerical attribute, value_counts for categorical attributes figure_base_path: (string), the folder for holding figures Returns: string, markdown formatted content """ row_template = template.TABLE_DESCRIPTIVE_ROW_TEMPLATE stats_template = template.TABLE_DESCRIPTIVE_STATS_TEMPLATE metrics = base_analysis.smetrics attribute_type = base_analysis.features[0].type # Make sure the display order of each attribute is consistent common_order = query_constants.COMMON_ORDER if attribute_type == Attribute.NUMERICAL: detail_order = query_constants.NUMERICAL_ORDER else: detail_order = query_constants.CATEGORICAL_ORDER # Use a OrderedDict to store the result result_holder = OrderedDict( [(item, 0) for item in common_order + detail_order]) for item in metrics: name = ScalarMetric.Name.Name(item.name) value = formatting.numeric_formatting(item.value) result_holder[name] = value # Construct the markdown formatted row row_stats_contents = [] for item in result_holder: row_stats_contents.append(stats_template.format( metric=item, value=result_holder[item] )) figure_path = visualization.plot_bar_chart(additional_analysis, figure_base_path) return row_template.format( name=attribute_name, type=Attribute.Type.Name(attribute_type), stats=' <br/> '.join(row_stats_contents), url=figure_path, alt_text=attribute_name, ) def create_table_from_table_metric(table_metric: TableMetric) -> Text: """Create a table for a TableMetric object. Currently, this function is used for Contingency_Table and TABLE_DESCRIPTIVE Examples: &#x200B;|Cash|Credit Card|No Charge|Unknown|Mobile|Prcard :-----:|:-----:|:-----:|:-----:|:-----:|:-----:|:-----: frequency|108114952.0|74475448.0|797730.0|369844.0|255082.0|192063.0 Args: table_metric: (analysis_entity_pb2.TableMetric) Returns: string """ supported_metric = { TableMetric.CONTINGENCY_TABLE, TableMetric.TABLE_DESCRIPTIVE } assert table_metric.name in supported_metric table_template = template.TABLE_TEMPLATE headers = ['&#x200B;'] + list(table_metric.column_indexes) header_string = "|".join(headers) header_separator = "|".join([":-----:" for i in range(len(headers))]) table_content = [] for row in table_metric.rows: # row header is in BOLD row_header = template.BOLD.format( content=str(row.row_index).strip()) row_values = [row_header] + [formatting.numeric_formatting(item.value) for item in row.cells] table_content.append("|".join(row_values)) table_content_string = "\n".join(table_content) return table_template.format( header=header_string, header_separator=header_separator, table_content=table_content_string ) def create_pairwise_metric_table( row_list: Set[Text], column_list: Set[Text], name_value_map: Dict[Text, float], same_match_value ) -> Text: """Construct table for pair-wise computed metrics, e.g., PEARSON_CORRELATION, ANOVA, CHI_SQUARE, INFORMATION_GAIN Examples: &#x200B;|tips|tolls|trip_total :-----:|:-----:|:-----:|:-----: tips|1|0.0001942405360750854|0.1952170878648758 tolls|0.0001942405360750854|1|0.22858665883541107 trip_total|0.1952170878648758|0.22858665883541107|1 Args: row_list: (List[str]), list of attribute names for table header column_list: (List[str]), list of attribute names for table row name name_value_map: (Dict[str, float]), map of name -> value same_match_value: value if the column and row name are the same. This could be either float or 'NA' depends on whether the computation of A-v.s.-A makes sense Returns: string """ table_template = template.TABLE_TEMPLATE headers = ['&#x200B;'] + list(column_list) header_string = "|".join(headers) header_separator = "|".join([":-----:" for i in range(len(headers))]) table_content = [] for row_name in row_list: # row header is in BOLD row_values = [template.BOLD.format(content=row_name.strip())] for col_name in column_list: # same_match_value is used when row_name == column_name if row_name == col_name: value = same_match_value else: value = name_value_map[row_name + '-' + col_name] # if the same_match_value is string, simply append it if isinstance(value, str): row_values.append(same_match_value) else: row_values.append(formatting.numeric_formatting(value)) table_content.append("|".join(row_values)) table_content_string = "\n".join(table_content) return table_template.format( header=header_string, header_separator=header_separator, table_content=table_content_string ) def create_no_order_pair_metric_section( analysis_list: List[Analysis], same_match_value: Union[Text, float], figure_base_path: Text, table_name: Text = "NA" ) -> Text: """Create metric table for pairwise comparison Args: analysis_list: (List[analysis_entity_pb2.Analysis]) same_match_value: (Union[str, float]) figure_base_path: (string), the folder for holding figures table_name: (str) Returns: string """ attribute_list = set() # a dictionary with {attributeone-attributetwo: metric_value} analysis_name_value_map = {} for item in analysis_list: value = item.smetrics[0].value name_list = [att.name for att in item.features] attribute_list.update(name_list) analysis_name_value_map['-'.join(name_list)] = value analysis_name_value_map['-'.join(reversed(name_list))] = value table_content = create_pairwise_metric_table( row_list=attribute_list, column_list=attribute_list, name_value_map=analysis_name_value_map, same_match_value=same_match_value) if table_name != "NA": figure_path = visualization.plot_heat_map_for_metric_table( heat_map_name=table_name, row_list=attribute_list, column_list=attribute_list, name_value_map=analysis_name_value_map, same_match_value=same_match_value, figure_base_path=figure_base_path) figure_content = template.IMAGE_TEMPLATE.format( url=figure_path, alt_text=table_name ) else: figure_content = "" return table_content + figure_content def create_order_pair_metric_section( analysis_list: List[Analysis], same_match_value: Union[Text, float] ) -> Text:
row_list=row_list, column_list=column_list, name_value_map=analysis_name_value_map, same_match_value=same_match_value) def create_target_metrics_highlight( target_name: Text, metric_name_list: List[Text], metric_analysis_list: List[List[Analysis]] ) -> Text: # pylint: disable-msg=too-many-locals """Create the content for highlight section regarding a target attribute
"""Create metric table for pairwise comparison Args: analysis_list: (List[analysis_entity_pb2.Analysis]) same_match_value: (Union[str, float]) Returns: string """ row_list = set() column_list = set() # a dictionary with {attributeone-attributetwo: metric_value} analysis_name_value_map = {} for item in analysis_list: value = item.smetrics[0].value name_list = [att.name for att in item.features] row_list.add(name_list[0]) column_list.add(name_list[1]) analysis_name_value_map['-'.join(name_list)] = value return create_pairwise_metric_table(
identifier_body
utils.py
def create_table_descriptive_row_from_analysis( attribute_name: Text, base_analysis: Analysis, additional_analysis: Analysis, figure_base_path: Text ) -> Text: # pylint: disable-msg=too-many-locals """Create makrdown formatted descriptive analysis result Args: attribute_name: (string), name of the attribute base_analysis: (analysis_entity_pb2.Analysis), analysis holding all the metrics additional_analysis: (analysis_entity_pb2.Analysis), histogram for numerical attribute, value_counts for categorical attributes figure_base_path: (string), the folder for holding figures Returns: string, markdown formatted content """ row_template = template.TABLE_DESCRIPTIVE_ROW_TEMPLATE stats_template = template.TABLE_DESCRIPTIVE_STATS_TEMPLATE metrics = base_analysis.smetrics attribute_type = base_analysis.features[0].type # Make sure the display order of each attribute is consistent common_order = query_constants.COMMON_ORDER if attribute_type == Attribute.NUMERICAL: detail_order = query_constants.NUMERICAL_ORDER else: detail_order = query_constants.CATEGORICAL_ORDER # Use a OrderedDict to store the result result_holder = OrderedDict( [(item, 0) for item in common_order + detail_order]) for item in metrics: name = ScalarMetric.Name.Name(item.name) value = formatting.numeric_formatting(item.value) result_holder[name] = value # Construct the markdown formatted row row_stats_contents = [] for item in result_holder: row_stats_contents.append(stats_template.format( metric=item, value=result_holder[item] )) figure_path = visualization.plot_bar_chart(additional_analysis, figure_base_path) return row_template.format( name=attribute_name, type=Attribute.Type.Name(attribute_type), stats=' <br/> '.join(row_stats_contents), url=figure_path, alt_text=attribute_name, ) def create_table_from_table_metric(table_metric: TableMetric) -> Text: """Create a table for a TableMetric object. Currently, this function is used for Contingency_Table and TABLE_DESCRIPTIVE Examples: &#x200B;|Cash|Credit Card|No Charge|Unknown|Mobile|Prcard :-----:|:-----:|:-----:|:-----:|:-----:|:-----:|:-----: frequency|108114952.0|74475448.0|797730.0|369844.0|255082.0|192063.0 Args: table_metric: (analysis_entity_pb2.TableMetric) Returns: string """ supported_metric = { TableMetric.CONTINGENCY_TABLE, TableMetric.TABLE_DESCRIPTIVE } assert table_metric.name in supported_metric table_template = template.TABLE_TEMPLATE headers = ['&#x200B;'] + list(table_metric.column_indexes) header_string = "|".join(headers) header_separator = "|".join([":-----:" for i in range(len(headers))]) table_content = [] for row in table_metric.rows: # row header is in BOLD row_header = template.BOLD.format( content=str(row.row_index).strip()) row_values = [row_header] + [formatting.numeric_formatting(item.value) for item in row.cells] table_content.append("|".join(row_values)) table_content_string = "\n".join(table_content) return table_template.format( header=header_string, header_separator=header_separator, table_content=table_content_string ) def
( row_list: Set[Text], column_list: Set[Text], name_value_map: Dict[Text, float], same_match_value ) -> Text: """Construct table for pair-wise computed metrics, e.g., PEARSON_CORRELATION, ANOVA, CHI_SQUARE, INFORMATION_GAIN Examples: &#x200B;|tips|tolls|trip_total :-----:|:-----:|:-----:|:-----: tips|1|0.0001942405360750854|0.1952170878648758 tolls|0.0001942405360750854|1|0.22858665883541107 trip_total|0.1952170878648758|0.22858665883541107|1 Args: row_list: (List[str]), list of attribute names for table header column_list: (List[str]), list of attribute names for table row name name_value_map: (Dict[str, float]), map of name -> value same_match_value: value if the column and row name are the same. This could be either float or 'NA' depends on whether the computation of A-v.s.-A makes sense Returns: string """ table_template = template.TABLE_TEMPLATE headers = ['&#x200B;'] + list(column_list) header_string = "|".join(headers) header_separator = "|".join([":-----:" for i in range(len(headers))]) table_content = [] for row_name in row_list: # row header is in BOLD row_values = [template.BOLD.format(content=row_name.strip())] for col_name in column_list: # same_match_value is used when row_name == column_name if row_name == col_name: value = same_match_value else: value = name_value_map[row_name + '-' + col_name] # if the same_match_value is string, simply append it if isinstance(value, str): row_values.append(same_match_value) else: row_values.append(formatting.numeric_formatting(value)) table_content.append("|".join(row_values)) table_content_string = "\n".join(table_content) return table_template.format( header=header_string, header_separator=header_separator, table_content=table_content_string ) def create_no_order_pair_metric_section( analysis_list: List[Analysis], same_match_value: Union[Text, float], figure_base_path: Text, table_name: Text = "NA" ) -> Text: """Create metric table for pairwise comparison Args: analysis_list: (List[analysis_entity_pb2.Analysis]) same_match_value: (Union[str, float]) figure_base_path: (string), the folder for holding figures table_name: (str) Returns: string """ attribute_list = set() # a dictionary with {attributeone-attributetwo: metric_value} analysis_name_value_map = {} for item in analysis_list: value = item.smetrics[0].value name_list = [att.name for att in item.features] attribute_list.update(name_list) analysis_name_value_map['-'.join(name_list)] = value analysis_name_value_map['-'.join(reversed(name_list))] = value table_content = create_pairwise_metric_table( row_list=attribute_list, column_list=attribute_list, name_value_map=analysis_name_value_map, same_match_value=same_match_value) if table_name != "NA": figure_path = visualization.plot_heat_map_for_metric_table( heat_map_name=table_name, row_list=attribute_list, column_list=attribute_list, name_value_map=analysis_name_value_map, same_match_value=same_match_value, figure_base_path=figure_base_path) figure_content = template.IMAGE_TEMPLATE.format( url=figure_path, alt_text=table_name ) else: figure_content = "" return table_content + figure_content def create_order_pair_metric_section( analysis_list: List[Analysis], same_match_value: Union[Text, float] ) -> Text: """Create metric table for pairwise comparison Args: analysis_list: (List[analysis_entity_pb2.Analysis]) same_match_value: (Union[str, float]) Returns: string """ row_list = set() column_list = set() # a dictionary with {attributeone-attributetwo: metric_value} analysis_name_value_map = {} for item in analysis_list: value = item.smetrics[0].value name_list = [att.name for att in item.features] row_list.add(name_list[0]) column_list.add(name_list[1]) analysis_name_value_map['-'.join(name_list)] = value return create_pairwise_metric_table( row_list=row_list, column_list=column_list, name_value_map=analysis_name_value_map, same_match_value=same_match_value) def create_target_metrics_highlight( target_name: Text, metric_name_list: List[Text], metric_analysis_list: List[List[Analysis]] ) -> Text: # pylint: disable-msg=too-many-locals """Create the content for highlight section regarding a target attribute
create_pairwise_metric_table
identifier_name
utils.py
.append(stats_template.format( metric=item, value=result_holder[item] )) figure_path = visualization.plot_bar_chart(additional_analysis, figure_base_path) return row_template.format( name=attribute_name, type=Attribute.Type.Name(attribute_type), stats=' <br/> '.join(row_stats_contents), url=figure_path, alt_text=attribute_name, ) def create_table_from_table_metric(table_metric: TableMetric) -> Text: """Create a table for a TableMetric object. Currently, this function is used for Contingency_Table and TABLE_DESCRIPTIVE Examples: &#x200B;|Cash|Credit Card|No Charge|Unknown|Mobile|Prcard :-----:|:-----:|:-----:|:-----:|:-----:|:-----:|:-----: frequency|108114952.0|74475448.0|797730.0|369844.0|255082.0|192063.0 Args: table_metric: (analysis_entity_pb2.TableMetric) Returns: string """ supported_metric = { TableMetric.CONTINGENCY_TABLE, TableMetric.TABLE_DESCRIPTIVE } assert table_metric.name in supported_metric table_template = template.TABLE_TEMPLATE headers = ['&#x200B;'] + list(table_metric.column_indexes) header_string = "|".join(headers) header_separator = "|".join([":-----:" for i in range(len(headers))]) table_content = [] for row in table_metric.rows: # row header is in BOLD row_header = template.BOLD.format( content=str(row.row_index).strip()) row_values = [row_header] + [formatting.numeric_formatting(item.value) for item in row.cells] table_content.append("|".join(row_values)) table_content_string = "\n".join(table_content) return table_template.format( header=header_string, header_separator=header_separator, table_content=table_content_string ) def create_pairwise_metric_table( row_list: Set[Text], column_list: Set[Text], name_value_map: Dict[Text, float], same_match_value ) -> Text: """Construct table for pair-wise computed metrics, e.g., PEARSON_CORRELATION, ANOVA, CHI_SQUARE, INFORMATION_GAIN Examples: &#x200B;|tips|tolls|trip_total :-----:|:-----:|:-----:|:-----: tips|1|0.0001942405360750854|0.1952170878648758 tolls|0.0001942405360750854|1|0.22858665883541107 trip_total|0.1952170878648758|0.22858665883541107|1 Args: row_list: (List[str]), list of attribute names for table header column_list: (List[str]), list of attribute names for table row name name_value_map: (Dict[str, float]), map of name -> value same_match_value: value if the column and row name are the same. This could be either float or 'NA' depends on whether the computation of A-v.s.-A makes sense Returns: string """ table_template = template.TABLE_TEMPLATE headers = ['&#x200B;'] + list(column_list) header_string = "|".join(headers) header_separator = "|".join([":-----:" for i in range(len(headers))]) table_content = [] for row_name in row_list: # row header is in BOLD row_values = [template.BOLD.format(content=row_name.strip())] for col_name in column_list: # same_match_value is used when row_name == column_name if row_name == col_name: value = same_match_value else: value = name_value_map[row_name + '-' + col_name] # if the same_match_value is string, simply append it if isinstance(value, str): row_values.append(same_match_value) else: row_values.append(formatting.numeric_formatting(value)) table_content.append("|".join(row_values)) table_content_string = "\n".join(table_content) return table_template.format( header=header_string, header_separator=header_separator, table_content=table_content_string ) def create_no_order_pair_metric_section( analysis_list: List[Analysis], same_match_value: Union[Text, float], figure_base_path: Text, table_name: Text = "NA" ) -> Text: """Create metric table for pairwise comparison Args: analysis_list: (List[analysis_entity_pb2.Analysis]) same_match_value: (Union[str, float]) figure_base_path: (string), the folder for holding figures table_name: (str) Returns: string """ attribute_list = set() # a dictionary with {attributeone-attributetwo: metric_value} analysis_name_value_map = {} for item in analysis_list: value = item.smetrics[0].value name_list = [att.name for att in item.features] attribute_list.update(name_list) analysis_name_value_map['-'.join(name_list)] = value analysis_name_value_map['-'.join(reversed(name_list))] = value table_content = create_pairwise_metric_table( row_list=attribute_list, column_list=attribute_list, name_value_map=analysis_name_value_map, same_match_value=same_match_value) if table_name != "NA": figure_path = visualization.plot_heat_map_for_metric_table( heat_map_name=table_name, row_list=attribute_list, column_list=attribute_list, name_value_map=analysis_name_value_map, same_match_value=same_match_value, figure_base_path=figure_base_path) figure_content = template.IMAGE_TEMPLATE.format( url=figure_path, alt_text=table_name ) else: figure_content = "" return table_content + figure_content def create_order_pair_metric_section( analysis_list: List[Analysis], same_match_value: Union[Text, float] ) -> Text: """Create metric table for pairwise comparison Args: analysis_list: (List[analysis_entity_pb2.Analysis]) same_match_value: (Union[str, float]) Returns: string """ row_list = set() column_list = set() # a dictionary with {attributeone-attributetwo: metric_value} analysis_name_value_map = {} for item in analysis_list: value = item.smetrics[0].value name_list = [att.name for att in item.features] row_list.add(name_list[0]) column_list.add(name_list[1]) analysis_name_value_map['-'.join(name_list)] = value return create_pairwise_metric_table( row_list=row_list, column_list=column_list, name_value_map=analysis_name_value_map, same_match_value=same_match_value) def create_target_metrics_highlight( target_name: Text, metric_name_list: List[Text], metric_analysis_list: List[List[Analysis]] ) -> Text: # pylint: disable-msg=too-many-locals """Create the content for highlight section regarding a target attribute Args: target_name: (string) metric_name_list: (List(string) metric_analysis_list: (List[List[analysis_entity_pb2.Analysis]]) Returns: """ assert len(metric_name_list) == len(metric_analysis_list) # Every metric should have the same length, i.e., target v.s. remaining assert len({len(item) for item in metric_analysis_list}) == 1 name_enrich = { 'ANOVA': 'ANOVA P-value', 'CHI_SQUARE': 'Chi-square P-value', 'INFORMATION_GAIN': 'Information Gain', 'PEARSON_CORRELATION': 'Correlation Coefficient' } table_template = template.TARGET_METRIC_HIGHLIGHT_TEMPLATE row_template = template.TARGET_METRIC_HIGHLIGHT_ROW_TEMPLATE num_metrics = len(metric_name_list) enrich_name_list = [name_enrich[item] if item in name_enrich else item for item in metric_name_list] metric_names_str = '|'.join(enrich_name_list) separator_str = ':-----:|' * num_metrics attribute_set = set() metric_holders = {metric: {} for metric in metric_name_list} for i in range(num_metrics): for analysis in metric_analysis_list[i]:
metric_name = Analysis.Name.Name(analysis.name) attribute_name = [att.name for att in analysis.features if att.name != target_name][0] attribute_set.add(attribute_name) metric_value = analysis.smetrics[0].value metric_holders[metric_name][attribute_name] = metric_value
conditional_block
receiver.rs
.loss_list .iter_mut() .filter(|lle| lle.feedback_time < now - lle.k * rtt) { pak.k += 1; pak.feedback_time = now; ret.push(pak.seq_num); } ret }; if seq_nums.is_empty() { return Ok(()); } // send the nak self.send_nak(cx, seq_nums.into_iter())?; Ok(()) } // checks the timers // if a timer was triggered, then an RSFutureTimeout will be returned // if not, the socket is given back fn check_timers(&mut self, cx: &mut Context) -> Result<(), Error> { // see if we need to ACK or NAK if let Poll::Ready(Some(_)) = self.ack_interval().poll_next(cx) { self.on_ack_event(cx)?; } if let Poll::Ready(_) = self.nak_interval().poll(cx) { self.on_nak_event(cx)?; } // no need to do anything specific let _ = self.release_delay().poll(cx); Ok(()) } // handles a SRT control packet fn handle_srt_control_packet(&mut self, pack: &SrtControlPacket) -> Result<(), Error> { use self::SrtControlPacket::*; match pack { HandshakeRequest(_) | HandshakeResponse(_) => { warn!("Received handshake SRT packet, HSv5 expected"); } _ => unimplemented!(), } Ok(()) } // handles an incomming a packet fn handle_packet( &mut self, cx: &mut Context, packet: &Packet, from: &SocketAddr, ) -> Result<(), Error> { // We don't care about packets from elsewhere if *from != self.settings.remote { info!("Packet received from unknown address: {:?}", from); return Ok(()); } if self.settings.local_sockid != packet.dest_sockid() { // packet isn't applicable info!( "Packet send to socket id ({}) that does not match local ({})", packet.dest_sockid().0, self.settings.local_sockid.0 ); return Ok(()); } trace!("Received packet: {:?}", packet); match packet { Packet::Control(ctrl) => { // handle the control packet match &ctrl.control_type { ControlTypes::Ack { .. } => warn!("Receiver received ACK packet, unusual"), ControlTypes::Ack2(seq_num) => self.handle_ack2(*seq_num)?, ControlTypes::DropRequest { .. } => unimplemented!(), ControlTypes::Handshake(_) => { if let Some(ret) = self.hs_returner.as_ref() { if let Some(pack) = (*ret)(&packet) { self.send_to_remote(cx, pack)?; } } } ControlTypes::KeepAlive => {} // TODO: actually reset EXP etc ControlTypes::Nak { .. } => warn!("Receiver received NAK packet, unusual"), ControlTypes::Shutdown => { info!("Shutdown packet received, flushing receiver..."); self.shutdown_flag = true; } // end of stream ControlTypes::Srt(srt_packet) => { self.handle_srt_control_packet(srt_packet)?; } } } Packet::Data(data) => self.handle_data_packet(cx, &data)?, }; Ok(()) } fn handle_ack2(&mut self, seq_num: i32) -> Result<(), Error> { // 1) Locate the related ACK in the ACK History Window according to the // ACK sequence number in this ACK2. let id_in_wnd = match self .ack_history_window .as_slice() .binary_search_by(|entry| entry.ack_seq_num.cmp(&seq_num)) { Ok(i) => Some(i), Err(_) => None, }; if let Some(id) = id_in_wnd { let AckHistoryEntry { timestamp: send_timestamp, ack_number, .. } = self.ack_history_window[id]; // 2) Update the largest ACK number ever been acknowledged. self.lr_ack_acked = (seq_num, ack_number); // 3) Calculate new rtt according to the ACK2 arrival time and the ACK // departure time, and update the RTT value as: RTT = (RTT * 7 + // rtt) / 8 let immediate_rtt = self.get_timestamp_now() - send_timestamp; self.rtt = (self.rtt * 7 + immediate_rtt) / 8; // 4) Update RTTVar by: RTTVar = (RTTVar * 3 + abs(RTT - rtt)) / 4. self.rtt_variance = (self.rtt_variance * 3 + (self.rtt_variance - immediate_rtt).abs()) / 4; // 5) Update both ACK and NAK period to 4 * RTT + RTTVar + SYN. let ack_us = 4 * self.rtt as u64 + self.rtt_variance as u64 + 10_000; self.ack_interval = interval(Duration::from_micros(ack_us)); } else { warn!( "ACK sequence number in ACK2 packet not found in ACK history: {}", seq_num ); } Ok(()) } fn handle_data_packet(&mut self, cx: &mut Context, data: &DataPacket) -> Result<(), Error> { let now = self.get_timestamp_now(); // 1) Reset the ExpCount to 1. If there is no unacknowledged data // packet, or if this is an ACK or NAK control packet, reset the EXP // timer. self.exp_count = 1; // 2&3 don't apply // 4) If the sequence number of the current data packet is 16n + 1, // where n is an integer, record the time interval between this if data.seq_number % 16 == 0 { self.probe_time = Some(now) } else if data.seq_number % 16 == 1 { // if there is an entry if let Some(pt) = self.probe_time { // calculate and insert self.packet_pair_window.push((data.seq_number, now - pt)); // reset self.probe_time = None; } } // 5) Record the packet arrival time in PKT History Window. self.packet_history_window.push((data.seq_number, now)); // 6) // a. If the sequence number of the current data packet is greater // than LRSN, put all the sequence numbers between (but // excluding) these two values into the receiver's loss list and // send them to the sender in an NAK packet. match data.seq_number.cmp(&self.lrsn) { Ordering::Greater => { // lrsn is the latest packet received, so nak the one after that for i in seq_num_range(self.lrsn, data.seq_number) { self.loss_list.push(LossListEntry { seq_num: i, feedback_time: now, // k is initialized at 2, as stated on page 12 (very end) k: 2, }) } self.send_nak(cx, seq_num_range(self.lrsn, data.seq_number))?; } // b. If the sequence number is less than LRSN, remove it from the // receiver's loss list. Ordering::Less => { match self.loss_list[..].binary_search_by(|ll| ll.seq_num.cmp(&data.seq_number)) { Ok(i) => { self.loss_list.remove(i); } Err(_) => { debug!( "Packet received that's not in the loss list: {:?}, loss_list={:?}", data.seq_number, self.loss_list .iter() .map(|ll| ll.seq_num.as_raw()) .collect::<Vec<_>>() ); } }; } Ordering::Equal => {} } // record that we got this packet self.lrsn = cmp::max(data.seq_number + 1, self.lrsn); // we've already gotten this packet, drop it if self.buffer.next_release() > data.seq_number { debug!("Received packet {:?} twice", data.seq_number); return Ok(()); } self.buffer.add(data.clone()); trace!( "Received data packet seq_num={}, loc={:?}, buffer={:?}", data.seq_number, data.message_loc, self.buffer, ); Ok(()) } // send a NAK, and return the future fn send_nak<I>(&mut self, cx: &mut Context, lost_seq_nums: I) -> Result<(), Error> where I: Iterator<Item = SeqNumber>,
{ let vec: Vec<_> = lost_seq_nums.collect(); debug!("Sending NAK for={:?}", vec); let pack = self.make_control_packet(ControlTypes::Nak( compress_loss_list(vec.iter().cloned()).collect(), )); self.send_to_remote(cx, pack)?; Ok(()) }
identifier_body
receiver.rs
hs_returner, sock, rtt: 10_000, rtt_variance: 1_000, listen_timeout: Duration::from_secs(1), loss_list: Vec::new(), ack_history_window: Vec::new(), packet_history_window: Vec::new(), packet_pair_window: Vec::new(), ack_interval: interval(Duration::from_millis(10)), nak_interval: delay_for(Duration::from_millis(10)), lrsn: init_seq_num, // at start, we have received everything until the first packet, exclusive (aka nothing) next_ack: 1, exp_count: 1, probe_time: None, timeout_timer: delay_for(Duration::from_secs(1)), lr_ack_acked: (0, init_seq_num), buffer: RecvBuffer::new(init_seq_num), shutdown_flag: false, release_delay: delay_for(Duration::from_secs(0)), // start with an empty delay send_wrapper: SinkSendWrapper::new(), } } pub fn settings(&self) -> &ConnectionSettings { &self.settings } pub fn remote(&self) -> SocketAddr { self.settings.remote } fn timeout_timer(&mut self) -> Pin<&mut Delay> { Pin::new(&mut self.timeout_timer) } fn sock(&mut self) -> Pin<&mut T> { Pin::new(&mut self.sock) } fn ack_interval(&mut self) -> Pin<&mut Interval> { Pin::new(&mut self.ack_interval) } fn nak_interval(&mut self) -> Pin<&mut Delay> { Pin::new(&mut self.nak_interval) } fn release_delay(&mut self) -> Pin<&mut Delay> { Pin::new(&mut self.release_delay) } fn
(&mut self, cx: &mut Context, packet: Packet) -> Result<(), Error> { self.send_wrapper .send(&mut self.sock, (packet, self.settings.remote), cx) } fn reset_timeout(&mut self) { self.timeout_timer.reset(time::Instant::from_std( Instant::now() + self.listen_timeout, )) } fn on_ack_event(&mut self, cx: &mut Context) -> Result<(), Error> { // get largest inclusive received packet number let ack_number = match self.loss_list.first() { // There is an element in the loss list Some(i) => i.seq_num, // No elements, use lrsn, as it's already exclusive None => self.lrsn, }; // 2) If (a) the ACK number equals to the largest ACK number ever // acknowledged by ACK2 if ack_number == self.lr_ack_acked.1 { // stop (do not send this ACK). return Ok(()); } // make sure this ACK number is greater or equal to a one sent previously if let Some(w) = self.ack_history_window.last() { assert!(w.ack_number <= ack_number); } trace!( "Sending ACK; ack_num={:?}, lr_ack_acked={:?}", ack_number, self.lr_ack_acked.1 ); if let Some(&AckHistoryEntry { ack_number: last_ack_number, timestamp: last_timestamp, .. }) = self.ack_history_window.first() { // or, (b) it is equal to the ACK number in the // last ACK if last_ack_number == ack_number && // and the time interval between this two ACK packets is // less than 2 RTTs, (self.get_timestamp_now() - last_timestamp) < (self.rtt * 2) { // stop (do not send this ACK). return Ok(()); } } // 3) Assign this ACK a unique increasing ACK sequence number. let ack_seq_num = self.next_ack; self.next_ack += 1; // 4) Calculate the packet arrival speed according to the following // algorithm: let packet_recv_rate = { if self.packet_history_window.len() < 16 { 0 } else { // Calculate the median value of the last 16 packet arrival // intervals (AI) using the values stored in PKT History Window. let mut last_16: Vec<_> = self.packet_history_window [self.packet_history_window.len() - 16..] .iter() .map(|&(_, ts)| ts) .collect(); last_16.sort(); // the median timestamp let ai = last_16[last_16.len() / 2]; // In these 16 values, remove those either greater than AI*8 or // less than AI/8. let filtered: Vec<i32> = last_16 .iter() .filter(|&&n| n / 8 < ai && n > ai / 8) .cloned() .collect(); // If more than 8 values are left, calculate the // average of the left values AI', and the packet arrival speed is // 1/AI' (number of packets per second). Otherwise, return 0. if filtered.len() > 8 { (filtered.iter().fold(0i64, |sum, &val| sum + i64::from(val)) / (filtered.len() as i64)) as i32 } else { 0 } } }; // 5) Calculate the estimated link capacity according to the following algorithm: let est_link_cap = { if self.packet_pair_window.len() < 16 { 0 } else { // Calculate the median value of the last 16 packet pair // intervals (PI) using the values in Packet Pair Window, and the // link capacity is 1/PI (number of packets per second). let pi = { let mut last_16: Vec<_> = self.packet_pair_window [self.packet_pair_window.len() - 16..] .iter() .map(|&(_, time)| time) .collect(); last_16.sort(); last_16[last_16.len() / 2] }; // Multiply by 1M because pi is in microseconds // pi is in us/packet (1.0e6 / (pi as f32)) as i32 } }; // Pack the ACK packet with RTT, RTT Variance, and flow window size (available // receiver buffer size). let ack = self.make_control_packet(ControlTypes::Ack { ack_seq_num, ack_number, rtt: Some(self.rtt), rtt_variance: Some(self.rtt_variance), buffer_available: None, // TODO: add this packet_recv_rate: Some(packet_recv_rate), est_link_cap: Some(est_link_cap), }); // add it to the ack history let now = self.get_timestamp_now(); self.ack_history_window.push(AckHistoryEntry { ack_number, ack_seq_num, timestamp: now, }); self.send_to_remote(cx, ack)?; Ok(()) } fn on_nak_event(&mut self, cx: &mut Context) -> Result<(), Error> { // reset NAK timer, rtt and variance are in us, so convert to ns // NAK is used to trigger a negative acknowledgement (NAK). Its period // is dynamically updated to 4 * RTT_+ RTTVar + SYN, where RTTVar is the // variance of RTT samples. let nak_interval_us = 4 * self.rtt as u64 + self.rtt_variance as u64 + 10_000; self.nak_interval.reset(time::Instant::from_std( Instant::now() + Duration::from_micros(nak_interval_us), )); // Search the receiver's loss list, find out all those sequence numbers // whose last feedback time is k*RTT before, where k is initialized as 2 // and increased by 1 each time the number is fed back. Compress // (according to section 6.4) and send these numbers back to the sender // in an NAK packet. let now = self.get_timestamp_now(); // increment k and change feedback time, returning sequence numbers let seq_nums = { let mut ret = Vec::new(); let rtt = self.rtt; for pak in self .loss_list .iter_mut() .filter(|lle| lle.feedback_time < now - lle.k * rtt) { pak.k += 1; pak.feedback_time = now; ret.push(pak.seq_num); } ret }; if seq_nums.is_empty() { return Ok(()); } // send the nak self.send_nak(cx, seq_nums.into_iter())?; Ok(()) } // checks the timers // if a timer was triggered, then an RSFutureTimeout will be returned // if not, the socket is given back fn check_timers(&mut self, cx:
send_to_remote
identifier_name
receiver.rs
::new(&mut self.ack_interval) } fn nak_interval(&mut self) -> Pin<&mut Delay> { Pin::new(&mut self.nak_interval) } fn release_delay(&mut self) -> Pin<&mut Delay> { Pin::new(&mut self.release_delay) } fn send_to_remote(&mut self, cx: &mut Context, packet: Packet) -> Result<(), Error> { self.send_wrapper .send(&mut self.sock, (packet, self.settings.remote), cx) } fn reset_timeout(&mut self) { self.timeout_timer.reset(time::Instant::from_std( Instant::now() + self.listen_timeout, )) } fn on_ack_event(&mut self, cx: &mut Context) -> Result<(), Error> { // get largest inclusive received packet number let ack_number = match self.loss_list.first() { // There is an element in the loss list Some(i) => i.seq_num, // No elements, use lrsn, as it's already exclusive None => self.lrsn, }; // 2) If (a) the ACK number equals to the largest ACK number ever // acknowledged by ACK2 if ack_number == self.lr_ack_acked.1 { // stop (do not send this ACK). return Ok(()); } // make sure this ACK number is greater or equal to a one sent previously if let Some(w) = self.ack_history_window.last() { assert!(w.ack_number <= ack_number); } trace!( "Sending ACK; ack_num={:?}, lr_ack_acked={:?}", ack_number, self.lr_ack_acked.1 ); if let Some(&AckHistoryEntry { ack_number: last_ack_number, timestamp: last_timestamp, .. }) = self.ack_history_window.first() { // or, (b) it is equal to the ACK number in the // last ACK if last_ack_number == ack_number && // and the time interval between this two ACK packets is // less than 2 RTTs, (self.get_timestamp_now() - last_timestamp) < (self.rtt * 2) { // stop (do not send this ACK). return Ok(()); } } // 3) Assign this ACK a unique increasing ACK sequence number. let ack_seq_num = self.next_ack; self.next_ack += 1; // 4) Calculate the packet arrival speed according to the following // algorithm: let packet_recv_rate = { if self.packet_history_window.len() < 16 { 0 } else { // Calculate the median value of the last 16 packet arrival // intervals (AI) using the values stored in PKT History Window. let mut last_16: Vec<_> = self.packet_history_window [self.packet_history_window.len() - 16..] .iter() .map(|&(_, ts)| ts) .collect(); last_16.sort(); // the median timestamp let ai = last_16[last_16.len() / 2]; // In these 16 values, remove those either greater than AI*8 or // less than AI/8. let filtered: Vec<i32> = last_16 .iter() .filter(|&&n| n / 8 < ai && n > ai / 8) .cloned() .collect(); // If more than 8 values are left, calculate the // average of the left values AI', and the packet arrival speed is // 1/AI' (number of packets per second). Otherwise, return 0. if filtered.len() > 8 { (filtered.iter().fold(0i64, |sum, &val| sum + i64::from(val)) / (filtered.len() as i64)) as i32 } else { 0 } } }; // 5) Calculate the estimated link capacity according to the following algorithm: let est_link_cap = { if self.packet_pair_window.len() < 16 { 0 } else { // Calculate the median value of the last 16 packet pair // intervals (PI) using the values in Packet Pair Window, and the // link capacity is 1/PI (number of packets per second). let pi = { let mut last_16: Vec<_> = self.packet_pair_window [self.packet_pair_window.len() - 16..] .iter() .map(|&(_, time)| time) .collect(); last_16.sort(); last_16[last_16.len() / 2] }; // Multiply by 1M because pi is in microseconds // pi is in us/packet (1.0e6 / (pi as f32)) as i32 } }; // Pack the ACK packet with RTT, RTT Variance, and flow window size (available // receiver buffer size). let ack = self.make_control_packet(ControlTypes::Ack { ack_seq_num, ack_number, rtt: Some(self.rtt), rtt_variance: Some(self.rtt_variance), buffer_available: None, // TODO: add this packet_recv_rate: Some(packet_recv_rate), est_link_cap: Some(est_link_cap), }); // add it to the ack history let now = self.get_timestamp_now(); self.ack_history_window.push(AckHistoryEntry { ack_number, ack_seq_num, timestamp: now, }); self.send_to_remote(cx, ack)?; Ok(()) } fn on_nak_event(&mut self, cx: &mut Context) -> Result<(), Error> { // reset NAK timer, rtt and variance are in us, so convert to ns // NAK is used to trigger a negative acknowledgement (NAK). Its period // is dynamically updated to 4 * RTT_+ RTTVar + SYN, where RTTVar is the // variance of RTT samples. let nak_interval_us = 4 * self.rtt as u64 + self.rtt_variance as u64 + 10_000; self.nak_interval.reset(time::Instant::from_std( Instant::now() + Duration::from_micros(nak_interval_us), )); // Search the receiver's loss list, find out all those sequence numbers // whose last feedback time is k*RTT before, where k is initialized as 2 // and increased by 1 each time the number is fed back. Compress // (according to section 6.4) and send these numbers back to the sender // in an NAK packet. let now = self.get_timestamp_now(); // increment k and change feedback time, returning sequence numbers let seq_nums = { let mut ret = Vec::new(); let rtt = self.rtt; for pak in self .loss_list .iter_mut() .filter(|lle| lle.feedback_time < now - lle.k * rtt) { pak.k += 1; pak.feedback_time = now; ret.push(pak.seq_num); } ret }; if seq_nums.is_empty() { return Ok(()); } // send the nak self.send_nak(cx, seq_nums.into_iter())?; Ok(()) } // checks the timers // if a timer was triggered, then an RSFutureTimeout will be returned // if not, the socket is given back fn check_timers(&mut self, cx: &mut Context) -> Result<(), Error> { // see if we need to ACK or NAK if let Poll::Ready(Some(_)) = self.ack_interval().poll_next(cx) { self.on_ack_event(cx)?; } if let Poll::Ready(_) = self.nak_interval().poll(cx) { self.on_nak_event(cx)?; } // no need to do anything specific let _ = self.release_delay().poll(cx); Ok(()) } // handles a SRT control packet fn handle_srt_control_packet(&mut self, pack: &SrtControlPacket) -> Result<(), Error> { use self::SrtControlPacket::*; match pack { HandshakeRequest(_) | HandshakeResponse(_) => { warn!("Received handshake SRT packet, HSv5 expected"); } _ => unimplemented!(), } Ok(()) } // handles an incomming a packet fn handle_packet( &mut self, cx: &mut Context, packet: &Packet, from: &SocketAddr, ) -> Result<(), Error> { // We don't care about packets from elsewhere if *from != self.settings.remote { info!("Packet received from unknown address: {:?}", from); return Ok(()); } if self.settings.local_sockid != packet.dest_sockid() { // packet isn't applicable
info!( "Packet send to socket id ({}) that does not match local ({})", packet.dest_sockid().0, self.settings.local_sockid.0 );
random_line_split
cluster.go
Amazon *ClusterAmazon `json:"amazon"` } // ClusterAmazon offers Amazon-specific settings for that instance pool type ClusterAmazon struct { // This fields contains ARNs for additional IAM policies to be added to // this instance pool AdditionalIAMPolicies []string `json:"additionalIAMPolicies,omitempty"` // When set to true, AWS Elastic Block Storage volumes are encrypted EBSEncrypted *bool `json:"ebsEncrypted,omitempty"` } type ClusterKubernetes struct { Zone string `json:"zone,omitempty"` Version string `json:"version,omitempty"` PodCIDR string `json:"podCIDR,omitempty"` ServiceCIDR string `json:"serviceCIDR,omitempty"` ClusterAutoscaler *ClusterKubernetesClusterAutoscaler `json:"clusterAutoscaler,omitempty"` Tiller *ClusterKubernetesTiller `json:"tiller,omitempty"` Dashboard *ClusterKubernetesDashboard `json:"dashboard,omitempty"` PodSecurityPolicy *ClusterPodSecurityPolicy `json:"podSecurityPolicy,omitempty"` Prometheus *ClusterKubernetesPrometheus `json:"prometheus,omitempty"` Grafana *ClusterKubernetesGrafana `json:"grafana,omiempty"` Heapster *ClusterKubernetesHeapster `json:"heapster,omiempty"` InfluxDB *ClusterKubernetesInfluxDB `json:"influxDB,omiempty"` EncryptionProvider *ClusterEncryptionProvider `json:"encryptionProvider"` APIServer *ClusterKubernetesAPIServer `json:"apiServer,omitempty"` Kubelet *ClusterKubernetesKubelet `json:"kubelet,omitempty"` Scheduler *ClusterKubernetesScheduler `json:"scheduler,omitempty"` Proxy *ClusterKubernetesProxy `json:"proxy,omitempty"` ControllerManager *ClusterKubernetesControllerManager `json:"controllerManager,omitempty"` Calico *ClusterKubernetesCalico `json:"calico,omitempty"` GlobalFeatureGates map[string]bool `json:"globalFeatureGates,omitempty"` Hyperkube *bool `json:"hyperkube,omitempty"` } type ClusterKubernetesClusterAutoscaler struct { Enabled bool `json:"enabled,omitempty"` Image string `json:"image,omitempty"` Version string `json:"version,omitempty"` ScaleDownUtilizationThreshold *float64 `json:"scaleDownUtilizationThreshold,omitempty"` Overprovisioning *ClusterKubernetesClusterAutoscalerOverprovisioning `json:"overprovisioning,omitempty"` } type ClusterKubernetesClusterAutoscalerOverprovisioning struct { Enabled bool `json:"enabled,omitempty"` Image string `json:"image,omitempty"` Version string `json:"version,omitempty"` ReservedMillicoresPerReplica int `json:"reservedMillicoresPerReplica,omitempty"` ReservedMegabytesPerReplica int `json:"reservedMegabytesPerReplica,omitempty"` CoresPerReplica int `json:"coresPerReplica,omitempty"` NodesPerReplica int `json:"nodesPerReplica,omitempty"` ReplicaCount int `json:"replicaCount,omitempty"` } type ClusterKubernetesTiller struct { Enabled bool `json:"enabled,omitempty"` Image string `json:"image,omitempty"` Version string `json:"version,omitempty"` } type ClusterKubernetesDashboard struct { Enabled bool `json:"enabled,omitempty"` Image string `json:"image,omitempty"` Version string `json:"version,omitempty"` } type ClusterKubernetesAPIServer struct { // expose the API server through a public load balancer Public bool `json:"public,omitempty"` AllowCIDRs []string `json:"allowCIDRs,omitempty"` // create DNS record for the private load balancer, and optionally lock it down PrivateRecord bool `json:"privateRecord,omitempty"` PrivateAllowCIDRs []string `json:"privateAllowCIDRs,omitempty"` EnableAdmissionControllers []string `json:"enableAdmissionControllers,omitempty"` DisableAdmissionControllers []string `json:"disableAdmissionControllers,omitempty"` // OIDC OIDC *ClusterKubernetesAPIServerOIDC `json:"oidc,omitempty"` // AWS specific options Amazon *ClusterKubernetesAPIServerAmazon `json:"amazon,omitempty"` AuthTokenWebhookFile string `json:"authTokenWebhookFile,omitempty"` FeatureGates map[string]bool `json:"featureGates,omitempty"` } type ClusterKubernetesAPIServerOIDC struct { // The client ID for the OpenID Connect client, must be set if oidc-issuer-url is set. ClientID string `json:"clientID,omitempty" hiera:"kubernetes::apiserver::oidc_client_id"` // If provided, the name of a custom OpenID Connect claim for specifying // user groups. The claim value is expected to be a string or array of // strings. This flag is experimental, please see the authentication // documentation for further details. GroupsClaim string `json:"groupsClaim,omitempty" hiera:"kubernetes::apiserver::oidc_groups_claim"` // If provided, all groups will be prefixed with this value to prevent // conflicts with other authentication strategies. GroupsPrefix string `json:"groupsPrefix,omitempty" hiera:"kubernetes::apiserver::oidc_groups_prefix"` // The URL of the OpenID issuer, only HTTPS scheme will be accepted. If // set, it will be used to verify the OIDC JSON Web Token (JWT). IssuerURL string `json:"issuerURL,omitempty" hiera:"kubernetes::apiserver::oidc_issuer_url"` // Comma-separated list of allowed JOSE asymmetric signing algorithms. JWTs // with a 'alg' header value not in this list will be rejected. Values are // defined by RFC 7518 https://tools.ietf.org/html/rfc7518#section-3.1. // (default [RS256]) SigningAlgs []string `json:"signingAlgs,omitempty" hiera:"kubernetes::apiserver::oidc_signing_algs"` // The OpenID claim to use as the user name. Note that claims other than // the default ('sub') is not guaranteed to be unique and immutable. This // flag is experimental, please see the authentication documentation for // further details. (default "sub") UsernameClaim string `json:"usernameClaim,omitempty" hiera:"kubernetes::apiserver::oidc_username_claim"` // If provided, all usernames will be prefixed with this value. If not // provided, username claims other than 'email' are prefixed by the issuer // URL to avoid clashes. To skip any prefixing, provide the value '-'. UsernamePrefix string `json:"usernamePrefix,omitempty" hiera:"kubernetes::apiserver::oidc_username_prefix"` } type ClusterKubernetesAPIServerAmazon struct { PublicELBAccessLogs *ClusterKubernetesAPIServerAmazonAccessLogs `json:"publicELBAccessLogs,omitempty"` InternalELBAccessLogs *ClusterKubernetesAPIServerAmazonAccessLogs `json:"internalELBAccessLogs,omitempty"` AwsIAMAuthenticatorInit bool `json:"awsIAMAuthenticatorInit,omitempty"` } type ClusterKubernetesAPIServerAmazonAccessLogs struct { Enabled *bool `json:"enabled,omitempty"` Bucket string `json:"bucket,omitempty"` BucketPrefix string `json:"bucketPrefix,omitempty"` Interval *int `json:"interval,omitempty"` } type ClusterPodSecurityPolicy struct { Enabled bool `json:"enabled,omitempty"` } type ClusterEncryptionProvider struct { Enabled bool `json:"enabled,omitempty"` Version string `json:"version,omitempty"` } // Configure the cluster internal deployment of prometheus type ClusterKubernetesPrometheus struct { // Enable a cluster internal prometheus deployment, default: true Enabled bool `json:"enabled,omitempty"` // Mode defines which components are installed Mode string `json:"mode,omitempty"` } type ClusterKubernetesGrafana struct { Enabled bool `json:"enabled,omitempty"` } type ClusterKubernetesHeapster struct { Enabled bool `json:"enabled,omitempty"` } type ClusterKubernetesInfluxDB struct { Enabled bool `json:"enabled,omitempty"` } type ClusterVaultHelper struct { URL string `json:"url,omitempty"` } type ClusterKubernetesScheduler struct { FeatureGates map[string]bool `json:"featureGates,omitempty"` } type ClusterKubernetesKubelet struct { FeatureGates map[string]bool `json:"featureGates,omitempty"` } type ClusterKubernetesProxy struct { FeatureGates map[string]bool `json:"featureGates,omitempty"` } type ClusterKubernetesControllerManager struct { FeatureGates map[string]bool `json:"featureGates,omitempty"` } type ClusterKubernetesCalicoBackend string type ClusterKubernetesCalico struct { Backend ClusterKubernetesCalicoBackend `json:"backend"` EnableTypha bool `json:"enableTypha"` TyphaReplicas *int `json:"typhaReplicas"` } // +k8s:openapi-gen=true // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object type ClusterList struct { metav1.TypeMeta `json:",inline"` metav1.ObjectMeta `json:"metadata,omitempty"` Items []Cluster `json:"items"` } func NewCluster(name string) *Cluster {
return &Cluster{ ObjectMeta: metav1.ObjectMeta{ Name: name, }, } }
identifier_body
cluster.go
Type string `json:"-"` // This specifies if a cluster is a hub, single or multi // Amazon specific options Amazon *ClusterAmazon `json:"amazon"` } // ClusterAmazon offers Amazon-specific settings for that instance pool type ClusterAmazon struct { // This fields contains ARNs for additional IAM policies to be added to // this instance pool AdditionalIAMPolicies []string `json:"additionalIAMPolicies,omitempty"` // When set to true, AWS Elastic Block Storage volumes are encrypted EBSEncrypted *bool `json:"ebsEncrypted,omitempty"` } type ClusterKubernetes struct { Zone string `json:"zone,omitempty"` Version string `json:"version,omitempty"` PodCIDR string `json:"podCIDR,omitempty"` ServiceCIDR string `json:"serviceCIDR,omitempty"` ClusterAutoscaler *ClusterKubernetesClusterAutoscaler `json:"clusterAutoscaler,omitempty"` Tiller *ClusterKubernetesTiller `json:"tiller,omitempty"` Dashboard *ClusterKubernetesDashboard `json:"dashboard,omitempty"` PodSecurityPolicy *ClusterPodSecurityPolicy `json:"podSecurityPolicy,omitempty"` Prometheus *ClusterKubernetesPrometheus `json:"prometheus,omitempty"` Grafana *ClusterKubernetesGrafana `json:"grafana,omiempty"` Heapster *ClusterKubernetesHeapster `json:"heapster,omiempty"` InfluxDB *ClusterKubernetesInfluxDB `json:"influxDB,omiempty"` EncryptionProvider *ClusterEncryptionProvider `json:"encryptionProvider"` APIServer *ClusterKubernetesAPIServer `json:"apiServer,omitempty"` Kubelet *ClusterKubernetesKubelet `json:"kubelet,omitempty"` Scheduler *ClusterKubernetesScheduler `json:"scheduler,omitempty"` Proxy *ClusterKubernetesProxy `json:"proxy,omitempty"` ControllerManager *ClusterKubernetesControllerManager `json:"controllerManager,omitempty"` Calico *ClusterKubernetesCalico `json:"calico,omitempty"` GlobalFeatureGates map[string]bool `json:"globalFeatureGates,omitempty"` Hyperkube *bool `json:"hyperkube,omitempty"` } type ClusterKubernetesClusterAutoscaler struct { Enabled bool `json:"enabled,omitempty"` Image string `json:"image,omitempty"` Version string `json:"version,omitempty"` ScaleDownUtilizationThreshold *float64 `json:"scaleDownUtilizationThreshold,omitempty"` Overprovisioning *ClusterKubernetesClusterAutoscalerOverprovisioning `json:"overprovisioning,omitempty"` } type ClusterKubernetesClusterAutoscalerOverprovisioning struct { Enabled bool `json:"enabled,omitempty"` Image string `json:"image,omitempty"` Version string `json:"version,omitempty"` ReservedMillicoresPerReplica int `json:"reservedMillicoresPerReplica,omitempty"` ReservedMegabytesPerReplica int `json:"reservedMegabytesPerReplica,omitempty"` CoresPerReplica int `json:"coresPerReplica,omitempty"` NodesPerReplica int `json:"nodesPerReplica,omitempty"` ReplicaCount int `json:"replicaCount,omitempty"` } type ClusterKubernetesTiller struct { Enabled bool `json:"enabled,omitempty"` Image string `json:"image,omitempty"` Version string `json:"version,omitempty"` } type ClusterKubernetesDashboard struct { Enabled bool `json:"enabled,omitempty"` Image string `json:"image,omitempty"` Version string `json:"version,omitempty"` } type ClusterKubernetesAPIServer struct { // expose the API server through a public load balancer Public bool `json:"public,omitempty"` AllowCIDRs []string `json:"allowCIDRs,omitempty"` // create DNS record for the private load balancer, and optionally lock it down PrivateRecord bool `json:"privateRecord,omitempty"` PrivateAllowCIDRs []string `json:"privateAllowCIDRs,omitempty"` EnableAdmissionControllers []string `json:"enableAdmissionControllers,omitempty"` DisableAdmissionControllers []string `json:"disableAdmissionControllers,omitempty"` // OIDC OIDC *ClusterKubernetesAPIServerOIDC `json:"oidc,omitempty"` // AWS specific options Amazon *ClusterKubernetesAPIServerAmazon `json:"amazon,omitempty"` AuthTokenWebhookFile string `json:"authTokenWebhookFile,omitempty"` FeatureGates map[string]bool `json:"featureGates,omitempty"` } type ClusterKubernetesAPIServerOIDC struct { // The client ID for the OpenID Connect client, must be set if oidc-issuer-url is set. ClientID string `json:"clientID,omitempty" hiera:"kubernetes::apiserver::oidc_client_id"` // If provided, the name of a custom OpenID Connect claim for specifying // user groups. The claim value is expected to be a string or array of // strings. This flag is experimental, please see the authentication // documentation for further details. GroupsClaim string `json:"groupsClaim,omitempty" hiera:"kubernetes::apiserver::oidc_groups_claim"` // If provided, all groups will be prefixed with this value to prevent // conflicts with other authentication strategies. GroupsPrefix string `json:"groupsPrefix,omitempty" hiera:"kubernetes::apiserver::oidc_groups_prefix"` // The URL of the OpenID issuer, only HTTPS scheme will be accepted. If // set, it will be used to verify the OIDC JSON Web Token (JWT). IssuerURL string `json:"issuerURL,omitempty" hiera:"kubernetes::apiserver::oidc_issuer_url"` // Comma-separated list of allowed JOSE asymmetric signing algorithms. JWTs // with a 'alg' header value not in this list will be rejected. Values are // defined by RFC 7518 https://tools.ietf.org/html/rfc7518#section-3.1. // (default [RS256]) SigningAlgs []string `json:"signingAlgs,omitempty" hiera:"kubernetes::apiserver::oidc_signing_algs"` // The OpenID claim to use as the user name. Note that claims other than // the default ('sub') is not guaranteed to be unique and immutable. This // flag is experimental, please see the authentication documentation for // further details. (default "sub") UsernameClaim string `json:"usernameClaim,omitempty" hiera:"kubernetes::apiserver::oidc_username_claim"` // If provided, all usernames will be prefixed with this value. If not // provided, username claims other than 'email' are prefixed by the issuer // URL to avoid clashes. To skip any prefixing, provide the value '-'. UsernamePrefix string `json:"usernamePrefix,omitempty" hiera:"kubernetes::apiserver::oidc_username_prefix"` } type ClusterKubernetesAPIServerAmazon struct { PublicELBAccessLogs *ClusterKubernetesAPIServerAmazonAccessLogs `json:"publicELBAccessLogs,omitempty"` InternalELBAccessLogs *ClusterKubernetesAPIServerAmazonAccessLogs `json:"internalELBAccessLogs,omitempty"` AwsIAMAuthenticatorInit bool `json:"awsIAMAuthenticatorInit,omitempty"` } type ClusterKubernetesAPIServerAmazonAccessLogs struct { Enabled *bool `json:"enabled,omitempty"` Bucket string `json:"bucket,omitempty"` BucketPrefix string `json:"bucketPrefix,omitempty"` Interval *int `json:"interval,omitempty"` } type ClusterPodSecurityPolicy struct { Enabled bool `json:"enabled,omitempty"` } type ClusterEncryptionProvider struct { Enabled bool `json:"enabled,omitempty"` Version string `json:"version,omitempty"` } // Configure the cluster internal deployment of prometheus type ClusterKubernetesPrometheus struct { // Enable a cluster internal prometheus deployment, default: true Enabled bool `json:"enabled,omitempty"` // Mode defines which components are installed Mode string `json:"mode,omitempty"` } type ClusterKubernetesGrafana struct { Enabled bool `json:"enabled,omitempty"` } type ClusterKubernetesHeapster struct { Enabled bool `json:"enabled,omitempty"` } type ClusterKubernetesInfluxDB struct { Enabled bool `json:"enabled,omitempty"` } type ClusterVaultHelper struct { URL string `json:"url,omitempty"` } type ClusterKubernetesScheduler struct { FeatureGates map[string]bool `json:"featureGates,omitempty"` } type ClusterKubernetesKubelet struct { FeatureGates map[string]bool `json:"featureGates,omitempty"` } type ClusterKubernetesProxy struct { FeatureGates map[string]bool `json:"featureGates,omitempty"` } type ClusterKubernetesControllerManager struct { FeatureGates map[string]bool `json:"featureGates,omitempty"` } type ClusterKubernetesCalicoBackend string type ClusterKubernetesCalico struct { Backend ClusterKubernetesCalicoBackend `json:"backend"` EnableTypha bool `json:"enableTypha"` TyphaReplicas *int `json:"typhaReplicas"` } // +k8s:openapi-gen=true // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object type ClusterList struct { metav1.TypeMeta `json:",inline"` metav1.ObjectMeta `json:"metadata,omitempty"` Items []Cluster `json:"items"` } func N
ewCluster(
identifier_name
cluster.go
Azure = "azure" CloudGoogle = "google" CloudBaremetal = "baremetal" CloudDigitalOcean = "digitalocean" ) const ( ClusterTypeHub = "hub" ClusterTypeClusterSingle = "cluster-single" ClusterTypeClusterMulti = "cluster-multi" ) const ( // represents Terraform in a destroy state StateDestroy = "destroy" ) const ( PrometheusModeFull = "Full" PrometheusModeExternalExportersOnly = "ExternalExportersOnly" PrometheusModeExternalScrapeTargetsOnly = "ExternalScrapeTargetsOnly" ) const ( CalicoBackendEtcd ClusterKubernetesCalicoBackend = "etcd" CalicoBackendKubernetes ClusterKubernetesCalicoBackend = "kubernetes" ) // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object // +k8s:openapi-gen=true // +resource:path=clusters type Cluster struct { metav1.TypeMeta `json:",inline"` metav1.ObjectMeta `json:"metadata,omitempty"` CloudId string `json:"cloudId,omitempty"` InstancePools []InstancePool `json:"instancePools,omitempty"` Cloud string `json:"cloud,omitempty"` Location string `json:"location,omitempty"` Network *Network `json:"network,omitempty"` LoggingSinks []*LoggingSink `json:"loggingSinks,omitempty"` Values *Values `json:"values,omitempty"` KubernetesAPI *KubernetesAPI `json:"kubernetesAPI,omitempty"` GroupIdentifier string `json:"groupIdentifier,omitempty"` VaultHelper *ClusterVaultHelper `json:"vaultHelper,omitempty"` Environment string `json:"environment,omitempty"` Kubernetes *ClusterKubernetes `json:"kubernetes,omitempty"` Type string `json:"-"` // This specifies if a cluster is a hub, single or multi // Amazon specific options Amazon *ClusterAmazon `json:"amazon"` } // ClusterAmazon offers Amazon-specific settings for that instance pool type ClusterAmazon struct { // This fields contains ARNs for additional IAM policies to be added to // this instance pool AdditionalIAMPolicies []string `json:"additionalIAMPolicies,omitempty"` // When set to true, AWS Elastic Block Storage volumes are encrypted EBSEncrypted *bool `json:"ebsEncrypted,omitempty"` } type ClusterKubernetes struct { Zone string `json:"zone,omitempty"` Version string `json:"version,omitempty"` PodCIDR string `json:"podCIDR,omitempty"` ServiceCIDR string `json:"serviceCIDR,omitempty"` ClusterAutoscaler *ClusterKubernetesClusterAutoscaler `json:"clusterAutoscaler,omitempty"` Tiller *ClusterKubernetesTiller `json:"tiller,omitempty"` Dashboard *ClusterKubernetesDashboard `json:"dashboard,omitempty"` PodSecurityPolicy *ClusterPodSecurityPolicy `json:"podSecurityPolicy,omitempty"` Prometheus *ClusterKubernetesPrometheus `json:"prometheus,omitempty"` Grafana *ClusterKubernetesGrafana `json:"grafana,omiempty"` Heapster *ClusterKubernetesHeapster `json:"heapster,omiempty"` InfluxDB *ClusterKubernetesInfluxDB `json:"influxDB,omiempty"` EncryptionProvider *ClusterEncryptionProvider `json:"encryptionProvider"` APIServer *ClusterKubernetesAPIServer `json:"apiServer,omitempty"` Kubelet *ClusterKubernetesKubelet `json:"kubelet,omitempty"` Scheduler *ClusterKubernetesScheduler `json:"scheduler,omitempty"` Proxy *ClusterKubernetesProxy `json:"proxy,omitempty"` ControllerManager *ClusterKubernetesControllerManager `json:"controllerManager,omitempty"` Calico *ClusterKubernetesCalico `json:"calico,omitempty"` GlobalFeatureGates map[string]bool `json:"globalFeatureGates,omitempty"` Hyperkube *bool `json:"hyperkube,omitempty"` } type ClusterKubernetesClusterAutoscaler struct { Enabled bool `json:"enabled,omitempty"` Image string `json:"image,omitempty"` Version string `json:"version,omitempty"` ScaleDownUtilizationThreshold *float64 `json:"scaleDownUtilizationThreshold,omitempty"` Overprovisioning *ClusterKubernetesClusterAutoscalerOverprovisioning `json:"overprovisioning,omitempty"` } type ClusterKubernetesClusterAutoscalerOverprovisioning struct { Enabled bool `json:"enabled,omitempty"` Image string `json:"image,omitempty"` Version string `json:"version,omitempty"` ReservedMillicoresPerReplica int `json:"reservedMillicoresPerReplica,omitempty"` ReservedMegabytesPerReplica int `json:"reservedMegabytesPerReplica,omitempty"` CoresPerReplica int `json:"coresPerReplica,omitempty"` NodesPerReplica int `json:"nodesPerReplica,omitempty"` ReplicaCount int `json:"replicaCount,omitempty"` } type ClusterKubernetesTiller struct { Enabled bool `json:"enabled,omitempty"` Image string `json:"image,omitempty"` Version string `json:"version,omitempty"` } type ClusterKubernetesDashboard struct { Enabled bool `json:"enabled,omitempty"`
// expose the API server through a public load balancer Public bool `json:"public,omitempty"` AllowCIDRs []string `json:"allowCIDRs,omitempty"` // create DNS record for the private load balancer, and optionally lock it down PrivateRecord bool `json:"privateRecord,omitempty"` PrivateAllowCIDRs []string `json:"privateAllowCIDRs,omitempty"` EnableAdmissionControllers []string `json:"enableAdmissionControllers,omitempty"` DisableAdmissionControllers []string `json:"disableAdmissionControllers,omitempty"` // OIDC OIDC *ClusterKubernetesAPIServerOIDC `json:"oidc,omitempty"` // AWS specific options Amazon *ClusterKubernetesAPIServerAmazon `json:"amazon,omitempty"` AuthTokenWebhookFile string `json:"authTokenWebhookFile,omitempty"` FeatureGates map[string]bool `json:"featureGates,omitempty"` } type ClusterKubernetesAPIServerOIDC struct { // The client ID for the OpenID Connect client, must be set if oidc-issuer-url is set. ClientID string `json:"clientID,omitempty" hiera:"kubernetes::apiserver::oidc_client_id"` // If provided, the name of a custom OpenID Connect claim for specifying // user groups. The claim value is expected to be a string or array of // strings. This flag is experimental, please see the authentication // documentation for further details. GroupsClaim string `json:"groupsClaim,omitempty" hiera:"kubernetes::apiserver::oidc_groups_claim"` // If provided, all groups will be prefixed with this value to prevent // conflicts with other authentication strategies. GroupsPrefix string `json:"groupsPrefix,omitempty" hiera:"kubernetes::apiserver::oidc_groups_prefix"` // The URL of the OpenID issuer, only HTTPS scheme will be accepted. If // set, it will be used to verify the OIDC JSON Web Token (JWT). IssuerURL string `json:"issuerURL,omitempty" hiera:"kubernetes::apiserver::oidc_issuer_url"` // Comma-separated list of allowed JOSE asymmetric signing algorithms. JWTs // with a 'alg' header value not in this list will be rejected. Values are // defined by RFC 7518 https://tools.ietf.org/html/rfc7518#section-3.1. // (default [RS256]) SigningAlgs []string `json:"signingAlgs,omitempty" hiera:"kubernetes::apiserver::oidc_signing_algs"` // The OpenID claim to use as the user name. Note that claims other than // the default ('sub') is not guaranteed to be unique and immutable. This // flag is experimental, please see the authentication documentation for // further details. (default "sub") UsernameClaim string `json:"usernameClaim,omitempty" hiera:"kubernetes::apiserver::oidc_username_claim"` // If provided, all usernames will be prefixed with this value. If not // provided, username claims other than 'email' are prefixed by the issuer // URL to avoid clashes. To skip any prefixing, provide the value '-'. UsernamePrefix string `json:"usernamePrefix,omitempty" hiera:"kubernetes::apiserver::oidc_username_prefix"` } type ClusterKubernetesAPIServerAmazon struct { PublicELBAccessLogs *ClusterKubernetesAPIServerAmazonAccessLogs `json:"publicELBAccessLogs,omitempty"` InternalELBAccessLogs *ClusterKubernetesAPIServerAmazonAccessLogs `json:"internalELBAccessLogs,omitempty"` AwsIAMAuthenticatorInit bool `json:"awsIAMAuthenticatorInit,omitempty"` } type ClusterKubernetesAPIServerAmazonAccessLogs struct { Enabled *bool `json:"enabled,omitempty"` Bucket string `json:"bucket,omitempty"` BucketPrefix string `json:"bucketPrefix,omitempty"` Interval *int `json:"interval,omitempty"` } type ClusterPod
Image string `json:"image,omitempty"` Version string `json:"version,omitempty"` } type ClusterKubernetesAPIServer struct {
random_line_split
orderbook.go
"suspended" } func updatePos(ord *Order) { securityId := ord.Security.Id p := getPos(ord.Acc, securityId) var outstand *float64 if ord.Side == "buy" { outstand = &p.OutstandBuyQty } else { outstand = &p.OutstandSellQty } switch ord.St { case "unconfirmed", "unconfirmed_replace": *outstand += ord.Qty - ord.CumQty case "filled", "partial": if ord.LastQty > 0 && ord.Type != "otc" { *outstand -= ord.LastQty if *outstand < 0 { log.Printf("Outstand < 0: %s", ord) *outstand = 0 } } qty := ord.LastQty if ord.Side == "buy" { p.BuyQty += ord.LastQty p.BuyValue += ord.LastQty * ord.LastPx } else { qty = -qty p.SellQty += ord.LastQty p.SellValue += ord.LastQty * ord.LastPx } qty0 := p.Qty px := ord.LastPx multiplier := ord.Security.Rate * ord.Security.Multiplier if (qty0 > 0) && (qty < 0) { // sell trade to cover position if qty0 > -qty { p.RealizedPnl += (px - p.AvgPx) * -qty * multiplier } else { p.RealizedPnl += (px - p.AvgPx) * qty0 * multiplier p.AvgPx = px } } else if (qty0 < 0) && (qty > 0) { // buy trade to cover position if -qty0 > qty { p.RealizedPnl += (p.AvgPx - px) * qty * multiplier } else { p.RealizedPnl += (p.AvgPx - px) * -qty0 * multiplier p.AvgPx = px } } else { // open position p.AvgPx = (qty0*p.AvgPx + qty*px) / (qty0 + qty) } p.Qty += qty if p.Qty == 0 { p.AvgPx = 0 } default: *outstand -= ord.Qty - ord.CumQty if *outstand < 0 { log.Printf("Outstand < 0: %s", ord) *outstand = 0 } } } var seqNum int64 = 0 var offlineDone = false var onlineCache [][]interface{} func ParseOffline(msg []interface{}) { if msg[1].(string) == "complete" { for _, msg := range onlineCache { ParseOrder(msg, false) } onlineCache = onlineCache[:0] offlineDone = true log.Print("offline done") } } func ParseOrder(msg []interface{}, isOnline bool) { if isOnline && !offlineDone { onlineCache = append(onlineCache, msg) return } clOrdId := int64(msg[1].(float64)) // tm := int64(msg[2].(float64)) seq := int64(msg[3].(float64)) if seq <= seqNum { return } seqNum = seq switch st := msg[4].(string); st { case "unconfirmed", "unconfirmed_replace": securityId := int64(msg[5].(float64)) security := SecurityMapById[securityId] if security == nil { log.Println("not found security", securityId) return } // aid := int(msg[6].(float64)) // userId := int(msg[7].(float64)) acc := int(msg[8].(float64)) // brokerAcc := int(msg[9].(float65)) qty := msg[10].(float64) px := msg[11].(float64) side := msg[12].(string) // ordType := msg[13].(string) // tif := msg[14].(string) ord := Order{ Id: clOrdId, St: st, Security: security, Acc: acc, Qty: qty, Px: px, Side: side, } if st == "unconfirmed_replace" { origClOrdId := int64(msg[14].(float64)) ord.OrigClOrdId = origClOrdId } orders[clOrdId] = &ord updatePos(&ord) case "filled", "partial": qty := msg[5].(float64) px := msg[6].(float64) // tradeId = msg[7].(string) execTransType := msg[8].(string) if execTransType == "cancel" { qty = -qty } ord := orders[clOrdId] if ord != nil { ord.AvgPx = (ord.CumQty*ord.AvgPx + qty*px) / (ord.CumQty + qty) ord.CumQty += qty ord.CumQty = math.Round(ord.CumQty*1e6) / 1e6 if ord.CumQty > ord.Qty { log.Printf("overfill found: %s", msg) } ord.LastQty = qty ord.LastPx = px if ord.CumQty >= ord.Qty { st = "filled" } else { st = "partial" } ord.St = st updatePos(ord) } else { log.Println("not found order for", clOrdId) } case "canceled": case "cancelled": case "expired": case "done_for_day": case "calculated": ord := orders[clOrdId] if ord != nil { st0 := ord.St ord.St = st if isLive(st0) { updatePos(ord) } } else { log.Println("can not find order for", clOrdId) } case "new", "pending", "replaced", "suspended": ord := orders[clOrdId] if ord != nil { if st == "replaced" { old := orders[ord.OrigClOrdId] if old == nil { log.Println("can not find order for", ord.OrigClOrdId) } else { old.St = st } st = "confirmed" } ord.St = st } else { log.Println("can not find order for", clOrdId) } case "new_rejected", "replace_rejected": ord := orders[clOrdId] if ord != nil { st0 := ord.St ord.St = st if isLive(st0) { updatePos(ord) } } else { log.Println("can not find order for", clOrdId) } case "risk_rejected": ord := orders[clOrdId] if ord != nil { ord.St = st updatePos(ord) } } } func ParseTarget(msg []interface{}) { acc := int(msg[1].(float64)) // msg[2] is acc name for _, v := range Positions[acc] { v.Target = 0. } if len(msg) < 4 { return } if _, ok := msg[3].([]interface{}); !ok { return } for _, v := range msg[3].([]interface{}) { t := v.([]interface{}) securityId := int64(t[0].(float64)) target := t[1].(float64) getPos(acc, securityId).Target = target } } func ParseBod(msg []interface{}) { acc := int(msg[1].(float64)) securityId := int64(msg[2].(float64)) qty := msg[3].(float64) avgPx := msg[4].(float64) commission := msg[5].(float64) realizedPnl := msg[6].(float64) // brokerAcc := int(msg[7]) // tm := int64(msg[8].(float64)) p := getPos(acc, securityId) p.Qty = qty p.AvgPx = avgPx p.Commission = commission p.RealizedPnl = realizedPnl p.Bod.Qty = qty p.Bod.AvgPx = avgPx p.Bod.Commission = commission p.Bod.RealizedPnl = realizedPnl } func
ParsePnl
identifier_name
orderbook.go
tmp = make(map[int64]*Position) Positions[acc] = tmp } p := tmp[securityId] if p == nil { p = &Position{} p.Acc = acc p.Security = SecurityMapById[securityId] if p.Security == nil { log.Println("unknown securityId", securityId) return p } tmp[securityId] = p used := usedSecurities[securityId] if !used { Request([]interface{}{"sub", securityId}) usedSecurities[securityId] = true } } return p } func isLive(st string) bool { if st == "" { return true } st = strings.ToLower(st) return strings.HasPrefix(st, "pending") || strings.HasPrefix(st, "unconfirmed") || strings.HasPrefix(st, "partial") || st == "new" || st == "suspended" } func updatePos(ord *Order) { securityId := ord.Security.Id p := getPos(ord.Acc, securityId) var outstand *float64 if ord.Side == "buy" { outstand = &p.OutstandBuyQty } else { outstand = &p.OutstandSellQty } switch ord.St { case "unconfirmed", "unconfirmed_replace": *outstand += ord.Qty - ord.CumQty case "filled", "partial": if ord.LastQty > 0 && ord.Type != "otc" { *outstand -= ord.LastQty if *outstand < 0 { log.Printf("Outstand < 0: %s", ord) *outstand = 0 } } qty := ord.LastQty if ord.Side == "buy" { p.BuyQty += ord.LastQty p.BuyValue += ord.LastQty * ord.LastPx } else { qty = -qty p.SellQty += ord.LastQty p.SellValue += ord.LastQty * ord.LastPx } qty0 := p.Qty px := ord.LastPx multiplier := ord.Security.Rate * ord.Security.Multiplier if (qty0 > 0) && (qty < 0) { // sell trade to cover position if qty0 > -qty { p.RealizedPnl += (px - p.AvgPx) * -qty * multiplier } else { p.RealizedPnl += (px - p.AvgPx) * qty0 * multiplier p.AvgPx = px } } else if (qty0 < 0) && (qty > 0) { // buy trade to cover position if -qty0 > qty { p.RealizedPnl += (p.AvgPx - px) * qty * multiplier } else { p.RealizedPnl += (p.AvgPx - px) * -qty0 * multiplier p.AvgPx = px } } else { // open position p.AvgPx = (qty0*p.AvgPx + qty*px) / (qty0 + qty) } p.Qty += qty if p.Qty == 0 { p.AvgPx = 0 } default: *outstand -= ord.Qty - ord.CumQty if *outstand < 0 { log.Printf("Outstand < 0: %s", ord) *outstand = 0 } } } var seqNum int64 = 0 var offlineDone = false var onlineCache [][]interface{} func ParseOffline(msg []interface{}) { if msg[1].(string) == "complete" { for _, msg := range onlineCache { ParseOrder(msg, false) } onlineCache = onlineCache[:0] offlineDone = true log.Print("offline done") } } func ParseOrder(msg []interface{}, isOnline bool) { if isOnline && !offlineDone { onlineCache = append(onlineCache, msg) return } clOrdId := int64(msg[1].(float64)) // tm := int64(msg[2].(float64)) seq := int64(msg[3].(float64)) if seq <= seqNum { return } seqNum = seq switch st := msg[4].(string); st { case "unconfirmed", "unconfirmed_replace": securityId := int64(msg[5].(float64)) security := SecurityMapById[securityId] if security == nil { log.Println("not found security", securityId) return } // aid := int(msg[6].(float64)) // userId := int(msg[7].(float64)) acc := int(msg[8].(float64)) // brokerAcc := int(msg[9].(float65)) qty := msg[10].(float64) px := msg[11].(float64) side := msg[12].(string) // ordType := msg[13].(string) // tif := msg[14].(string) ord := Order{ Id: clOrdId, St: st, Security: security, Acc: acc, Qty: qty, Px: px, Side: side, } if st == "unconfirmed_replace" { origClOrdId := int64(msg[14].(float64)) ord.OrigClOrdId = origClOrdId } orders[clOrdId] = &ord updatePos(&ord) case "filled", "partial": qty := msg[5].(float64) px := msg[6].(float64) // tradeId = msg[7].(string) execTransType := msg[8].(string) if execTransType == "cancel" { qty = -qty } ord := orders[clOrdId] if ord != nil { ord.AvgPx = (ord.CumQty*ord.AvgPx + qty*px) / (ord.CumQty + qty) ord.CumQty += qty ord.CumQty = math.Round(ord.CumQty*1e6) / 1e6 if ord.CumQty > ord.Qty { log.Printf("overfill found: %s", msg) } ord.LastQty = qty ord.LastPx = px if ord.CumQty >= ord.Qty { st = "filled" } else { st = "partial" } ord.St = st updatePos(ord) } else { log.Println("not found order for", clOrdId) } case "canceled": case "cancelled": case "expired": case "done_for_day": case "calculated": ord := orders[clOrdId] if ord != nil { st0 := ord.St ord.St = st if isLive(st0) { updatePos(ord) } } else { log.Println("can not find order for", clOrdId) } case "new", "pending", "replaced", "suspended": ord := orders[clOrdId] if ord != nil { if st == "replaced" { old := orders[ord.OrigClOrdId] if old == nil { log.Println("can not find order for", ord.OrigClOrdId) } else { old.St = st } st = "confirmed" } ord.St = st } else { log.Println("can not find order for", clOrdId) } case "new_rejected", "replace_rejected": ord := orders[clOrdId] if ord != nil { st0 := ord.St ord.St = st if isLive(st0) { updatePos(ord) } } else { log.Println("can not find order for", clOrdId) } case "risk_rejected": ord := orders[clOrdId] if ord != nil { ord.St = st updatePos(ord) } } } func ParseTarget(msg []interface{})
{ acc := int(msg[1].(float64)) // msg[2] is acc name for _, v := range Positions[acc] { v.Target = 0. } if len(msg) < 4 { return } if _, ok := msg[3].([]interface{}); !ok { return } for _, v := range msg[3].([]interface{}) { t := v.([]interface{}) securityId := int64(t[0].(float64)) target := t[1].(float64) getPos(acc, securityId).Target = target } }
identifier_body
orderbook.go
Sector: msg[13].(string), IndustryGroup: msg[14].(string), Industry: msg[15].(string), SubIndustry: msg[16].(string), Bbgid: msg[17].(string), Cusip: msg[18].(string), Sedol: msg[19].(string), Isin: msg[20].(string), } if sec.Market == "CURRENCY"
if sec.Multiplier <= 0 { sec.Multiplier = 1 } if sec.Rate <= 0 { sec.Rate = 1 } sec0 := SecurityMapById[sec.Id] if sec0 == nil { SecurityMapById[sec.Id] = sec } else { *sec0 = *sec sec = sec0 } tmp := SecurityMapByMarket[sec.Market] if tmp == nil { tmp = make(map[string]*Security) SecurityMapByMarket[sec.Market] = tmp } tmp[sec.Symbol] = sec } type Order struct { Id int64 OrigClOrdId int64 // Tm int64 // Seq int64 St string Security *Security // UserId int Acc int Qty float64 Px float64 Side string Type string // Tif string CumQty float64 AvgPx float64 LastQty float64 LastPx float64 } var orders = make(map[int64]*Order) type PositionBase struct { Qty float64 AvgPx float64 Commission float64 RealizedPnl float64 } type Position struct { PositionBase Bod PositionBase OutstandBuyQty float64 OutstandSellQty float64 BuyQty float64 BuyValue float64 SellQty float64 SellValue float64 Security *Security Acc int Target float64 } var Positions = make(map[int]map[int64]*Position) var usedSecurities = make(map[int64]bool) func getPos(acc int, securityId int64) *Position { tmp := Positions[acc] if tmp == nil { tmp = make(map[int64]*Position) Positions[acc] = tmp } p := tmp[securityId] if p == nil { p = &Position{} p.Acc = acc p.Security = SecurityMapById[securityId] if p.Security == nil { log.Println("unknown securityId", securityId) return p } tmp[securityId] = p used := usedSecurities[securityId] if !used { Request([]interface{}{"sub", securityId}) usedSecurities[securityId] = true } } return p } func isLive(st string) bool { if st == "" { return true } st = strings.ToLower(st) return strings.HasPrefix(st, "pending") || strings.HasPrefix(st, "unconfirmed") || strings.HasPrefix(st, "partial") || st == "new" || st == "suspended" } func updatePos(ord *Order) { securityId := ord.Security.Id p := getPos(ord.Acc, securityId) var outstand *float64 if ord.Side == "buy" { outstand = &p.OutstandBuyQty } else { outstand = &p.OutstandSellQty } switch ord.St { case "unconfirmed", "unconfirmed_replace": *outstand += ord.Qty - ord.CumQty case "filled", "partial": if ord.LastQty > 0 && ord.Type != "otc" { *outstand -= ord.LastQty if *outstand < 0 { log.Printf("Outstand < 0: %s", ord) *outstand = 0 } } qty := ord.LastQty if ord.Side == "buy" { p.BuyQty += ord.LastQty p.BuyValue += ord.LastQty * ord.LastPx } else { qty = -qty p.SellQty += ord.LastQty p.SellValue += ord.LastQty * ord.LastPx } qty0 := p.Qty px := ord.LastPx multiplier := ord.Security.Rate * ord.Security.Multiplier if (qty0 > 0) && (qty < 0) { // sell trade to cover position if qty0 > -qty { p.RealizedPnl += (px - p.AvgPx) * -qty * multiplier } else { p.RealizedPnl += (px - p.AvgPx) * qty0 * multiplier p.AvgPx = px } } else if (qty0 < 0) && (qty > 0) { // buy trade to cover position if -qty0 > qty { p.RealizedPnl += (p.AvgPx - px) * qty * multiplier } else { p.RealizedPnl += (p.AvgPx - px) * -qty0 * multiplier p.AvgPx = px } } else { // open position p.AvgPx = (qty0*p.AvgPx + qty*px) / (qty0 + qty) } p.Qty += qty if p.Qty == 0 { p.AvgPx = 0 } default: *outstand -= ord.Qty - ord.CumQty if *outstand < 0 { log.Printf("Outstand < 0: %s", ord) *outstand = 0 } } } var seqNum int64 = 0 var offlineDone = false var onlineCache [][]interface{} func ParseOffline(msg []interface{}) { if msg[1].(string) == "complete" { for _, msg := range onlineCache { ParseOrder(msg, false) } onlineCache = onlineCache[:0] offlineDone = true log.Print("offline done") } } func ParseOrder(msg []interface{}, isOnline bool) { if isOnline && !offlineDone { onlineCache = append(onlineCache, msg) return } clOrdId := int64(msg[1].(float64)) // tm := int64(msg[2].(float64)) seq := int64(msg[3].(float64)) if seq <= seqNum { return } seqNum = seq switch st := msg[4].(string); st { case "unconfirmed", "unconfirmed_replace": securityId := int64(msg[5].(float64)) security := SecurityMapById[securityId] if security == nil { log.Println("not found security", securityId) return } // aid := int(msg[6].(float64)) // userId := int(msg[7].(float64)) acc := int(msg[8].(float64)) // brokerAcc := int(msg[9].(float65)) qty := msg[10].(float64) px := msg[11].(float64) side := msg[12].(string) // ordType := msg[13].(string) // tif := msg[14].(string) ord := Order{ Id: clOrdId, St: st, Security: security, Acc: acc, Qty: qty, Px: px, Side: side, } if st == "unconfirmed_replace" { origClOrdId := int64(msg[14].(float64)) ord.OrigClOrdId = origClOrdId } orders[clOrdId] = &ord updatePos(&ord) case "filled", "partial": qty := msg[5].(float64) px := msg[6].(float64) // tradeId = msg[7].(string) execTransType := msg[8].(string) if execTransType == "cancel" { qty = -qty } ord := orders[clOrdId] if ord != nil { ord.AvgPx = (ord.CumQty*ord.AvgPx + qty*px) / (ord.CumQty + qty) ord.CumQty += qty ord.CumQty = math.Round(ord.CumQty*1e6) / 1e6 if ord.CumQty > ord.Qty { log.Printf("overfill found: %s",
{ sec.Market = "FX" }
conditional_block
orderbook.go
64 St string Security *Security // UserId int Acc int Qty float64 Px float64 Side string Type string // Tif string CumQty float64 AvgPx float64 LastQty float64 LastPx float64 } var orders = make(map[int64]*Order) type PositionBase struct { Qty float64 AvgPx float64 Commission float64 RealizedPnl float64 } type Position struct { PositionBase Bod PositionBase OutstandBuyQty float64 OutstandSellQty float64 BuyQty float64 BuyValue float64 SellQty float64 SellValue float64 Security *Security Acc int Target float64 } var Positions = make(map[int]map[int64]*Position) var usedSecurities = make(map[int64]bool) func getPos(acc int, securityId int64) *Position { tmp := Positions[acc] if tmp == nil { tmp = make(map[int64]*Position) Positions[acc] = tmp } p := tmp[securityId] if p == nil { p = &Position{} p.Acc = acc p.Security = SecurityMapById[securityId] if p.Security == nil { log.Println("unknown securityId", securityId) return p } tmp[securityId] = p used := usedSecurities[securityId] if !used { Request([]interface{}{"sub", securityId}) usedSecurities[securityId] = true } } return p } func isLive(st string) bool { if st == "" { return true } st = strings.ToLower(st) return strings.HasPrefix(st, "pending") || strings.HasPrefix(st, "unconfirmed") || strings.HasPrefix(st, "partial") || st == "new" || st == "suspended" } func updatePos(ord *Order) { securityId := ord.Security.Id p := getPos(ord.Acc, securityId) var outstand *float64 if ord.Side == "buy" { outstand = &p.OutstandBuyQty } else { outstand = &p.OutstandSellQty } switch ord.St { case "unconfirmed", "unconfirmed_replace": *outstand += ord.Qty - ord.CumQty case "filled", "partial": if ord.LastQty > 0 && ord.Type != "otc" { *outstand -= ord.LastQty if *outstand < 0 { log.Printf("Outstand < 0: %s", ord) *outstand = 0 } } qty := ord.LastQty if ord.Side == "buy" { p.BuyQty += ord.LastQty p.BuyValue += ord.LastQty * ord.LastPx } else { qty = -qty p.SellQty += ord.LastQty p.SellValue += ord.LastQty * ord.LastPx } qty0 := p.Qty px := ord.LastPx multiplier := ord.Security.Rate * ord.Security.Multiplier if (qty0 > 0) && (qty < 0) { // sell trade to cover position if qty0 > -qty { p.RealizedPnl += (px - p.AvgPx) * -qty * multiplier } else { p.RealizedPnl += (px - p.AvgPx) * qty0 * multiplier p.AvgPx = px } } else if (qty0 < 0) && (qty > 0) { // buy trade to cover position if -qty0 > qty { p.RealizedPnl += (p.AvgPx - px) * qty * multiplier } else { p.RealizedPnl += (p.AvgPx - px) * -qty0 * multiplier p.AvgPx = px } } else { // open position p.AvgPx = (qty0*p.AvgPx + qty*px) / (qty0 + qty) } p.Qty += qty if p.Qty == 0 { p.AvgPx = 0 } default: *outstand -= ord.Qty - ord.CumQty if *outstand < 0 { log.Printf("Outstand < 0: %s", ord) *outstand = 0 } } } var seqNum int64 = 0 var offlineDone = false var onlineCache [][]interface{} func ParseOffline(msg []interface{}) { if msg[1].(string) == "complete" { for _, msg := range onlineCache { ParseOrder(msg, false) } onlineCache = onlineCache[:0] offlineDone = true log.Print("offline done") } } func ParseOrder(msg []interface{}, isOnline bool) { if isOnline && !offlineDone { onlineCache = append(onlineCache, msg) return } clOrdId := int64(msg[1].(float64)) // tm := int64(msg[2].(float64)) seq := int64(msg[3].(float64)) if seq <= seqNum { return } seqNum = seq switch st := msg[4].(string); st { case "unconfirmed", "unconfirmed_replace": securityId := int64(msg[5].(float64)) security := SecurityMapById[securityId] if security == nil { log.Println("not found security", securityId) return } // aid := int(msg[6].(float64)) // userId := int(msg[7].(float64)) acc := int(msg[8].(float64)) // brokerAcc := int(msg[9].(float65)) qty := msg[10].(float64) px := msg[11].(float64) side := msg[12].(string) // ordType := msg[13].(string) // tif := msg[14].(string) ord := Order{ Id: clOrdId, St: st, Security: security, Acc: acc, Qty: qty, Px: px, Side: side, } if st == "unconfirmed_replace" { origClOrdId := int64(msg[14].(float64)) ord.OrigClOrdId = origClOrdId } orders[clOrdId] = &ord updatePos(&ord) case "filled", "partial": qty := msg[5].(float64) px := msg[6].(float64) // tradeId = msg[7].(string) execTransType := msg[8].(string) if execTransType == "cancel" { qty = -qty } ord := orders[clOrdId] if ord != nil { ord.AvgPx = (ord.CumQty*ord.AvgPx + qty*px) / (ord.CumQty + qty) ord.CumQty += qty ord.CumQty = math.Round(ord.CumQty*1e6) / 1e6 if ord.CumQty > ord.Qty { log.Printf("overfill found: %s", msg) } ord.LastQty = qty ord.LastPx = px if ord.CumQty >= ord.Qty { st = "filled" } else { st = "partial" } ord.St = st updatePos(ord) } else { log.Println("not found order for", clOrdId) } case "canceled": case "cancelled": case "expired": case "done_for_day": case "calculated": ord := orders[clOrdId] if ord != nil { st0 := ord.St ord.St = st if isLive(st0) { updatePos(ord) } } else { log.Println("can not find order for", clOrdId) } case "new", "pending", "replaced", "suspended": ord := orders[clOrdId] if ord != nil { if st == "replaced" { old := orders[ord.OrigClOrdId] if old == nil { log.Println("can not find order for", ord.OrigClOrdId) } else { old.St = st } st = "confirmed" } ord.St = st } else { log.Println("can not find order for", clOrdId) }
random_line_split
main.rs
_lowercase().as_str() { "y" => break, "n" => return, _ => println!("Unknown response: {}", answer), } } for scan_position in config.scan_positions() { println!("Colorizing {}:", scan_position.name); let translations = config.translations(scan_position); if translations.is_empty() { println!(" - No translations found"); } else { for translation in translations { println!( " - Translation:\n - Infile: {}\n - Outfile: {}", translation.infile.display(), translation.outfile.display() ); config.colorize(scan_position, &translation); } } } println!("Complete!"); } struct Config { image_dir: PathBuf, keep_without_thermal: bool, las_dir: PathBuf, max_reflectance: f32, min_reflectance: f32, project: Project, rotate: bool, scan_position_names: Option<Vec<String>>, sync_to_pps: bool, temperature_gradient: Gradient<Rgb>, use_scanpos_names: bool, name_map: NameMap, } struct ImageGroup<'a> { camera_calibration: &'a CameraCalibration, image: &'a Image, irb: Irb, irb_path: PathBuf, mount_calibration: &'a MountCalibration, rotate: bool, } struct Translation { infile: PathBuf, outfile: PathBuf, } #[derive(Debug, Default, Deserialize)] struct NameMap { maps: Vec<FromTo>, } #[derive(Debug, Default, Deserialize)] struct FromTo { from: String, to: String, } impl Config { fn new(matches: &ArgMatches) -> Config { use std::fs::File; use std::io::Read; use toml; let project = Project::from_path(matches.value_of("PROJECT").unwrap()).unwrap(); let image_dir = PathBuf::from(matches.value_of("IMAGE_DIR").unwrap()); let las_dir = Path::new(matches.value_of("LAS_DIR").unwrap()).to_path_buf(); let min_reflectance = value_t!(matches, "min-reflectance", f32).unwrap(); let max_reflectance = value_t!(matches, "max-reflectance", f32).unwrap(); let min_temperature = value_t!(matches, "min-temperature", f32).unwrap(); let max_temperature = value_t!(matches, "max-temperature", f32).unwrap(); let min_temperature_color = Rgb::new(0.0, 0., 1.0); let max_temperature_color = Rgb::new(1.0, 0., 0.); let temperature_gradient = Gradient::with_domain(vec![ (min_temperature, min_temperature_color), (max_temperature, max_temperature_color), ]); let name_map = if let Some(name_map) = matches.value_of("name-map") { let mut s = String::new(); File::open(name_map) .unwrap() .read_to_string(&mut s) .unwrap(); toml::from_str(&s).unwrap() } else { NameMap::default() }; Config { image_dir: image_dir, keep_without_thermal: matches.is_present("keep-without-thermal"), las_dir: las_dir, max_reflectance: max_reflectance, min_reflectance: min_reflectance, project: project, rotate: matches.is_present("rotate"), scan_position_names: matches.values_of("scan-position").map(|values| { values.map(|name| name.to_string()).collect() }), sync_to_pps: matches.is_present("sync-to-pps"), temperature_gradient: temperature_gradient, use_scanpos_names: matches.is_present("use-scanpos-names"), name_map: name_map, } } fn translations(&self, scan_position: &ScanPosition) -> Vec<Translation> { let paths = scan_position.singlescan_rxp_paths(&self.project); if self.use_scanpos_names && paths.len() > 1 { panic!( "--use-scanpos-names was provided, but there are {} rxp files for scan position {}", paths.len(), scan_position.name ); } paths .into_iter() .map(|path| { Translation { outfile: self.outfile(scan_position, &path), infile: path, } }) .collect() } fn colorize(&self, scan_position: &ScanPosition, translation: &Translation) { use std::f64; let image_groups = self.image_groups(scan_position); let stream = Stream::from_path(&translation.infile) .sync_to_pps(self.sync_to_pps) .open() .unwrap(); let mut writer = las::Writer::from_path(&translation.outfile, self.las_header()).unwrap(); for point in stream { let point = point.expect("could not read rxp point"); let socs = Point::socs(point.x, point.y, point.z); let temperatures = image_groups .iter() .filter_map(|image_group| image_group.temperature(&socs)) .collect::<Vec<_>>(); let temperature = if temperatures.is_empty() { if self.keep_without_thermal { f64::NAN } else { continue; } } else { temperatures.iter().sum::<f64>() / temperatures.len() as f64 }; let glcs = socs.to_prcs(scan_position.sop).to_glcs(self.project.pop); let point = las::Point { x: glcs.x, y: glcs.y, z: glcs.z, intensity: self.to_intensity(point.reflectance), color: Some(self.to_color(temperature as f32)), gps_time: Some(temperature), ..Default::default() }; writer.write(point).expect("could not write las point"); } } fn scan_positions(&self) -> Vec<&ScanPosition> { let mut scan_positions: Vec<_> = if let Some(names) = self.scan_position_names.as_ref() { names .iter() .map(|name| self.project.scan_positions.get(name).unwrap()) .collect() } else { self.project.scan_positions.values().collect() }; scan_positions.sort_by_key(|s| &s.name); scan_positions } fn to_color(&self, n: f32) -> Color { let color = self.temperature_gradient.get(n); Color { red: (u16::MAX as f32 * color.red) as u16, green: (u16::MAX as f32 * color.green) as u16, blue: (u16::MAX as f32 * color.blue) as u16, } } fn to_intensity(&self, n: f32) -> u16 { (u16::MAX as f32 * (n - self.min_reflectance) / (self.max_reflectance - self.min_reflectance)) as u16 } fn las_header(&self) -> las::Header { let mut header = las::Header::default(); header.point_format = Format::new(3).unwrap(); header.transforms = las::Vector { x: las::Transform { scale: 0.001, offset: self.project.pop[(0, 3)], }, y: las::Transform { scale: 0.001, offset: self.project.pop[(1, 3)], }, z: las::Transform { scale: 0.001, offset: self.project.pop[(2, 3)], }, }; header } fn image_groups<'a>(&'a self, scan_position: &'a ScanPosition) -> Vec<ImageGroup<'a>> { let mut image_dir = self.image_dir.clone(); image_dir.push(&scan_position.name); match fs::read_dir(image_dir) { Ok(read_dir) => { read_dir .enumerate() .filter_map(|(i, entry)| { let entry = entry.unwrap(); if entry.path().extension().map(|e| e == "irb").unwrap_or( false, ) { let image = if let Some(name) = self.name_map(scan_position) { let image_name = format!("{} - Image{:03}", name, i + 1); scan_position.images.get(&image_name).expect(&format!( "Could not find image {}", image_name )) } else { scan_position.image_from_path(entry.path()).unwrap() }; let irb = Irb::from_path(entry.path().to_string_lossy().as_ref()) .unwrap(); let camera_calibration = image.camera_calibration(&self.project).unwrap(); let mount_calibration = image.mount_calibration(&self.project).unwrap(); Some(ImageGroup { camera_calibration: camera_calibration, image: image, irb: irb, irb_path: entry.path(), mount_calibration: mount_calibration, rotate: self.rotate, }) } else { None } }) .collect()
} Err(err) => { use std::io::ErrorKind;
random_line_split
main.rs
").unwrap()).unwrap(); let image_dir = PathBuf::from(matches.value_of("IMAGE_DIR").unwrap()); let las_dir = Path::new(matches.value_of("LAS_DIR").unwrap()).to_path_buf(); let min_reflectance = value_t!(matches, "min-reflectance", f32).unwrap(); let max_reflectance = value_t!(matches, "max-reflectance", f32).unwrap(); let min_temperature = value_t!(matches, "min-temperature", f32).unwrap(); let max_temperature = value_t!(matches, "max-temperature", f32).unwrap(); let min_temperature_color = Rgb::new(0.0, 0., 1.0); let max_temperature_color = Rgb::new(1.0, 0., 0.); let temperature_gradient = Gradient::with_domain(vec![ (min_temperature, min_temperature_color), (max_temperature, max_temperature_color), ]); let name_map = if let Some(name_map) = matches.value_of("name-map") { let mut s = String::new(); File::open(name_map) .unwrap() .read_to_string(&mut s) .unwrap(); toml::from_str(&s).unwrap() } else { NameMap::default() }; Config { image_dir: image_dir, keep_without_thermal: matches.is_present("keep-without-thermal"), las_dir: las_dir, max_reflectance: max_reflectance, min_reflectance: min_reflectance, project: project, rotate: matches.is_present("rotate"), scan_position_names: matches.values_of("scan-position").map(|values| { values.map(|name| name.to_string()).collect() }), sync_to_pps: matches.is_present("sync-to-pps"), temperature_gradient: temperature_gradient, use_scanpos_names: matches.is_present("use-scanpos-names"), name_map: name_map, } } fn translations(&self, scan_position: &ScanPosition) -> Vec<Translation> { let paths = scan_position.singlescan_rxp_paths(&self.project); if self.use_scanpos_names && paths.len() > 1 { panic!( "--use-scanpos-names was provided, but there are {} rxp files for scan position {}", paths.len(), scan_position.name ); } paths .into_iter() .map(|path| { Translation { outfile: self.outfile(scan_position, &path), infile: path, } }) .collect() } fn colorize(&self, scan_position: &ScanPosition, translation: &Translation) { use std::f64; let image_groups = self.image_groups(scan_position); let stream = Stream::from_path(&translation.infile) .sync_to_pps(self.sync_to_pps) .open() .unwrap(); let mut writer = las::Writer::from_path(&translation.outfile, self.las_header()).unwrap(); for point in stream { let point = point.expect("could not read rxp point"); let socs = Point::socs(point.x, point.y, point.z); let temperatures = image_groups .iter() .filter_map(|image_group| image_group.temperature(&socs)) .collect::<Vec<_>>(); let temperature = if temperatures.is_empty() { if self.keep_without_thermal { f64::NAN } else { continue; } } else { temperatures.iter().sum::<f64>() / temperatures.len() as f64 }; let glcs = socs.to_prcs(scan_position.sop).to_glcs(self.project.pop); let point = las::Point { x: glcs.x, y: glcs.y, z: glcs.z, intensity: self.to_intensity(point.reflectance), color: Some(self.to_color(temperature as f32)), gps_time: Some(temperature), ..Default::default() }; writer.write(point).expect("could not write las point"); } } fn scan_positions(&self) -> Vec<&ScanPosition> { let mut scan_positions: Vec<_> = if let Some(names) = self.scan_position_names.as_ref() { names .iter() .map(|name| self.project.scan_positions.get(name).unwrap()) .collect() } else { self.project.scan_positions.values().collect() }; scan_positions.sort_by_key(|s| &s.name); scan_positions } fn to_color(&self, n: f32) -> Color { let color = self.temperature_gradient.get(n); Color { red: (u16::MAX as f32 * color.red) as u16, green: (u16::MAX as f32 * color.green) as u16, blue: (u16::MAX as f32 * color.blue) as u16, } } fn to_intensity(&self, n: f32) -> u16 { (u16::MAX as f32 * (n - self.min_reflectance) / (self.max_reflectance - self.min_reflectance)) as u16 } fn las_header(&self) -> las::Header { let mut header = las::Header::default(); header.point_format = Format::new(3).unwrap(); header.transforms = las::Vector { x: las::Transform { scale: 0.001, offset: self.project.pop[(0, 3)], }, y: las::Transform { scale: 0.001, offset: self.project.pop[(1, 3)], }, z: las::Transform { scale: 0.001, offset: self.project.pop[(2, 3)], }, }; header } fn image_groups<'a>(&'a self, scan_position: &'a ScanPosition) -> Vec<ImageGroup<'a>> { let mut image_dir = self.image_dir.clone(); image_dir.push(&scan_position.name); match fs::read_dir(image_dir) { Ok(read_dir) => { read_dir .enumerate() .filter_map(|(i, entry)| { let entry = entry.unwrap(); if entry.path().extension().map(|e| e == "irb").unwrap_or( false, ) { let image = if let Some(name) = self.name_map(scan_position) { let image_name = format!("{} - Image{:03}", name, i + 1); scan_position.images.get(&image_name).expect(&format!( "Could not find image {}", image_name )) } else { scan_position.image_from_path(entry.path()).unwrap() }; let irb = Irb::from_path(entry.path().to_string_lossy().as_ref()) .unwrap(); let camera_calibration = image.camera_calibration(&self.project).unwrap(); let mount_calibration = image.mount_calibration(&self.project).unwrap(); Some(ImageGroup { camera_calibration: camera_calibration, image: image, irb: irb, irb_path: entry.path(), mount_calibration: mount_calibration, rotate: self.rotate, }) } else { None } }) .collect() } Err(err) => { use std::io::ErrorKind; match err.kind() { ErrorKind::NotFound => Vec::new(), _ => panic!("io error: {}", err), } } } } fn outfile<P: AsRef<Path>>(&self, scan_position: &ScanPosition, infile: P) -> PathBuf { let mut outfile = self.las_dir.clone(); if self.use_scanpos_names { outfile.push(Path::new(&scan_position.name).with_extension("las")); } else { outfile.push(infile.as_ref().with_extension("las").file_name().unwrap()); } outfile } fn name_map(&self, scan_position: &ScanPosition) -> Option<&str> { self.name_map .maps .iter() .find(|map| map.from == scan_position.name) .map(|map| map.to.as_str()) } } impl fmt::Display for Config { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { writeln!(f, "Configuration:")?; writeln!(f, " - project: {}", self.project.path.display())?; writeln!(f, " - image dir: {}", self.image_dir.display())?; writeln!(f, " - las dir: {}", self.las_dir.display())?; writeln!(f, " - scan positions:")?; for scan_position in self.scan_positions() { writeln!(f, " - name: {}", scan_position.name)?; let image_groups = self.image_groups(scan_position); if image_groups.is_empty() { writeln!(f, " - no images for this scan position")?; } else { writeln!(f, " - images:")?; for image_group in image_groups { writeln!(f, " - {}", image_group.irb_path.display())?; } } } Ok(()) } } impl<'a> ImageGroup<'a> { fn
temperature
identifier_name
main.rs
().flush().unwrap(); let answer: String = read!(); println!(); match answer.to_lowercase().as_str() { "y" => break, "n" => return, _ => println!("Unknown response: {}", answer), } } for scan_position in config.scan_positions() { println!("Colorizing {}:", scan_position.name); let translations = config.translations(scan_position); if translations.is_empty() { println!(" - No translations found"); } else { for translation in translations { println!( " - Translation:\n - Infile: {}\n - Outfile: {}", translation.infile.display(), translation.outfile.display() ); config.colorize(scan_position, &translation); } } } println!("Complete!"); } struct Config { image_dir: PathBuf, keep_without_thermal: bool, las_dir: PathBuf, max_reflectance: f32, min_reflectance: f32, project: Project, rotate: bool, scan_position_names: Option<Vec<String>>, sync_to_pps: bool, temperature_gradient: Gradient<Rgb>, use_scanpos_names: bool, name_map: NameMap, } struct ImageGroup<'a> { camera_calibration: &'a CameraCalibration, image: &'a Image, irb: Irb, irb_path: PathBuf, mount_calibration: &'a MountCalibration, rotate: bool, } struct Translation { infile: PathBuf, outfile: PathBuf, } #[derive(Debug, Default, Deserialize)] struct NameMap { maps: Vec<FromTo>, } #[derive(Debug, Default, Deserialize)] struct FromTo { from: String, to: String, } impl Config { fn new(matches: &ArgMatches) -> Config { use std::fs::File; use std::io::Read; use toml; let project = Project::from_path(matches.value_of("PROJECT").unwrap()).unwrap(); let image_dir = PathBuf::from(matches.value_of("IMAGE_DIR").unwrap()); let las_dir = Path::new(matches.value_of("LAS_DIR").unwrap()).to_path_buf(); let min_reflectance = value_t!(matches, "min-reflectance", f32).unwrap(); let max_reflectance = value_t!(matches, "max-reflectance", f32).unwrap(); let min_temperature = value_t!(matches, "min-temperature", f32).unwrap(); let max_temperature = value_t!(matches, "max-temperature", f32).unwrap(); let min_temperature_color = Rgb::new(0.0, 0., 1.0); let max_temperature_color = Rgb::new(1.0, 0., 0.); let temperature_gradient = Gradient::with_domain(vec![ (min_temperature, min_temperature_color), (max_temperature, max_temperature_color), ]); let name_map = if let Some(name_map) = matches.value_of("name-map") { let mut s = String::new(); File::open(name_map) .unwrap() .read_to_string(&mut s) .unwrap(); toml::from_str(&s).unwrap() } else { NameMap::default() }; Config { image_dir: image_dir, keep_without_thermal: matches.is_present("keep-without-thermal"), las_dir: las_dir, max_reflectance: max_reflectance, min_reflectance: min_reflectance, project: project, rotate: matches.is_present("rotate"), scan_position_names: matches.values_of("scan-position").map(|values| { values.map(|name| name.to_string()).collect() }), sync_to_pps: matches.is_present("sync-to-pps"), temperature_gradient: temperature_gradient, use_scanpos_names: matches.is_present("use-scanpos-names"), name_map: name_map, } } fn translations(&self, scan_position: &ScanPosition) -> Vec<Translation> { let paths = scan_position.singlescan_rxp_paths(&self.project); if self.use_scanpos_names && paths.len() > 1 { panic!( "--use-scanpos-names was provided, but there are {} rxp files for scan position {}", paths.len(), scan_position.name ); } paths .into_iter() .map(|path| { Translation { outfile: self.outfile(scan_position, &path), infile: path, } }) .collect() } fn colorize(&self, scan_position: &ScanPosition, translation: &Translation) { use std::f64; let image_groups = self.image_groups(scan_position); let stream = Stream::from_path(&translation.infile) .sync_to_pps(self.sync_to_pps) .open() .unwrap(); let mut writer = las::Writer::from_path(&translation.outfile, self.las_header()).unwrap(); for point in stream { let point = point.expect("could not read rxp point"); let socs = Point::socs(point.x, point.y, point.z); let temperatures = image_groups .iter() .filter_map(|image_group| image_group.temperature(&socs)) .collect::<Vec<_>>(); let temperature = if temperatures.is_empty() { if self.keep_without_thermal { f64::NAN } else { continue; } } else { temperatures.iter().sum::<f64>() / temperatures.len() as f64 }; let glcs = socs.to_prcs(scan_position.sop).to_glcs(self.project.pop); let point = las::Point { x: glcs.x, y: glcs.y, z: glcs.z, intensity: self.to_intensity(point.reflectance), color: Some(self.to_color(temperature as f32)), gps_time: Some(temperature), ..Default::default() }; writer.write(point).expect("could not write las point"); } } fn scan_positions(&self) -> Vec<&ScanPosition> { let mut scan_positions: Vec<_> = if let Some(names) = self.scan_position_names.as_ref() { names .iter() .map(|name| self.project.scan_positions.get(name).unwrap()) .collect() } else { self.project.scan_positions.values().collect() }; scan_positions.sort_by_key(|s| &s.name); scan_positions } fn to_color(&self, n: f32) -> Color { let color = self.temperature_gradient.get(n); Color { red: (u16::MAX as f32 * color.red) as u16, green: (u16::MAX as f32 * color.green) as u16, blue: (u16::MAX as f32 * color.blue) as u16, } } fn to_intensity(&self, n: f32) -> u16 { (u16::MAX as f32 * (n - self.min_reflectance) / (self.max_reflectance - self.min_reflectance)) as u16 } fn las_header(&self) -> las::Header { let mut header = las::Header::default(); header.point_format = Format::new(3).unwrap(); header.transforms = las::Vector { x: las::Transform { scale: 0.001, offset: self.project.pop[(0, 3)], }, y: las::Transform { scale: 0.001, offset: self.project.pop[(1, 3)], }, z: las::Transform { scale: 0.001, offset: self.project.pop[(2, 3)], }, }; header } fn image_groups<'a>(&'a self, scan_position: &'a ScanPosition) -> Vec<ImageGroup<'a>> { let mut image_dir = self.image_dir.clone(); image_dir.push(&scan_position.name); match fs::read_dir(image_dir) { Ok(read_dir) => { read_dir .enumerate() .filter_map(|(i, entry)| { let entry = entry.unwrap(); if entry.path().extension().map(|e| e == "irb").unwrap_or( false, ) { let image = if let Some(name) = self.name_map(scan_position) { let image_name = format!("{} - Image{:03}", name, i + 1); scan_position.images.get(&image_name).expect(&format!( "Could not find image {}", image_name )) } else { scan_position.image_from_path(entry.path()).unwrap() }; let irb = Irb::from_path(entry.path().to_string_lossy().as_ref()) .unwrap(); let camera_calibration = image.camera_calibration(&self.project).unwrap(); let mount_calibration = image.mount_calibration(&self.project).unwrap(); Some(ImageGroup { camera_calibration: camera_calibration, image: image, irb: irb, irb_path: entry.path(), mount_calibration: mount_calibration, rotate: self.rotate, }) } else
{ None }
conditional_block
data_collection.py
RUNNING = 1 FINISHED = 2 class DataCollection: def __init__(self, user, mode, iterations, museID = None): self.user = user self.museID = museID pygame.init() self.width = 600 self.height = 600 pygame.display.set_caption(user + ' Data Collection Session') self.screen = pygame.display.set_mode((self.width, self.height)) self.totalIterations = iterations self.passwords = self.generate_passwords(mode, iterations) self.mode = mode self.currentPassIndex = 0 self.currentCharIndex = 0 self.donePass = False self.inputSize = (300, 60) self.inputPosition = (self.width/2 - self.inputSize[0]/2, self.height/2 - self.inputSize[1]/2) font = pygame.font.Font(None, 50) inputRect = pygame.Rect(self.inputPosition[0], self.inputPosition[1], self.inputSize[0], self.inputSize[1]) self.input = TextBox(inputRect, clear_on_enter=True, inactive_on_enter=False, font=font) self.gameRunning = False self.state = DataCollectionState.MUSE_DISCONNECTED # 0 = Muse Disconnected, 1 = Session Running, 2 = Finished self.setup_marker_streaming() self.markers = [[]] # Each item is array of 2 items - timestamp + the key which was pressed. self.eegData = [[]] # Each item is array of timestamp + data for each channel. self.get_eeg_stream(0.5) self.startTime = time() # Timestamp of experiment start. self.finishTime = 0 # Timestamp of experiment finish. self.lastEEGSampleTime = self.startTime def setup_marker_streaming(self): streamName = self.user + ' Training Session Markers' self.markerInfo = StreamInfo(streamName, 'Keystroke Markers', 1, 0, 'string', str(uuid.uuid1())) self.markerOutlet = StreamOutlet(self.markerInfo) def get_eeg_stream(self, timeout): eeg_inlet_streams : StreamInlet = resolve_byprop('type', 'EEG', timeout=timeout) for stream in eeg_inlet_streams: if self.museID == None or not stream.name().find(self.museID) == -1: self.eegInlet = StreamInlet(stream) self.eegTimeCorrection = self.eegInlet.time_correction() self.state = DataCollectionState.RUNNING self.doneCheckEEG = True def push_marker(self, timestamp, currentChar): self.markerOutlet.push_sample(currentChar, timestamp) # Push key marker with timestamp via LSL for other programs. self.markers.append([timestamp, currentChar]) def pull_eeg_data(self, timeout = 0.0, max_samples = 360): samples, timestamps = self.eegInlet.pull_chunk(timeout, max_samples) # Pull samples. timestampCount = len(timestamps) if(timestampCount > 0):
def save_data(self): info = self.eegInlet.info() desc = info.desc() chanNum = info.channel_count() channels = desc.child('channels').first_child() channelNames = [channels.child_value('label')] for i in range(1, chanNum): channels = channels.next_sibling() channelNames.append(channels.child_value('label')) startTime = datetime.datetime.fromtimestamp(self.startTime).strftime(Constants.SESSION_FILE_DATETIME_FORMAT) finishTime = datetime.datetime.fromtimestamp(self.finishTime).strftime(Constants.SESSION_FILE_DATETIME_FORMAT) # Save EEG Data fileBase = os.path.join('session_data', self.user, self.mode.name, self.user + '_' + self.mode.name + '_' + startTime + '_' + finishTime) file = fileBase + '_EEG.csv' helpers.ensure_dir(file) with open(file, 'w', newline='') as csvfile: writer = csv.writer(csvfile, delimiter=',', quotechar='"', quoting=csv.QUOTE_MINIMAL) writer.writerow(['timestamp'] + channelNames) for data in self.eegData: writer.writerow(data) print('Saved EEG data to: ' + file) # Save Marker Data file = os.path.join('session_data', self.user, self.mode.name, self.user + '_' + self.mode.name + '_' + startTime + '_' + finishTime).replace(':','\ua789') file = fileBase + '_MRK.csv' helpers.ensure_dir(file) with open(file, 'w', newline='') as csvfile: writer = csv.writer(csvfile, delimiter=',', quotechar='"', quoting=csv.QUOTE_MINIMAL) writer.writerow(['timestamp','key marker']) for data in self.markers: writer.writerow(data) print('Saved Marker data to: ' + file) def generate_passwords(self, mode, iterations): passwords = [''] * iterations if mode == PasswordTypes.PIN_FIXED_4: length = 4 poolInit = '0123456789' elif mode == PasswordTypes.MIXED_FIXED_8: length = 8 poolInit = 'abcdefghijklmnopqrstuvwxyz' pool = poolInit # Calculate number of each character required for even distribution. freq = math.floor(iterations * length / len(pool)) poolTrack = [freq] * len(pool) # Keeps track of how many of each letter has been used. for i in range(iterations): for j in range(length): if len(poolTrack) != 0: index = random.randint(0, len(poolTrack) - 1) char = pool[index] poolTrack[index] -= 1 if poolTrack[index] == 0: poolTrack.pop(index) pool = pool.replace(char,'') # Once we've used the minimum required "freq" of each character, we simply do a random choice from the initial pool. else: char = random.choice(poolInit) passwords[i] += char.upper() return passwords def draw_static_ui(self): fontPassEnt = pygame.font.Font(None, 40) passEnt = 'Passwords Entered: ' passEntS = fontPassEnt.render(passEnt, 1, (0,0,0)) iter = str(self.currentPassIndex) + ' / ' + str(self.totalIterations) iterS = fontPassEnt.render(iter, 1, (0,0,0)) iterOffsetX = fontPassEnt.size(iter)[0] + 10 self.screen.blit(passEntS, (self.width - iterOffsetX - fontPassEnt.size(passEnt)[0] - 10, 10)) self.screen.blit(iterS, (self.width - iterOffsetX, 10)) if self.state == DataCollectionState.RUNNING: instruct = 'Type the password below, press ENTER when done:' elif self.state == DataCollectionState.MUSE_DISCONNECTED: instruct = 'Error: a Muse LSL stream must be active to continue (Muse ID: {0})'.format(self.museID) else: instruct = 'Finished session. This window will close in a moment.' fontInstruct = pygame.font.Font(None, 24) instructS = fontInstruct.render(instruct, 1, (0,0,0)) instructSize = fontInstruct.size(instruct) self.screen.blit(instructS, (self.width/2 - instructSize[0]/2, self.height/4 - instructSize[1]/2)) def process_input(self): for event in pygame.event.get(): if self.state == DataCollectionState.RUNNING: currentPass = self.passwords[self.currentPassIndex] currentChar = currentPass[self.currentCharIndex] if event.type == pygame.KEYDOWN: if (event.key == ord(currentChar) or event.key == ord(currentChar.lower())) and not self.donePass: newEvent = pygame.event.Event(pygame.KEYDOWN, {'unicode': currentChar.upper(),'key': ord(currentChar.upper()), 'mod': None}) self.input.get_event(newEvent) self.push_marker(float(time()), currentChar) if self.currentCharIndex < len(currentPass) - 1: self.currentCharIndex += 1 else: self.donePass = True elif event.key == pygame.K_RETURN and self.donePass: self.currentCharIndex = 0 self.currentPassIndex += 1 if self.currentPassIndex == self.totalIterations: self.state = DataCollectionState.FINISHED self.input.get_event(event) self.donePass = False if event.type == pygame.QUIT: pygame.quit() def process_logic(self): if self.state == DataCollectionState.MUSE_DISCONNECTED: if self.doneCheckEEG == True: self.doneCheckEEG = False threading.Thread(target = self.get_eeg_stream, kwargs={'timeout' : 5}).start() elif self.state == DataCollection
print('Number of samples: {0} | Time since last: {1}'.format(timestampCount, time() - self.lastEEGSampleTime)) self.lastEEGSampleTime = time() for i in range(0, len(timestamps)): self.eegData.append([timestamps[i]] + samples[i])
conditional_block
data_collection.py
RUNNING = 1 FINISHED = 2 class DataCollection: def __init__(self, user, mode, iterations, museID = None): self.user = user self.museID = museID pygame.init() self.width = 600 self.height = 600 pygame.display.set_caption(user + ' Data Collection Session') self.screen = pygame.display.set_mode((self.width, self.height)) self.totalIterations = iterations self.passwords = self.generate_passwords(mode, iterations) self.mode = mode self.currentPassIndex = 0 self.currentCharIndex = 0 self.donePass = False self.inputSize = (300, 60) self.inputPosition = (self.width/2 - self.inputSize[0]/2, self.height/2 - self.inputSize[1]/2) font = pygame.font.Font(None, 50) inputRect = pygame.Rect(self.inputPosition[0], self.inputPosition[1], self.inputSize[0], self.inputSize[1]) self.input = TextBox(inputRect, clear_on_enter=True, inactive_on_enter=False, font=font) self.gameRunning = False self.state = DataCollectionState.MUSE_DISCONNECTED # 0 = Muse Disconnected, 1 = Session Running, 2 = Finished self.setup_marker_streaming() self.markers = [[]] # Each item is array of 2 items - timestamp + the key which was pressed. self.eegData = [[]] # Each item is array of timestamp + data for each channel. self.get_eeg_stream(0.5) self.startTime = time() # Timestamp of experiment start. self.finishTime = 0 # Timestamp of experiment finish. self.lastEEGSampleTime = self.startTime
def setup_marker_streaming(self): streamName = self.user + ' Training Session Markers' self.markerInfo = StreamInfo(streamName, 'Keystroke Markers', 1, 0, 'string', str(uuid.uuid1())) self.markerOutlet = StreamOutlet(self.markerInfo) def get_eeg_stream(self, timeout): eeg_inlet_streams : StreamInlet = resolve_byprop('type', 'EEG', timeout=timeout) for stream in eeg_inlet_streams: if self.museID == None or not stream.name().find(self.museID) == -1: self.eegInlet = StreamInlet(stream) self.eegTimeCorrection = self.eegInlet.time_correction() self.state = DataCollectionState.RUNNING self.doneCheckEEG = True def push_marker(self, timestamp, currentChar): self.markerOutlet.push_sample(currentChar, timestamp) # Push key marker with timestamp via LSL for other programs. self.markers.append([timestamp, currentChar]) def pull_eeg_data(self, timeout = 0.0, max_samples = 360): samples, timestamps = self.eegInlet.pull_chunk(timeout, max_samples) # Pull samples. timestampCount = len(timestamps) if(timestampCount > 0): print('Number of samples: {0} | Time since last: {1}'.format(timestampCount, time() - self.lastEEGSampleTime)) self.lastEEGSampleTime = time() for i in range(0, len(timestamps)): self.eegData.append([timestamps[i]] + samples[i]) def save_data(self): info = self.eegInlet.info() desc = info.desc() chanNum = info.channel_count() channels = desc.child('channels').first_child() channelNames = [channels.child_value('label')] for i in range(1, chanNum): channels = channels.next_sibling() channelNames.append(channels.child_value('label')) startTime = datetime.datetime.fromtimestamp(self.startTime).strftime(Constants.SESSION_FILE_DATETIME_FORMAT) finishTime = datetime.datetime.fromtimestamp(self.finishTime).strftime(Constants.SESSION_FILE_DATETIME_FORMAT) # Save EEG Data fileBase = os.path.join('session_data', self.user, self.mode.name, self.user + '_' + self.mode.name + '_' + startTime + '_' + finishTime) file = fileBase + '_EEG.csv' helpers.ensure_dir(file) with open(file, 'w', newline='') as csvfile: writer = csv.writer(csvfile, delimiter=',', quotechar='"', quoting=csv.QUOTE_MINIMAL) writer.writerow(['timestamp'] + channelNames) for data in self.eegData: writer.writerow(data) print('Saved EEG data to: ' + file) # Save Marker Data file = os.path.join('session_data', self.user, self.mode.name, self.user + '_' + self.mode.name + '_' + startTime + '_' + finishTime).replace(':','\ua789') file = fileBase + '_MRK.csv' helpers.ensure_dir(file) with open(file, 'w', newline='') as csvfile: writer = csv.writer(csvfile, delimiter=',', quotechar='"', quoting=csv.QUOTE_MINIMAL) writer.writerow(['timestamp','key marker']) for data in self.markers: writer.writerow(data) print('Saved Marker data to: ' + file) def generate_passwords(self, mode, iterations): passwords = [''] * iterations if mode == PasswordTypes.PIN_FIXED_4: length = 4 poolInit = '0123456789' elif mode == PasswordTypes.MIXED_FIXED_8: length = 8 poolInit = 'abcdefghijklmnopqrstuvwxyz' pool = poolInit # Calculate number of each character required for even distribution. freq = math.floor(iterations * length / len(pool)) poolTrack = [freq] * len(pool) # Keeps track of how many of each letter has been used. for i in range(iterations): for j in range(length): if len(poolTrack) != 0: index = random.randint(0, len(poolTrack) - 1) char = pool[index] poolTrack[index] -= 1 if poolTrack[index] == 0: poolTrack.pop(index) pool = pool.replace(char,'') # Once we've used the minimum required "freq" of each character, we simply do a random choice from the initial pool. else: char = random.choice(poolInit) passwords[i] += char.upper() return passwords def draw_static_ui(self): fontPassEnt = pygame.font.Font(None, 40) passEnt = 'Passwords Entered: ' passEntS = fontPassEnt.render(passEnt, 1, (0,0,0)) iter = str(self.currentPassIndex) + ' / ' + str(self.totalIterations) iterS = fontPassEnt.render(iter, 1, (0,0,0)) iterOffsetX = fontPassEnt.size(iter)[0] + 10 self.screen.blit(passEntS, (self.width - iterOffsetX - fontPassEnt.size(passEnt)[0] - 10, 10)) self.screen.blit(iterS, (self.width - iterOffsetX, 10)) if self.state == DataCollectionState.RUNNING: instruct = 'Type the password below, press ENTER when done:' elif self.state == DataCollectionState.MUSE_DISCONNECTED: instruct = 'Error: a Muse LSL stream must be active to continue (Muse ID: {0})'.format(self.museID) else: instruct = 'Finished session. This window will close in a moment.' fontInstruct = pygame.font.Font(None, 24) instructS = fontInstruct.render(instruct, 1, (0,0,0)) instructSize = fontInstruct.size(instruct) self.screen.blit(instructS, (self.width/2 - instructSize[0]/2, self.height/4 - instructSize[1]/2)) def process_input(self): for event in pygame.event.get(): if self.state == DataCollectionState.RUNNING: currentPass = self.passwords[self.currentPassIndex] currentChar = currentPass[self.currentCharIndex] if event.type == pygame.KEYDOWN: if (event.key == ord(currentChar) or event.key == ord(currentChar.lower())) and not self.donePass: newEvent = pygame.event.Event(pygame.KEYDOWN, {'unicode': currentChar.upper(),'key': ord(currentChar.upper()), 'mod': None}) self.input.get_event(newEvent) self.push_marker(float(time()), currentChar) if self.currentCharIndex < len(currentPass) - 1: self.currentCharIndex += 1 else: self.donePass = True elif event.key == pygame.K_RETURN and self.donePass: self.currentCharIndex = 0 self.currentPassIndex += 1 if self.currentPassIndex == self.totalIterations: self.state = DataCollectionState.FINISHED self.input.get_event(event) self.donePass = False if event.type == pygame.QUIT: pygame.quit() def process_logic(self): if self.state == DataCollectionState.MUSE_DISCONNECTED: if self.doneCheckEEG == True: self.doneCheckEEG = False threading.Thread(target = self.get_eeg_stream, kwargs={'timeout' : 5}).start() elif self.state == DataCollection
random_line_split
data_collection.py
RUNNING = 1 FINISHED = 2 class DataCollection: def __init__(self, user, mode, iterations, museID = None): self.user = user self.museID = museID pygame.init() self.width = 600 self.height = 600 pygame.display.set_caption(user + ' Data Collection Session') self.screen = pygame.display.set_mode((self.width, self.height)) self.totalIterations = iterations self.passwords = self.generate_passwords(mode, iterations) self.mode = mode self.currentPassIndex = 0 self.currentCharIndex = 0 self.donePass = False self.inputSize = (300, 60) self.inputPosition = (self.width/2 - self.inputSize[0]/2, self.height/2 - self.inputSize[1]/2) font = pygame.font.Font(None, 50) inputRect = pygame.Rect(self.inputPosition[0], self.inputPosition[1], self.inputSize[0], self.inputSize[1]) self.input = TextBox(inputRect, clear_on_enter=True, inactive_on_enter=False, font=font) self.gameRunning = False self.state = DataCollectionState.MUSE_DISCONNECTED # 0 = Muse Disconnected, 1 = Session Running, 2 = Finished self.setup_marker_streaming() self.markers = [[]] # Each item is array of 2 items - timestamp + the key which was pressed. self.eegData = [[]] # Each item is array of timestamp + data for each channel. self.get_eeg_stream(0.5) self.startTime = time() # Timestamp of experiment start. self.finishTime = 0 # Timestamp of experiment finish. self.lastEEGSampleTime = self.startTime def setup_marker_streaming(self): streamName = self.user + ' Training Session Markers' self.markerInfo = StreamInfo(streamName, 'Keystroke Markers', 1, 0, 'string', str(uuid.uuid1())) self.markerOutlet = StreamOutlet(self.markerInfo) def get_eeg_stream(self, timeout): eeg_inlet_streams : StreamInlet = resolve_byprop('type', 'EEG', timeout=timeout) for stream in eeg_inlet_streams: if self.museID == None or not stream.name().find(self.museID) == -1: self.eegInlet = StreamInlet(stream) self.eegTimeCorrection = self.eegInlet.time_correction() self.state = DataCollectionState.RUNNING self.doneCheckEEG = True def push_marker(self, timestamp, currentChar): self.markerOutlet.push_sample(currentChar, timestamp) # Push key marker with timestamp via LSL for other programs. self.markers.append([timestamp, currentChar]) def pull_eeg_data(self, timeout = 0.0, max_samples = 360): samples, timestamps = self.eegInlet.pull_chunk(timeout, max_samples) # Pull samples. timestampCount = len(timestamps) if(timestampCount > 0): print('Number of samples: {0} | Time since last: {1}'.format(timestampCount, time() - self.lastEEGSampleTime)) self.lastEEGSampleTime = time() for i in range(0, len(timestamps)): self.eegData.append([timestamps[i]] + samples[i]) def save_data(self): info = self.eegInlet.info() desc = info.desc() chanNum = info.channel_count() channels = desc.child('channels').first_child() channelNames = [channels.child_value('label')] for i in range(1, chanNum): channels = channels.next_sibling() channelNames.append(channels.child_value('label')) startTime = datetime.datetime.fromtimestamp(self.startTime).strftime(Constants.SESSION_FILE_DATETIME_FORMAT) finishTime = datetime.datetime.fromtimestamp(self.finishTime).strftime(Constants.SESSION_FILE_DATETIME_FORMAT) # Save EEG Data fileBase = os.path.join('session_data', self.user, self.mode.name, self.user + '_' + self.mode.name + '_' + startTime + '_' + finishTime) file = fileBase + '_EEG.csv' helpers.ensure_dir(file) with open(file, 'w', newline='') as csvfile: writer = csv.writer(csvfile, delimiter=',', quotechar='"', quoting=csv.QUOTE_MINIMAL) writer.writerow(['timestamp'] + channelNames) for data in self.eegData: writer.writerow(data) print('Saved EEG data to: ' + file) # Save Marker Data file = os.path.join('session_data', self.user, self.mode.name, self.user + '_' + self.mode.name + '_' + startTime + '_' + finishTime).replace(':','\ua789') file = fileBase + '_MRK.csv' helpers.ensure_dir(file) with open(file, 'w', newline='') as csvfile: writer = csv.writer(csvfile, delimiter=',', quotechar='"', quoting=csv.QUOTE_MINIMAL) writer.writerow(['timestamp','key marker']) for data in self.markers: writer.writerow(data) print('Saved Marker data to: ' + file) def generate_passwords(self, mode, iterations): passwords = [''] * iterations if mode == PasswordTypes.PIN_FIXED_4: length = 4 poolInit = '0123456789' elif mode == PasswordTypes.MIXED_FIXED_8: length = 8 poolInit = 'abcdefghijklmnopqrstuvwxyz' pool = poolInit # Calculate number of each character required for even distribution. freq = math.floor(iterations * length / len(pool)) poolTrack = [freq] * len(pool) # Keeps track of how many of each letter has been used. for i in range(iterations): for j in range(length): if len(poolTrack) != 0: index = random.randint(0, len(poolTrack) - 1) char = pool[index] poolTrack[index] -= 1 if poolTrack[index] == 0: poolTrack.pop(index) pool = pool.replace(char,'') # Once we've used the minimum required "freq" of each character, we simply do a random choice from the initial pool. else: char = random.choice(poolInit) passwords[i] += char.upper() return passwords def draw_static_ui(self): fontPassEnt = pygame.font.Font(None, 40) passEnt = 'Passwords Entered: ' passEntS = fontPassEnt.render(passEnt, 1, (0,0,0)) iter = str(self.currentPassIndex) + ' / ' + str(self.totalIterations) iterS = fontPassEnt.render(iter, 1, (0,0,0)) iterOffsetX = fontPassEnt.size(iter)[0] + 10 self.screen.blit(passEntS, (self.width - iterOffsetX - fontPassEnt.size(passEnt)[0] - 10, 10)) self.screen.blit(iterS, (self.width - iterOffsetX, 10)) if self.state == DataCollectionState.RUNNING: instruct = 'Type the password below, press ENTER when done:' elif self.state == DataCollectionState.MUSE_DISCONNECTED: instruct = 'Error: a Muse LSL stream must be active to continue (Muse ID: {0})'.format(self.museID) else: instruct = 'Finished session. This window will close in a moment.' fontInstruct = pygame.font.Font(None, 24) instructS = fontInstruct.render(instruct, 1, (0,0,0)) instructSize = fontInstruct.size(instruct) self.screen.blit(instructS, (self.width/2 - instructSize[0]/2, self.height/4 - instructSize[1]/2)) def process_input(self): for event in pygame.event.get(): if self.state == DataCollectionState.RUNNING: currentPass = self.passwords[self.currentPassIndex] currentChar = currentPass[self.currentCharIndex] if event.type == pygame.KEYDOWN: if (event.key == ord(currentChar) or event.key == ord(currentChar.lower())) and not self.donePass: newEvent = pygame.event.Event(pygame.KEYDOWN, {'unicode': currentChar.upper(),'key': ord(currentChar.upper()), 'mod': None}) self.input.get_event(newEvent) self.push_marker(float(time()), currentChar) if self.currentCharIndex < len(currentPass) - 1: self.currentCharIndex += 1 else: self.donePass = True elif event.key == pygame.K_RETURN and self.donePass: self.currentCharIndex = 0 self.currentPassIndex += 1 if self.currentPassIndex == self.totalIterations: self.state = DataCollectionState.FINISHED self.input.get_event(event) self.donePass = False if event.type == pygame.QUIT: pygame.quit() def
(self): if self.state == DataCollectionState.MUSE_DISCONNECTED: if self.doneCheckEEG == True: self.doneCheckEEG = False threading.Thread(target = self.get_eeg_stream, kwargs={'timeout' : 5}).start() elif self.state == DataCollection
process_logic
identifier_name
data_collection.py
RUNNING = 1 FINISHED = 2 class DataCollection:
self.state = DataCollectionState.MUSE_DISCONNECTED # 0 = Muse Disconnected, 1 = Session Running, 2 = Finished self.setup_marker_streaming() self.markers = [[]] # Each item is array of 2 items - timestamp + the key which was pressed. self.eegData = [[]] # Each item is array of timestamp + data for each channel. self.get_eeg_stream(0.5) self.startTime = time() # Timestamp of experiment start. self.finishTime = 0 # Timestamp of experiment finish. self.lastEEGSampleTime = self.startTime def setup_marker_streaming(self): streamName = self.user + ' Training Session Markers' self.markerInfo = StreamInfo(streamName, 'Keystroke Markers', 1, 0, 'string', str(uuid.uuid1())) self.markerOutlet = StreamOutlet(self.markerInfo) def get_eeg_stream(self, timeout): eeg_inlet_streams : StreamInlet = resolve_byprop('type', 'EEG', timeout=timeout) for stream in eeg_inlet_streams: if self.museID == None or not stream.name().find(self.museID) == -1: self.eegInlet = StreamInlet(stream) self.eegTimeCorrection = self.eegInlet.time_correction() self.state = DataCollectionState.RUNNING self.doneCheckEEG = True def push_marker(self, timestamp, currentChar): self.markerOutlet.push_sample(currentChar, timestamp) # Push key marker with timestamp via LSL for other programs. self.markers.append([timestamp, currentChar]) def pull_eeg_data(self, timeout = 0.0, max_samples = 360): samples, timestamps = self.eegInlet.pull_chunk(timeout, max_samples) # Pull samples. timestampCount = len(timestamps) if(timestampCount > 0): print('Number of samples: {0} | Time since last: {1}'.format(timestampCount, time() - self.lastEEGSampleTime)) self.lastEEGSampleTime = time() for i in range(0, len(timestamps)): self.eegData.append([timestamps[i]] + samples[i]) def save_data(self): info = self.eegInlet.info() desc = info.desc() chanNum = info.channel_count() channels = desc.child('channels').first_child() channelNames = [channels.child_value('label')] for i in range(1, chanNum): channels = channels.next_sibling() channelNames.append(channels.child_value('label')) startTime = datetime.datetime.fromtimestamp(self.startTime).strftime(Constants.SESSION_FILE_DATETIME_FORMAT) finishTime = datetime.datetime.fromtimestamp(self.finishTime).strftime(Constants.SESSION_FILE_DATETIME_FORMAT) # Save EEG Data fileBase = os.path.join('session_data', self.user, self.mode.name, self.user + '_' + self.mode.name + '_' + startTime + '_' + finishTime) file = fileBase + '_EEG.csv' helpers.ensure_dir(file) with open(file, 'w', newline='') as csvfile: writer = csv.writer(csvfile, delimiter=',', quotechar='"', quoting=csv.QUOTE_MINIMAL) writer.writerow(['timestamp'] + channelNames) for data in self.eegData: writer.writerow(data) print('Saved EEG data to: ' + file) # Save Marker Data file = os.path.join('session_data', self.user, self.mode.name, self.user + '_' + self.mode.name + '_' + startTime + '_' + finishTime).replace(':','\ua789') file = fileBase + '_MRK.csv' helpers.ensure_dir(file) with open(file, 'w', newline='') as csvfile: writer = csv.writer(csvfile, delimiter=',', quotechar='"', quoting=csv.QUOTE_MINIMAL) writer.writerow(['timestamp','key marker']) for data in self.markers: writer.writerow(data) print('Saved Marker data to: ' + file) def generate_passwords(self, mode, iterations): passwords = [''] * iterations if mode == PasswordTypes.PIN_FIXED_4: length = 4 poolInit = '0123456789' elif mode == PasswordTypes.MIXED_FIXED_8: length = 8 poolInit = 'abcdefghijklmnopqrstuvwxyz' pool = poolInit # Calculate number of each character required for even distribution. freq = math.floor(iterations * length / len(pool)) poolTrack = [freq] * len(pool) # Keeps track of how many of each letter has been used. for i in range(iterations): for j in range(length): if len(poolTrack) != 0: index = random.randint(0, len(poolTrack) - 1) char = pool[index] poolTrack[index] -= 1 if poolTrack[index] == 0: poolTrack.pop(index) pool = pool.replace(char,'') # Once we've used the minimum required "freq" of each character, we simply do a random choice from the initial pool. else: char = random.choice(poolInit) passwords[i] += char.upper() return passwords def draw_static_ui(self): fontPassEnt = pygame.font.Font(None, 40) passEnt = 'Passwords Entered: ' passEntS = fontPassEnt.render(passEnt, 1, (0,0,0)) iter = str(self.currentPassIndex) + ' / ' + str(self.totalIterations) iterS = fontPassEnt.render(iter, 1, (0,0,0)) iterOffsetX = fontPassEnt.size(iter)[0] + 10 self.screen.blit(passEntS, (self.width - iterOffsetX - fontPassEnt.size(passEnt)[0] - 10, 10)) self.screen.blit(iterS, (self.width - iterOffsetX, 10)) if self.state == DataCollectionState.RUNNING: instruct = 'Type the password below, press ENTER when done:' elif self.state == DataCollectionState.MUSE_DISCONNECTED: instruct = 'Error: a Muse LSL stream must be active to continue (Muse ID: {0})'.format(self.museID) else: instruct = 'Finished session. This window will close in a moment.' fontInstruct = pygame.font.Font(None, 24) instructS = fontInstruct.render(instruct, 1, (0,0,0)) instructSize = fontInstruct.size(instruct) self.screen.blit(instructS, (self.width/2 - instructSize[0]/2, self.height/4 - instructSize[1]/2)) def process_input(self): for event in pygame.event.get(): if self.state == DataCollectionState.RUNNING: currentPass = self.passwords[self.currentPassIndex] currentChar = currentPass[self.currentCharIndex] if event.type == pygame.KEYDOWN: if (event.key == ord(currentChar) or event.key == ord(currentChar.lower())) and not self.donePass: newEvent = pygame.event.Event(pygame.KEYDOWN, {'unicode': currentChar.upper(),'key': ord(currentChar.upper()), 'mod': None}) self.input.get_event(newEvent) self.push_marker(float(time()), currentChar) if self.currentCharIndex < len(currentPass) - 1: self.currentCharIndex += 1 else: self.donePass = True elif event.key == pygame.K_RETURN and self.donePass: self.currentCharIndex = 0 self.currentPassIndex += 1 if self.currentPassIndex == self.totalIterations: self.state = DataCollectionState.FINISHED self.input.get_event(event) self.donePass = False if event.type == pygame.QUIT: pygame.quit() def process_logic(self): if self.state == DataCollectionState.MUSE_DISCONNECTED: if self.doneCheckEEG == True: self.doneCheckEEG = False threading.Thread(target = self.get_eeg_stream, kwargs={'timeout' : 5}).start() elif self.state == DataCollectionState
def __init__(self, user, mode, iterations, museID = None): self.user = user self.museID = museID pygame.init() self.width = 600 self.height = 600 pygame.display.set_caption(user + ' Data Collection Session') self.screen = pygame.display.set_mode((self.width, self.height)) self.totalIterations = iterations self.passwords = self.generate_passwords(mode, iterations) self.mode = mode self.currentPassIndex = 0 self.currentCharIndex = 0 self.donePass = False self.inputSize = (300, 60) self.inputPosition = (self.width/2 - self.inputSize[0]/2, self.height/2 - self.inputSize[1]/2) font = pygame.font.Font(None, 50) inputRect = pygame.Rect(self.inputPosition[0], self.inputPosition[1], self.inputSize[0], self.inputSize[1]) self.input = TextBox(inputRect, clear_on_enter=True, inactive_on_enter=False, font=font) self.gameRunning = False
identifier_body
mp4TagWriter.ts
atoms.length > 0) throw new Error("Buffer already parsed"); let offset = 0; let atom: Atom; while (true) { atom = this._readAtom(offset); if (!atom || atom.length < 1) break; this._atoms.push(atom); offset = atom.offset + atom.length; } if (this._atoms.length < 1) throw new Error("Buffer could not be parsed"); } setDuration(duration: number) { const mvhdAtom: Atom = this._findAtom(this._atoms, ["moov", "mvhd"]); if (!mvhdAtom) throw new Error("'mvhd' atom could not be found"); // version(4) + created(4) + modified(4) + timescale(4) const precedingDataLength = 16; this._bufferView.setUint32(mvhdAtom.offset + ATOM_HEAD_LENGTH + precedingDataLength, duration); }
(name: string, data: ArrayBuffer | string | number) { if (name.length > 4 || name.length < 1) throw new Error(`Unsupported atom name: '${name}'`); let dataBuffer: ArrayBuffer; if (data instanceof ArrayBuffer) { dataBuffer = data; } else if (typeof data === "string") { dataBuffer = this._getBufferFromString(data); } else if (typeof data === "number") { dataBuffer = new ArrayBuffer(4); const dataView = new DataView(dataBuffer); dataView.setUint32(0, data); } else { throw new Error(`Unsupported data: '${data}'`); } const atom: Atom = { name, length: ATOM_HEADER_LENGTH + dataBuffer.byteLength, data: dataBuffer, }; this._insertAtom(atom, this._metadataPath); } getBuffer() { const buffers: ArrayBuffer[] = []; let bufferIndex = 0; // we don't change the offsets, since it would add needless complexity without benefit for (const atom of this._atoms) { if (!atom.children) { // nothing has been added or removed const slice = this._buffer.slice(atom.offset, atom.offset + atom.length); buffers.push(slice); bufferIndex++; continue; } atom.length = ATOM_HEAD_LENGTH; const levels: AtomLevel[] = [{ parent: atom, offset: bufferIndex, childIndex: 0 }]; let levelIndex = 0; while (true) { const { parent, offset, childIndex } = levels[levelIndex]; if (childIndex >= parent.children.length) { // move one level up levelIndex--; levels.pop(); let parentHeadLength = ATOM_HEAD_LENGTH; if (parent.name === "meta") { parent.length += 4; parentHeadLength += 4; } else if (parent.name === "stsd") { parent.length += 8; parentHeadLength += 8; } // set length of parent in buffer this._bufferView.setUint32(parent.offset, parent.length); const parentHeader = this._buffer.slice(parent.offset, parent.offset + parentHeadLength); buffers.splice(offset, 0, parentHeader); // we completed the last parent - exit if (levelIndex < 0) break; // add our current parents length to new parent and move childIndex of new parent one ahead const newParent = levels[levelIndex].parent; newParent.length += parent.length; levels[levelIndex].childIndex++; continue; } const child = parent.children[childIndex]; if (child.children) { // move one level down child.length = ATOM_HEAD_LENGTH; levels.push({ parent: child, offset: bufferIndex, childIndex: 0 }); levelIndex++; continue; } else if (child.data) { // add new data to buffer const headerBuffer = this._getHeaderBufferFromAtom(child); buffers.push(headerBuffer); buffers.push(child.data); } else { // add entire child to buffer const slice = this._buffer.slice(child.offset, child.offset + child.length); buffers.push(slice); } bufferIndex++; parent.length += child.length; // move one child ahead levels[levelIndex].childIndex++; } } this._bufferView = null; this._buffer = null; this._atoms = []; return concatArrayBuffers(buffers); } private _insertAtom(atom: Atom, path: string[]) { if (!path) throw new Error("Path can not be empty"); const parentAtom = this._findAtom(this._atoms, path); if (!parentAtom) throw new Error(`Parent atom at path '${path.join(" > ")}' could not be found`); if (parentAtom.children === undefined) { parentAtom.children = this._readChildAtoms(parentAtom); } let offset = parentAtom.offset + ATOM_HEAD_LENGTH; if (parentAtom.name === "meta") { offset += 4; } else if (parentAtom.name === "stsd") { offset += 8; } if (parentAtom.children.length > 0) { const lastChild = parentAtom.children[parentAtom.children.length - 1]; offset = lastChild.offset + lastChild.length; } atom.offset = offset; parentAtom.children.push(atom); } private _findAtom(atoms: Atom[], path: string[]): Atom | null { if (!path || path.length < 1) throw new Error("Path can not be empty"); const curPath = [...path]; const curName = curPath.shift(); const curElem = atoms.find((i) => i.name === curName); if (curPath.length < 1) return curElem; if (!curElem) return null; if (curElem.children === undefined) { curElem.children = this._readChildAtoms(curElem); } if (curElem.children.length < 1) return null; return this._findAtom(curElem.children, curPath); } private _readChildAtoms(atom: Atom): Atom[] { const children: Atom[] = []; const childEnd = atom.offset + atom.length; let childOffset = atom.offset + ATOM_HEAD_LENGTH; if (atom.name === "meta") { childOffset += 4; } else if (atom.name === "stsd") { childOffset += 8; } while (true) { if (childOffset >= childEnd) break; const childAtom = this._readAtom(childOffset); if (!childAtom || childAtom.length < 1) break; childOffset = childAtom.offset + childAtom.length; children.push(childAtom); } return children; } private _readAtom(offset: number): Atom { const begin = offset; const end = offset + ATOM_HEAD_LENGTH; const buffer = this._buffer.slice(begin, end); if (buffer.byteLength < ATOM_HEAD_LENGTH) { return { length: buffer.byteLength, offset, }; } const dataView = new DataView(buffer); let length = dataView.getUint32(0, false); let name = ""; for (let i = 0; i < 4; i++) { name += String.fromCharCode(dataView.getUint8(4 + i)); } return { name, length, offset, }; } private _getHeaderBufferFromAtom(atom: Atom) { if (!atom || atom.length < 1 || !atom.name || !atom.data) throw new Error("Can not compute header buffer for this atom"); const headerBuffer = new ArrayBuffer(ATOM_HEADER_LENGTH); const headerBufferView = new DataView(headerBuffer); // length at 0, length = 4 headerBufferView.setUint32(0, atom.length); // name at 4, length = 4 const nameChars = this._getCharCodes(atom.name); for (let i = 0; i < nameChars.length; i++) { headerBufferView.setUint8(4 + i, nameChars[i]); } // data length at 8, length = 4 headerBufferView.setUint32(8, ATOM_DATA_HEAD_LENGTH + atom.data.byteLength); // data name at 12, length = 4 const dataNameChars = this._getCharCodes("data"); for (let i = 0; i < dataNameChars.length; i++) { headerBufferView.setUint8(12 + i, dataNameChars[i]); } // data flags at 16, length = 4 headerBufferView.setUint32(16, this._getFlags(atom.name)); return headerBuffer; } private _getBufferFromString(input: string): ArrayBuffer { // return new TextEncoder().encode(input).buffer; const buffer = new ArrayBuffer(input.length); const bufferView = new DataView(buffer); const chars = this._getCharCodes(input); for (let i = 0; i < chars.length; i++) { bufferView.setUint8(i, chars[i]); } return buffer; } private _getCharCodes(input: string) { const chars: number[] = []; for (let i =
addMetadataAtom
identifier_name
mp4TagWriter.ts
.length > 0) throw new Error("Buffer already parsed"); let offset = 0; let atom: Atom; while (true) { atom = this._readAtom(offset); if (!atom || atom.length < 1) break; this._atoms.push(atom); offset = atom.offset + atom.length; } if (this._atoms.length < 1) throw new Error("Buffer could not be parsed"); } setDuration(duration: number) { const mvhdAtom: Atom = this._findAtom(this._atoms, ["moov", "mvhd"]); if (!mvhdAtom) throw new Error("'mvhd' atom could not be found"); // version(4) + created(4) + modified(4) + timescale(4) const precedingDataLength = 16; this._bufferView.setUint32(mvhdAtom.offset + ATOM_HEAD_LENGTH + precedingDataLength, duration); } addMetadataAtom(name: string, data: ArrayBuffer | string | number) { if (name.length > 4 || name.length < 1) throw new Error(`Unsupported atom name: '${name}'`); let dataBuffer: ArrayBuffer; if (data instanceof ArrayBuffer) { dataBuffer = data; } else if (typeof data === "string") { dataBuffer = this._getBufferFromString(data); } else if (typeof data === "number") { dataBuffer = new ArrayBuffer(4); const dataView = new DataView(dataBuffer); dataView.setUint32(0, data); } else { throw new Error(`Unsupported data: '${data}'`); } const atom: Atom = { name, length: ATOM_HEADER_LENGTH + dataBuffer.byteLength, data: dataBuffer, }; this._insertAtom(atom, this._metadataPath); } getBuffer() { const buffers: ArrayBuffer[] = []; let bufferIndex = 0; // we don't change the offsets, since it would add needless complexity without benefit for (const atom of this._atoms) { if (!atom.children) { // nothing has been added or removed const slice = this._buffer.slice(atom.offset, atom.offset + atom.length); buffers.push(slice); bufferIndex++; continue; } atom.length = ATOM_HEAD_LENGTH; const levels: AtomLevel[] = [{ parent: atom, offset: bufferIndex, childIndex: 0 }]; let levelIndex = 0; while (true) { const { parent, offset, childIndex } = levels[levelIndex]; if (childIndex >= parent.children.length) { // move one level up levelIndex--; levels.pop(); let parentHeadLength = ATOM_HEAD_LENGTH; if (parent.name === "meta") { parent.length += 4; parentHeadLength += 4; } else if (parent.name === "stsd") { parent.length += 8; parentHeadLength += 8; } // set length of parent in buffer this._bufferView.setUint32(parent.offset, parent.length); const parentHeader = this._buffer.slice(parent.offset, parent.offset + parentHeadLength); buffers.splice(offset, 0, parentHeader); // we completed the last parent - exit if (levelIndex < 0) break; // add our current parents length to new parent and move childIndex of new parent one ahead const newParent = levels[levelIndex].parent; newParent.length += parent.length; levels[levelIndex].childIndex++; continue; } const child = parent.children[childIndex]; if (child.children) { // move one level down child.length = ATOM_HEAD_LENGTH; levels.push({ parent: child, offset: bufferIndex, childIndex: 0 }); levelIndex++; continue; } else if (child.data) { // add new data to buffer const headerBuffer = this._getHeaderBufferFromAtom(child); buffers.push(headerBuffer); buffers.push(child.data); } else { // add entire child to buffer const slice = this._buffer.slice(child.offset, child.offset + child.length); buffers.push(slice); } bufferIndex++; parent.length += child.length; // move one child ahead levels[levelIndex].childIndex++; } } this._bufferView = null; this._buffer = null; this._atoms = []; return concatArrayBuffers(buffers); } private _insertAtom(atom: Atom, path: string[]) { if (!path) throw new Error("Path can not be empty"); const parentAtom = this._findAtom(this._atoms, path); if (!parentAtom) throw new Error(`Parent atom at path '${path.join(" > ")}' could not be found`); if (parentAtom.children === undefined) { parentAtom.children = this._readChildAtoms(parentAtom); } let offset = parentAtom.offset + ATOM_HEAD_LENGTH; if (parentAtom.name === "meta") { offset += 4; } else if (parentAtom.name === "stsd") { offset += 8; } if (parentAtom.children.length > 0) { const lastChild = parentAtom.children[parentAtom.children.length - 1]; offset = lastChild.offset + lastChild.length; } atom.offset = offset; parentAtom.children.push(atom); } private _findAtom(atoms: Atom[], path: string[]): Atom | null { if (!path || path.length < 1) throw new Error("Path can not be empty"); const curPath = [...path]; const curName = curPath.shift(); const curElem = atoms.find((i) => i.name === curName); if (curPath.length < 1) return curElem; if (!curElem) return null; if (curElem.children === undefined) { curElem.children = this._readChildAtoms(curElem); } if (curElem.children.length < 1) return null; return this._findAtom(curElem.children, curPath); } private _readChildAtoms(atom: Atom): Atom[] { const children: Atom[] = []; const childEnd = atom.offset + atom.length; let childOffset = atom.offset + ATOM_HEAD_LENGTH; if (atom.name === "meta") { childOffset += 4; } else if (atom.name === "stsd") { childOffset += 8; } while (true) { if (childOffset >= childEnd) break; const childAtom = this._readAtom(childOffset); if (!childAtom || childAtom.length < 1) break; childOffset = childAtom.offset + childAtom.length; children.push(childAtom); } return children; } private _readAtom(offset: number): Atom
} return { name, length, offset, }; } private _getHeaderBufferFromAtom(atom: Atom) { if (!atom || atom.length < 1 || !atom.name || !atom.data) throw new Error("Can not compute header buffer for this atom"); const headerBuffer = new ArrayBuffer(ATOM_HEADER_LENGTH); const headerBufferView = new DataView(headerBuffer); // length at 0, length = 4 headerBufferView.setUint32(0, atom.length); // name at 4, length = 4 const nameChars = this._getCharCodes(atom.name); for (let i = 0; i < nameChars.length; i++) { headerBufferView.setUint8(4 + i, nameChars[i]); } // data length at 8, length = 4 headerBufferView.setUint32(8, ATOM_DATA_HEAD_LENGTH + atom.data.byteLength); // data name at 12, length = 4 const dataNameChars = this._getCharCodes("data"); for (let i = 0; i < dataNameChars.length; i++) { headerBufferView.setUint8(12 + i, dataNameChars[i]); } // data flags at 16, length = 4 headerBufferView.setUint32(16, this._getFlags(atom.name)); return headerBuffer; } private _getBufferFromString(input: string): ArrayBuffer { // return new TextEncoder().encode(input).buffer; const buffer = new ArrayBuffer(input.length); const bufferView = new DataView(buffer); const chars = this._getCharCodes(input); for (let i = 0; i < chars.length; i++) { bufferView.setUint8(i, chars[i]); } return buffer; } private _getCharCodes(input: string) { const chars: number[] = []; for (let i =
{ const begin = offset; const end = offset + ATOM_HEAD_LENGTH; const buffer = this._buffer.slice(begin, end); if (buffer.byteLength < ATOM_HEAD_LENGTH) { return { length: buffer.byteLength, offset, }; } const dataView = new DataView(buffer); let length = dataView.getUint32(0, false); let name = ""; for (let i = 0; i < 4; i++) { name += String.fromCharCode(dataView.getUint8(4 + i));
identifier_body
mp4TagWriter.ts
.length < 1) throw new Error(`Unsupported atom name: '${name}'`); let dataBuffer: ArrayBuffer; if (data instanceof ArrayBuffer) { dataBuffer = data; } else if (typeof data === "string") { dataBuffer = this._getBufferFromString(data); } else if (typeof data === "number") { dataBuffer = new ArrayBuffer(4); const dataView = new DataView(dataBuffer); dataView.setUint32(0, data); } else { throw new Error(`Unsupported data: '${data}'`); } const atom: Atom = { name, length: ATOM_HEADER_LENGTH + dataBuffer.byteLength, data: dataBuffer, }; this._insertAtom(atom, this._metadataPath); } getBuffer() { const buffers: ArrayBuffer[] = []; let bufferIndex = 0; // we don't change the offsets, since it would add needless complexity without benefit for (const atom of this._atoms) { if (!atom.children) { // nothing has been added or removed const slice = this._buffer.slice(atom.offset, atom.offset + atom.length); buffers.push(slice); bufferIndex++; continue; } atom.length = ATOM_HEAD_LENGTH; const levels: AtomLevel[] = [{ parent: atom, offset: bufferIndex, childIndex: 0 }]; let levelIndex = 0; while (true) { const { parent, offset, childIndex } = levels[levelIndex]; if (childIndex >= parent.children.length) { // move one level up levelIndex--; levels.pop(); let parentHeadLength = ATOM_HEAD_LENGTH; if (parent.name === "meta") { parent.length += 4; parentHeadLength += 4; } else if (parent.name === "stsd") { parent.length += 8; parentHeadLength += 8; } // set length of parent in buffer this._bufferView.setUint32(parent.offset, parent.length); const parentHeader = this._buffer.slice(parent.offset, parent.offset + parentHeadLength); buffers.splice(offset, 0, parentHeader); // we completed the last parent - exit if (levelIndex < 0) break; // add our current parents length to new parent and move childIndex of new parent one ahead const newParent = levels[levelIndex].parent; newParent.length += parent.length; levels[levelIndex].childIndex++; continue; } const child = parent.children[childIndex]; if (child.children) { // move one level down child.length = ATOM_HEAD_LENGTH; levels.push({ parent: child, offset: bufferIndex, childIndex: 0 }); levelIndex++; continue; } else if (child.data) { // add new data to buffer const headerBuffer = this._getHeaderBufferFromAtom(child); buffers.push(headerBuffer); buffers.push(child.data); } else { // add entire child to buffer const slice = this._buffer.slice(child.offset, child.offset + child.length); buffers.push(slice); } bufferIndex++; parent.length += child.length; // move one child ahead levels[levelIndex].childIndex++; } } this._bufferView = null; this._buffer = null; this._atoms = []; return concatArrayBuffers(buffers); } private _insertAtom(atom: Atom, path: string[]) { if (!path) throw new Error("Path can not be empty"); const parentAtom = this._findAtom(this._atoms, path); if (!parentAtom) throw new Error(`Parent atom at path '${path.join(" > ")}' could not be found`); if (parentAtom.children === undefined) { parentAtom.children = this._readChildAtoms(parentAtom); } let offset = parentAtom.offset + ATOM_HEAD_LENGTH; if (parentAtom.name === "meta") { offset += 4; } else if (parentAtom.name === "stsd") { offset += 8; } if (parentAtom.children.length > 0) { const lastChild = parentAtom.children[parentAtom.children.length - 1]; offset = lastChild.offset + lastChild.length; } atom.offset = offset; parentAtom.children.push(atom); } private _findAtom(atoms: Atom[], path: string[]): Atom | null { if (!path || path.length < 1) throw new Error("Path can not be empty"); const curPath = [...path]; const curName = curPath.shift(); const curElem = atoms.find((i) => i.name === curName); if (curPath.length < 1) return curElem; if (!curElem) return null; if (curElem.children === undefined) { curElem.children = this._readChildAtoms(curElem); } if (curElem.children.length < 1) return null; return this._findAtom(curElem.children, curPath); } private _readChildAtoms(atom: Atom): Atom[] { const children: Atom[] = []; const childEnd = atom.offset + atom.length; let childOffset = atom.offset + ATOM_HEAD_LENGTH; if (atom.name === "meta") { childOffset += 4; } else if (atom.name === "stsd") { childOffset += 8; } while (true) { if (childOffset >= childEnd) break; const childAtom = this._readAtom(childOffset); if (!childAtom || childAtom.length < 1) break; childOffset = childAtom.offset + childAtom.length; children.push(childAtom); } return children; } private _readAtom(offset: number): Atom { const begin = offset; const end = offset + ATOM_HEAD_LENGTH; const buffer = this._buffer.slice(begin, end); if (buffer.byteLength < ATOM_HEAD_LENGTH) { return { length: buffer.byteLength, offset, }; } const dataView = new DataView(buffer); let length = dataView.getUint32(0, false); let name = ""; for (let i = 0; i < 4; i++) { name += String.fromCharCode(dataView.getUint8(4 + i)); } return { name, length, offset, }; } private _getHeaderBufferFromAtom(atom: Atom) { if (!atom || atom.length < 1 || !atom.name || !atom.data) throw new Error("Can not compute header buffer for this atom"); const headerBuffer = new ArrayBuffer(ATOM_HEADER_LENGTH); const headerBufferView = new DataView(headerBuffer); // length at 0, length = 4 headerBufferView.setUint32(0, atom.length); // name at 4, length = 4 const nameChars = this._getCharCodes(atom.name); for (let i = 0; i < nameChars.length; i++) { headerBufferView.setUint8(4 + i, nameChars[i]); } // data length at 8, length = 4 headerBufferView.setUint32(8, ATOM_DATA_HEAD_LENGTH + atom.data.byteLength); // data name at 12, length = 4 const dataNameChars = this._getCharCodes("data"); for (let i = 0; i < dataNameChars.length; i++) { headerBufferView.setUint8(12 + i, dataNameChars[i]); } // data flags at 16, length = 4 headerBufferView.setUint32(16, this._getFlags(atom.name)); return headerBuffer; } private _getBufferFromString(input: string): ArrayBuffer { // return new TextEncoder().encode(input).buffer; const buffer = new ArrayBuffer(input.length); const bufferView = new DataView(buffer); const chars = this._getCharCodes(input); for (let i = 0; i < chars.length; i++) { bufferView.setUint8(i, chars[i]); } return buffer; } private _getCharCodes(input: string) { const chars: number[] = []; for (let i = 0; i < input.length; i++) { chars.push(input.charCodeAt(i)); } return chars; } private _getFlags(name: string) { switch (name) { case "covr": // 13 for jpeg, 14 for png return 13; case "trkn": case "disk": return 0; case "tmpo": case "cpil": case "rtng": return 21; default: return 1; } } } export class Mp4TagWriter implements TagWriter { private _mp4: Mp4; constructor(buffer: ArrayBuffer) { this._mp4 = new Mp4(buffer); this._mp4.parse(); } setTitle(title: string): void { if (!title) throw new Error("Invalid value for title");
this._mp4.addMetadataAtom("©nam", title); } setArtists(artists: string[]): void { if (!artists || artists.length < 1) throw new Error("Invalid value for artists");
random_line_split
mp4TagWriter.ts
.length > 0) throw new Error("Buffer already parsed"); let offset = 0; let atom: Atom; while (true) { atom = this._readAtom(offset); if (!atom || atom.length < 1) break; this._atoms.push(atom); offset = atom.offset + atom.length; } if (this._atoms.length < 1) throw new Error("Buffer could not be parsed"); } setDuration(duration: number) { const mvhdAtom: Atom = this._findAtom(this._atoms, ["moov", "mvhd"]); if (!mvhdAtom) throw new Error("'mvhd' atom could not be found"); // version(4) + created(4) + modified(4) + timescale(4) const precedingDataLength = 16; this._bufferView.setUint32(mvhdAtom.offset + ATOM_HEAD_LENGTH + precedingDataLength, duration); } addMetadataAtom(name: string, data: ArrayBuffer | string | number) { if (name.length > 4 || name.length < 1) throw new Error(`Unsupported atom name: '${name}'`); let dataBuffer: ArrayBuffer; if (data instanceof ArrayBuffer) { dataBuffer = data; } else if (typeof data === "string") { dataBuffer = this._getBufferFromString(data); } else if (typeof data === "number") { dataBuffer = new ArrayBuffer(4); const dataView = new DataView(dataBuffer); dataView.setUint32(0, data); } else { throw new Error(`Unsupported data: '${data}'`); } const atom: Atom = { name, length: ATOM_HEADER_LENGTH + dataBuffer.byteLength, data: dataBuffer, }; this._insertAtom(atom, this._metadataPath); } getBuffer() { const buffers: ArrayBuffer[] = []; let bufferIndex = 0; // we don't change the offsets, since it would add needless complexity without benefit for (const atom of this._atoms) { if (!atom.children) { // nothing has been added or removed const slice = this._buffer.slice(atom.offset, atom.offset + atom.length); buffers.push(slice); bufferIndex++; continue; } atom.length = ATOM_HEAD_LENGTH; const levels: AtomLevel[] = [{ parent: atom, offset: bufferIndex, childIndex: 0 }]; let levelIndex = 0; while (true) { const { parent, offset, childIndex } = levels[levelIndex]; if (childIndex >= parent.children.length) { // move one level up levelIndex--; levels.pop(); let parentHeadLength = ATOM_HEAD_LENGTH; if (parent.name === "meta") { parent.length += 4; parentHeadLength += 4; } else if (parent.name === "stsd") { parent.length += 8; parentHeadLength += 8; } // set length of parent in buffer this._bufferView.setUint32(parent.offset, parent.length); const parentHeader = this._buffer.slice(parent.offset, parent.offset + parentHeadLength); buffers.splice(offset, 0, parentHeader); // we completed the last parent - exit if (levelIndex < 0) break; // add our current parents length to new parent and move childIndex of new parent one ahead const newParent = levels[levelIndex].parent; newParent.length += parent.length; levels[levelIndex].childIndex++; continue; } const child = parent.children[childIndex]; if (child.children) { // move one level down child.length = ATOM_HEAD_LENGTH; levels.push({ parent: child, offset: bufferIndex, childIndex: 0 }); levelIndex++; continue; } else if (child.data) { // add new data to buffer const headerBuffer = this._getHeaderBufferFromAtom(child); buffers.push(headerBuffer); buffers.push(child.data); } else { // add entire child to buffer const slice = this._buffer.slice(child.offset, child.offset + child.length); buffers.push(slice); } bufferIndex++; parent.length += child.length; // move one child ahead levels[levelIndex].childIndex++; } } this._bufferView = null; this._buffer = null; this._atoms = []; return concatArrayBuffers(buffers); } private _insertAtom(atom: Atom, path: string[]) { if (!path) throw new Error("Path can not be empty"); const parentAtom = this._findAtom(this._atoms, path); if (!parentAtom) throw new Error(`Parent atom at path '${path.join(" > ")}' could not be found`); if (parentAtom.children === undefined) { parentAtom.children = this._readChildAtoms(parentAtom); } let offset = parentAtom.offset + ATOM_HEAD_LENGTH; if (parentAtom.name === "meta") { offset += 4; } else if (parentAtom.name === "stsd") { offset += 8; } if (parentAtom.children.length > 0) { const lastChild = parentAtom.children[parentAtom.children.length - 1]; offset = lastChild.offset + lastChild.length; } atom.offset = offset; parentAtom.children.push(atom); } private _findAtom(atoms: Atom[], path: string[]): Atom | null { if (!path || path.length < 1) throw new Error("Path can not be empty"); const curPath = [...path]; const curName = curPath.shift(); const curElem = atoms.find((i) => i.name === curName); if (curPath.length < 1) return curElem; if (!curElem) return null; if (curElem.children === undefined) { curElem.children = this._readChildAtoms(curElem); } if (curElem.children.length < 1) return null; return this._findAtom(curElem.children, curPath); } private _readChildAtoms(atom: Atom): Atom[] { const children: Atom[] = []; const childEnd = atom.offset + atom.length; let childOffset = atom.offset + ATOM_HEAD_LENGTH; if (atom.name === "meta") { childOffset += 4; } else if (atom.name === "stsd") { childOffset += 8; } while (true)
return children; } private _readAtom(offset: number): Atom { const begin = offset; const end = offset + ATOM_HEAD_LENGTH; const buffer = this._buffer.slice(begin, end); if (buffer.byteLength < ATOM_HEAD_LENGTH) { return { length: buffer.byteLength, offset, }; } const dataView = new DataView(buffer); let length = dataView.getUint32(0, false); let name = ""; for (let i = 0; i < 4; i++) { name += String.fromCharCode(dataView.getUint8(4 + i)); } return { name, length, offset, }; } private _getHeaderBufferFromAtom(atom: Atom) { if (!atom || atom.length < 1 || !atom.name || !atom.data) throw new Error("Can not compute header buffer for this atom"); const headerBuffer = new ArrayBuffer(ATOM_HEADER_LENGTH); const headerBufferView = new DataView(headerBuffer); // length at 0, length = 4 headerBufferView.setUint32(0, atom.length); // name at 4, length = 4 const nameChars = this._getCharCodes(atom.name); for (let i = 0; i < nameChars.length; i++) { headerBufferView.setUint8(4 + i, nameChars[i]); } // data length at 8, length = 4 headerBufferView.setUint32(8, ATOM_DATA_HEAD_LENGTH + atom.data.byteLength); // data name at 12, length = 4 const dataNameChars = this._getCharCodes("data"); for (let i = 0; i < dataNameChars.length; i++) { headerBufferView.setUint8(12 + i, dataNameChars[i]); } // data flags at 16, length = 4 headerBufferView.setUint32(16, this._getFlags(atom.name)); return headerBuffer; } private _getBufferFromString(input: string): ArrayBuffer { // return new TextEncoder().encode(input).buffer; const buffer = new ArrayBuffer(input.length); const bufferView = new DataView(buffer); const chars = this._getCharCodes(input); for (let i = 0; i < chars.length; i++) { bufferView.setUint8(i, chars[i]); } return buffer; } private _getCharCodes(input: string) { const chars: number[] = []; for (let i =
{ if (childOffset >= childEnd) break; const childAtom = this._readAtom(childOffset); if (!childAtom || childAtom.length < 1) break; childOffset = childAtom.offset + childAtom.length; children.push(childAtom); }
conditional_block
types.d.ts
asks/functionsaccessingociresources.htm | Accessing Other Oracle Cloud Infrastructure Resources from Running Functions} * for more information. Use Resouce Principal by setting * {@link IAMConfig#useResourcePrincipal} property to true.</li> * </ol> * <p> * Note that when using Instance Principal or Resource Principal you must * specify compartiment id (OCID), either as * {@link Config#compartment} property of the initial configuration or as * <em>opt.compartment</em> for each {@link NoSQLClient} method call. Note * that you must use compartment id (OCID) and not compartment name. This * also means that you may not prefix table name with compartment name when * calling methods of {@link NoSQLClient}. These restrictions do not apply * when using specific user identity, which is best when naming flexibility is * desired. * <p> * To use specific user's identity, you must provide the following credentials: * <ul> * <li>Tenancy OCID. This is Oracle Cloud ID (OCID) for your tenancy. See * {@link https://docs.cloud.oracle.com/iaas/Content/General/Concepts/identifiers.htm | Resource Identifiers} * for information on OCIDs.</li> * <li>User's OCID. This is Oracle Cloud ID (OCID) for the user in your * tenancy. See * {@link https://docs.cloud.oracle.com/iaas/Content/General/Concepts/identifiers.htm | Resource Identifiers} * for information on OCIDs.</li> * <li>API Signing Key. This is public-private key pair used to sign the API * requests, see {@link https://docs.cloud.oracle.com/iaas/Content/API/Concepts/apisigningkey.htm | Required Keys and OCIDs}. * In particular, private key is needed to generate the request signature.</li> * <li>Public Key Fingerprint. This is an identifier of the public key of the * API Signing Key pair.</li> * <li>Passphrase for the private key of API Signing Key pair if the private * key is encrypted.</li> * </ul> * <p> * See {@link https://docs.cloud.oracle.com/iaas/Content/API/Concepts/apisigningkey.htm | Required Keys and OCIDs} * for detailed description of the above credentials and the steps you need to * perform to enable signing of API requests, which are: * <ul> * <li>Generate the key pair described above.</li> * <li>Upload public key.</li> * <li>Obtain tenancy and user OCIDs and public key fingerprint.</li> * </ul>
* <p> * You may provide these credentials in one of the following ways, in order of * increased security: * <ul> * <li>Directly as properties of {@link IAMConfig}. * In this case, set properties {@link tenantId}, {@link userId}, * {@link privateKey} or {@link privateKeyFile}, {@link fingerprint} and * {@link passphrase} (if private key is encrypted)</li> * <li>As part of an OCI configuration file. See * {@link https://docs.cloud.oracle.com/iaas/Content/API/Concepts/sdkconfig.htm | SDK and CLI Configuration File} * for information on OCI configuration file and what entries are used for the * required credentials. In this case, you may set properties * {@link configFile} and/or {@link profileName}. If not set, * appropriate default values will be used, see property descriptions.</li> * <li>Specify your own credentials provider in the form of * {@link IAMCredentialsProvider} that implements {@link loadIAMCredentials} * function. This allows you to store and retrieve credentials in a secure * manner. In this case, specify {@link credentialsProvider} property.</li> * </ul> * Note that the private key must be in PEM format. You may provide a path * to the PEM key file. Alternatively, except when using OCI configuration * file, you may provide PEM encoded private key directly as <em>Buffer</em> * or <em>string</em>. Note that the {@link passphrase} must be provided if * the private key is encrypted. * <p> * <p> * The driver will determine the method of authorization as follows: * <ol> * <li>If {@link IAMConfig#useResourcePrincipal} is set to <em>true</em>, then * Resource Principal authentication will be used. No other properties listed * below are allowed for Resource Prinicpal authorization.</li> * <li>If {@link IAMConfig#useInstancePrincipal} is set to <em>true</em>, then * Instance Principal authentication will be used. You may also set * {@link IAMConfig#federationEndpoint}, although it is not requred and in * most cases federation endpoint will be auto-detected. No other properties * listed below are allowed for Instance Principal authorization.</li> * <li>If {@link useSessionToken} is set to <em>true</em>, then session * token-based authentication will be used. Note that this method uses OCI * config file, so settings to properties {@link configFile} and * {@link profileName} also apply. See {@link useSessionToken} for more * information. * <li>If {@link IAMConfig} has any of user identity properties such as * {@link tenantId}, {@link userId}, {@link privateKey}, {@link fingerprint} * or {@link passphrase}, the driver assumes that you are using a specific * user's identity and that the credentials are provided directly in * {@link IAMConfig}. All required user's credentials, as described above, * must be present as properties of {@link IAMConfig}, otherwise * {@link NoSQLArgumentError} will result.</li> * <li>If {@link IAMConfig} has {@link credentialsProvider} property, the * driver assumes that you are using a specific user's identity and the * credentials are obtained through the credentials provider which must be in * the form of {@link IAMCredentialsProvider}. In this case the credentials * must not be set directly in {@link IAMConfig}.</li> * <li>If none of the above, the driver assumes that you are using a specific * user's identity and the credentials are stored in OCI config * file and will use {@link configFile} and {@link profileName} properties * if present, otherwise it will assume their default values. In particular, * if you specify {@link Config#serviceType} as {@link ServiceType.CLOUD} * and omit {@link Config#auth} alltogether, the driver will use IAM * authorization with default OCI config file and default profile name.</li> * </ol> * <p> * Note that if using an OCI configuration file, you may also specify region * identifier in the same profile as your credentials. In this case, you need * not specify either region or endpoint in {@link Config}. In particular, * if you use the default OCI config file (<em>~/.oci/config</em>) and default * profile name (<em>DEFAULT</em>) and do not need to customize any other * configuration properties, you may create {@link NoSQLClient} instance * without providing configuration to {@link NoSQLClient} constructor. * See {@link NoSQLClient} for more information. * <p> * If using Resource Principal, you also need not specify either region or * endpoint in {@link Config}, as Resource Principal's region will be used. * In fact, when running in Functions service, you may only access NoSQL * service in the same region as the running function, so when using Resource * Principal, it is preferable not to specify either region or endpoint in * {@link Config}. * <p> * Generated authorization signature is valid for a period of time and is * cached for effeciency. The caching behavior may be customized with * properties {@link IAMConfig#durationSeconds} and * {@link IAMConfig#refreshAheadMs}. See their property descriptions for * details. * * @see {@link AuthConfig} * @see {@link IAMCredentials} * @see {@link IAMCredentialsProvider} * @see {@page connect-cloud.md} * * @example * JSON {@link Config} object supplying user's credentials directly * (sensitiveinfo not shown). * ```json * { * "region": "us-phoenix-1", * "auth": { * "iam": { * "tenantId": "ocid1.tenancy.oc...................", * "userId": "ocid1.user.oc.....................", * "fingerprint": "aa:aa:aa:aa:.....", * "privateKeyFile": "~/myapp/security/oci_api_key.pem", * "passphrase": "..............." * } * } * } * ``` * * @example * JSON {@link Config} object supplying user's credentials through OCI * configuration file. * ```json * { * "region": "us-phoenix-1", * "auth": { * "iam": { * "configFile": "~/myapp/.oci/config", * "profileName": "John" * } * } * } * ``` * * @example * Javascript {@link Config} object supplying user's credentials via custom * credentials provider. * ```js *
random_line_split
mqttclient.go
o *handler) handle(client mqtt.Client, msg mqtt.Message) { // We extract the count and write that out first to simplify checking for missing values var m Message var resp Session if err := json.Unmarshal(msg.Payload(), &resp); err != nil { fmt.Printf("Message could not be parsed (%s): %s", msg.Payload(), err) return } fmt.Println(resp) switch resp.Type { case CMDMSG_OFFER: enc.Decode(resp.Data, &m) Notice(m) case CMDMSG_DISC: var devcmd DiscoveryCmd enc.Decode(resp.Data, &devcmd) DiscoveryDev(&devcmd) case CMDMSG_WAKE: var fing Fing enc.Decode(resp.Data, &fing) wakemac(fing) case CMDMSG_UPDATE: var newver *versionUpdate GetUpdateMyself(newver) case CMDMSG_MR2: var mr2info Mr2Msg enc.Decode(resp.Data, &mr2info) Mr2HostPort(&mr2info) } } func Mr2HostPort(mr2info *Mr2Msg) { arg := fmt.Sprintf("client -s %s -p %s -P %d -c %s", mr2info.ServerAddr, mr2info.Password, mr2info.ExposePort, mr2info.ExposeAddr) fmt.Println("mr2", arg) err := fmt.Errorf("") //err := sys.CmdRun("./mr2", arg) if err != nil { CmdFeedBack(CMDMSG_MR2, 0, err.Error(), time.Now().String()) return } else { CmdFeedBack(CMDMSG_MR2, 1, "成功", time.Now().String()) } } func wakemac(fing Fing) { for _, v := range fing.Devices { wol.Wake(v.Mac, "", "", "") } } func DiscoveryDev(devcmd *DiscoveryCmd) { go func() { switch devcmd.DevType { case DEVICE_IP: dev := DiscoveryDevice()
req := &Session{} req.Type = "discoveryrsp" req.DeviceId = "kvm1" req.Data = enc.Encode(dev) //enc.Encode(answer) answermsg := PublishMsg{ Topic: "discoveryrsp", Msg: req, } fmt.Println("discoveryrsp", answermsg) SendMsg(answermsg) //response) case DEVICE_ONVIF: case DEVICE_SNMP: case DEVICE_MODBUS: case DEVICE_BACNET: case DEVICE_CAN: case DEVICE_UPCA: } }() } func CmdFeedBack(cmdstr string, status int, err string, sid string) { resp := ResponseMsg{ Cmdstr: cmdstr, Status: status, Err: err, Sid: sid, } req := &Session{} req.Type = "cmdFeedback" req.DeviceId = "kvm1" req.Data = enc.Encode(resp) //enc.Encode(answer) answermsg := PublishMsg{ Topic: "cmdFeedback", Msg: req, } fmt.Println("cmdFeedback", answermsg) SendMsg(answermsg) //response) } func GetCurrentPath() string { getwd, err := os.Getwd() if err != nil { fmt.Print(err.Error()) } else { fmt.Print(getwd) } return getwd } func GetUpdateMyself(newver *versionUpdate) { if newver.ForceUpdate == 1 { if newver.DownLoadUrl != "" { go func() { filepath := GetCurrentPath() + "/" + newver.Version fileext := ".zip" filename, err := DownloadFile(filepath, newver.DownLoadUrl, fileext) if err != nil { CmdFeedBack(CMDMSG_UPDATE, 0, err.Error(), "1") } else { fmt.Println("Download Finished") CmdFeedBack(CMDMSG_UPDATE, 1, "Download Finished", "1") if IsZip(filename) { err = Unzip(filename, filepath) if err != nil { CmdFeedBack(CMDMSG_UPDATE, 0, err.Error(), "1") } else { CmdFeedBack(CMDMSG_UPDATE, 1, "zip ok", "1") } } //以版本号建立新目录并解压包 //后期写更新配置文件告诉Process守护进程需要更新 //并立即退出或者按某种策略更新 //守护进程在判断到需要更新软件时就按更新配置文件的新路径执行程序 } }() } } } /* func SendMsgAnswer(msg Answer) { msgChans <- msg fmt.Print("SendMsg OK") //mqttclient.Publish(Config.Mqtt.PUBTOPIC+"/"+Config.Report.SN, Config.Mqtt.QOS, false, msg) } */ func SendMsg(msg PublishMsg) { msgChans <- msg fmt.Print("SendMsg OK") //mqttclient.Publish(Config.Mqtt.PUBTOPIC+"/"+Config.Report.SN, Config.Mqtt.QOS, false, msg) } func StartMqtt() { // Enable logging by uncommenting the below // mqtt.ERROR = log.New(os.Stdout, "[ERROR] ", 0) // mqtt.CRITICAL = log.New(os.Stdout, "[CRITICAL] ", 0) // mqtt.WARN = log.New(os.Stdout, "[WARN] ", 0) // mqtt.DEBUG = log.New(os.Stdout, "[DEBUG] ", 0) // Create a handler that will deal with incoming messages h := NewHandler() defer h.Close() msgChans = make(chan PublishMsg, 10) // Now we establish the connection to the mqtt broker opts := mqtt.NewClientOptions() opts.AddBroker(config.Config.Mqtt.SERVERADDRESS) opts.SetClientID(config.Config.Mqtt.CLIENTID) opts.ConnectTimeout = time.Second // Minimal delays on connect opts.WriteTimeout = time.Second // Minimal delays on writes opts.KeepAlive = 30 // Keepalive every 10 seconds so we quickly detect network outages opts.PingTimeout = time.Second // local broker so response should be quick // Automate connection management (will keep trying to connect and will reconnect if network drops) opts.ConnectRetry = true opts.AutoReconnect = true // If using QOS2 and CleanSession = FALSE then it is possible that we will receive messages on topics that we // have not subscribed to here (if they were previously subscribed to they are part of the session and survive // disconnect/reconnect). Adding a DefaultPublishHandler lets us detect this. opts.DefaultPublishHandler = func(_ mqtt.Client, msg mqtt.Message) { fmt.Printf("UNEXPECTED MESSAGE: %s\n", msg) } // Log events opts.OnConnectionLost = func(cl mqtt.Client, err error) { fmt.Println("connection lost") } opts.OnConnect = func(c mqtt.Client) { fmt.Println("connection established") // Establish the subscription - doing this here means that it willSUB happen every time a connection is established // (useful if opts.CleanSession is TRUE or the broker does not reliably store session data) t := c.Subscribe(config.Config.Mqtt.SUBTOPIC, config.Config.Mqtt.QOS, h.handle) // the connection handler is called in a goroutine so blocking here would hot cause an issue. However as blocking // in other handlers does cause problems its best to just assume we should not block go func() { _ = t.Wait() // Can also use '<-t.Done()' in releases > 1.2.0 if t.Error() != nil { fmt.Printf("ERROR SUBSCRIBING: %s\n", t.Error()) } else { fmt.Println("subscribed to: ", config.Config.Mqtt.SUBTOPIC) } }() } opts.OnReconnecting = func(mqtt.Client, *mqtt.ClientOptions) { fmt.Println("attempting to reconnect") } // // Connect to the broker // client := mqtt.NewClient(opts) // If using QOS2 and CleanSession = FALSE then messages may be transmitted to us before the subscribe completes. // Adding routes prior to connecting is a way of ensuring that these messages are processed client.AddRoute(config.Config.Mqtt.SUBTOPIC, h.handle) if token := client.Connect(); token.Wait() && token.Error() != nil { panic(token.Error()) } fmt.Println("Connection is up") done := make(chan struct{}) var wg sync.WaitGroup wg.Add(1) go func() { var count uint64 for { select { case data := <-msgChans: msg, err := json.Marshal(data.Msg) if err != nil { panic(err) } //t := client.Publish(Config.Mqtt.PUBTOPIC+"/"+Config.Report.SN, Config.Mqtt.QOS, false, msg) t := client.Publish(config.Config.Mqtt.PUBTOPIC+"/"+data.Topic, config.Config.Mqtt.QOS, false, msg) go func() { _ = t.Wait() // Can also use '<-t.Done()' in releases > 1.2.0 if t.Error() != nil { fmt.Printf("msg PUBLISHING: %s\n", t.Error().Error())
random_line_split
mqttclient.go
*handler) handle(client mqtt.Client, msg mqtt.Message) { // We extract the count and write that out first to simplify checking for missing values var m Message var resp Session if err := json.Unmarshal(msg.Payload(), &resp); err != nil { fmt.Printf("Message could not be parsed (%s): %s", msg.Payload(), err) return } fmt.Println(resp) switch resp.Type { case CMDMSG_OFFER: enc.Decode(resp.Data, &m) Notice(m) case CMDMSG_DISC: var devcmd DiscoveryCmd enc.Decode(resp.Data, &devcmd) DiscoveryDev(&devcmd) case CMDMSG_WAKE: var fing Fing enc.Decode(resp.Data, &fing) wakemac(fing) case CMDMSG_UPDATE: var newver *versionUpdate GetUpdateMyself(newver) case CMDMSG_MR2: var mr2info Mr2Msg enc.Decode(resp.Data, &mr2info) Mr2HostPort(&mr2info) } } func Mr2HostPort(mr2info *Mr2Msg) { arg := fmt.Sprintf("client -s %s -p %s -P %d -c %s", mr2info.ServerAddr, mr2info.Password, mr2info.ExposePort, mr2info.ExposeAddr) fmt.Println("mr2", arg) err := fmt.Errorf("") //err := sys.CmdRun("./mr2", arg) if err != nil { CmdFeedBack(CMDMSG_MR2, 0, err.Error(), time.Now().String()) return } else { CmdFeedBack(CMDMSG_MR2, 1, "成功", time.Now().String()) } } func wakemac(fing Fing) { for _, v := range fing.Devices { wol.Wake(v.Mac, "", "", "") } } func DiscoveryDev(devcmd *DiscoveryCmd) { go func() { switch devcmd.DevType { case DEVICE_IP: dev := DiscoveryDevice() req := &Session{} req.Type = "discoveryrsp" req.DeviceId = "kvm1" req.Data = enc.Encode(dev) //enc.Encode(answer) answermsg := PublishMsg{ Topic: "discoveryrsp", Msg: req, } fmt.Println("discoveryrsp", answermsg) SendMsg(answermsg) //response) case DEVICE_ONVIF: case DEVICE_SNMP: case DEVICE_MODBUS: case DEVICE_BACNET: case DEVICE_CAN: case DEVICE_UPCA: } }() } func CmdFeedBack(cmdstr string, status int, err string, sid string) { resp := ResponseMsg{ Cmdstr: cmdstr, Status: status, Err: err, Sid: sid, } req := &Session{} req.Type = "cmdFeedback" req.DeviceId = "kvm1" req.Data = enc.Encode(resp) //enc.Encode(answer) answermsg := PublishMsg{ Topic: "cmdFeedback", Msg: req, } fmt.Println("cmdFeedback", answermsg) SendMsg(answermsg) //response) } func GetCurrentPath() string { getwd, err := os.Getwd() if err != nil { fmt.Print(err.Error()) } else { fmt.Print(getwd) } return getwd } func GetUpdateMyself(newver *versionUpdate) { if newver.ForceUpdate == 1 { if newver.DownLoadUrl != "" { go func() { filepath := GetCurrentPath() + "/" + newver.Version fileext := ".zip" filename, err := DownloadFile(filepath, newver.DownLoadUrl, fileext) if err != nil { CmdFeedBack(CMDMSG_UPDATE, 0, err.Error(), "1") } else { fmt.Println("Download Finished") CmdFeedBack(CMDMSG_UPDATE, 1, "Download Finished", "1") if IsZip(filename) { err = Unzip(filename, filepath) if err != nil { CmdFeedBack(CMDMSG_UPDATE, 0, err.Error(), "1") } else { CmdFeedBack(CMDMSG_UPDATE, 1, "zip ok", "1") } } //以版本号建立新目录并解压包 //后期写更新配置文件告诉Process守护进程需要更新 //并立即退出或者按某种策略更新 //守护进程在判断到需要更新软件时就按更新配置文件的新路径执行程序 } }() } } } /* func SendMsgAnswer(msg Answer) { msgChans <- msg fmt.Print("SendMsg OK") //mqttclient.Publish(Config.Mqtt.PUBTOPIC+"/"+Config.Report.SN, Config.Mqtt.QOS, false, msg) } */ func SendMsg(msg PublishMsg) { msgChans <- msg fmt.Print("SendMsg OK") //mqttclient.Publish(Config.Mqtt.PUBTOPIC+"/"+Config.Report.SN, Config.Mqtt.QOS, false, msg) } func Star
t, "[CRITICAL] ", 0) // mqtt.WARN = log.New(os.Stdout, "[WARN] ", 0) // mqtt.DEBUG = log.New(os.Stdout, "[DEBUG] ", 0) // Create a handler that will deal with incoming messages h := NewHandler() defer h.Close() msgChans = make(chan PublishMsg, 10) // Now we establish the connection to the mqtt broker opts := mqtt.NewClientOptions() opts.AddBroker(config.Config.Mqtt.SERVERADDRESS) opts.SetClientID(config.Config.Mqtt.CLIENTID) opts.ConnectTimeout = time.Second // Minimal delays on connect opts.WriteTimeout = time.Second // Minimal delays on writes opts.KeepAlive = 30 // Keepalive every 10 seconds so we quickly detect network outages opts.PingTimeout = time.Second // local broker so response should be quick // Automate connection management (will keep trying to connect and will reconnect if network drops) opts.ConnectRetry = true opts.AutoReconnect = true // If using QOS2 and CleanSession = FALSE then it is possible that we will receive messages on topics that we // have not subscribed to here (if they were previously subscribed to they are part of the session and survive // disconnect/reconnect). Adding a DefaultPublishHandler lets us detect this. opts.DefaultPublishHandler = func(_ mqtt.Client, msg mqtt.Message) { fmt.Printf("UNEXPECTED MESSAGE: %s\n", msg) } // Log events opts.OnConnectionLost = func(cl mqtt.Client, err error) { fmt.Println("connection lost") } opts.OnConnect = func(c mqtt.Client) { fmt.Println("connection established") // Establish the subscription - doing this here means that it willSUB happen every time a connection is established // (useful if opts.CleanSession is TRUE or the broker does not reliably store session data) t := c.Subscribe(config.Config.Mqtt.SUBTOPIC, config.Config.Mqtt.QOS, h.handle) // the connection handler is called in a goroutine so blocking here would hot cause an issue. However as blocking // in other handlers does cause problems its best to just assume we should not block go func() { _ = t.Wait() // Can also use '<-t.Done()' in releases > 1.2.0 if t.Error() != nil { fmt.Printf("ERROR SUBSCRIBING: %s\n", t.Error()) } else { fmt.Println("subscribed to: ", config.Config.Mqtt.SUBTOPIC) } }() } opts.OnReconnecting = func(mqtt.Client, *mqtt.ClientOptions) { fmt.Println("attempting to reconnect") } // // Connect to the broker // client := mqtt.NewClient(opts) // If using QOS2 and CleanSession = FALSE then messages may be transmitted to us before the subscribe completes. // Adding routes prior to connecting is a way of ensuring that these messages are processed client.AddRoute(config.Config.Mqtt.SUBTOPIC, h.handle) if token := client.Connect(); token.Wait() && token.Error() != nil { panic(token.Error()) } fmt.Println("Connection is up") done := make(chan struct{}) var wg sync.WaitGroup wg.Add(1) go func() { var count uint64 for { select { case data := <-msgChans: msg, err := json.Marshal(data.Msg) if err != nil { panic(err) } //t := client.Publish(Config.Mqtt.PUBTOPIC+"/"+Config.Report.SN, Config.Mqtt.QOS, false, msg) t := client.Publish(config.Config.Mqtt.PUBTOPIC+"/"+data.Topic, config.Config.Mqtt.QOS, false, msg) go func() { _ = t.Wait() // Can also use '<-t.Done()' in releases > 1.2.0 if t.Error() != nil { fmt.Printf("msg PUBLISHING: %s\n", t.Error().Error())
tMqtt() { // Enable logging by uncommenting the below // mqtt.ERROR = log.New(os.Stdout, "[ERROR] ", 0) // mqtt.CRITICAL = log.New(os.Stdou
identifier_body
mqttclient.go
o *handler) handle(client mqtt.Client, msg mqtt.Message) { // We extract the count and write that out first to simplify checking for missing values var m Message var resp Session if err := json.Unmarshal(msg.Payload(), &resp); err != nil { fmt.Printf("Message could not be parsed (%s): %s", msg.Payload(), err) return } fmt.Println(resp) switch resp.Type { case CMDMSG_OFFER: enc.Decode(resp.Data, &m) Notice(m) case CMDMSG_DISC: var devcmd DiscoveryCmd enc.Decode(resp.Data, &devcmd) DiscoveryDev(&devcmd) case CMDMSG_WAKE: var fing Fing enc.Decode(resp.Data, &fing) wakemac(fing) case CMDMSG_UPDATE: var newver *versionUpdate GetUpdateMyself(newver) case CMDMSG_MR2: var mr2info Mr2Msg enc.Decode(resp.Data, &mr2info) Mr2HostPort(&mr2info) } } func Mr2HostPort(mr2info *Mr2Msg) { arg := fmt.Sprintf("client -s %s -p %s -P %d -c %s", mr2info.ServerAddr, mr2info.Password, mr2info.ExposePort, mr2info.ExposeAddr) fmt.Println("mr2", arg) err := fmt.Errorf("") //err := sys.CmdRun("./mr2", arg) if err != nil { CmdFeedBack(CMDMSG_MR2, 0, err.Error(), time.Now().String()) return } else { CmdFeedBack(CMDMSG_MR2, 1, "成功", time.Now().String()) } } func wakemac(fing Fing) { for _, v := range fing.Devices { wol.Wake(v.Mac, "", "", "") } } func DiscoveryDev(devcmd *DiscoveryCmd) { go func() { switch devcmd.DevType { case DEVICE_IP: dev := DiscoveryDevice() req := &Session{} req.Type = "discoveryrsp" req.DeviceId = "kvm1" req.Data = enc.Encode(dev) //enc.Encode(answer) answermsg := PublishMsg{ Topic: "discoveryrsp", Msg: req, } fmt.Println("discoveryrsp", answermsg) SendMsg(answermsg) //response) case DEVICE_ONVIF: case DEVICE_SNMP: case DEVICE_MODBUS: case DEVICE_BACNET: case DEVICE_CAN: case DEVICE_UPCA: } }() } func CmdFeedBack(cmdstr string, status int, err string, sid string) { resp := ResponseMsg{ Cmdstr: cmdstr, Status: status, Err: err, Sid: sid, } req := &Session{} req.Type = "cmdFeedback" req.DeviceId = "kvm1" req.Data = enc.Encode(resp) //enc.Encode(answer) answermsg := PublishMsg{ Topic: "cmdFeedback", Msg: req, } fmt.Println("cmdFeedback", answermsg) SendMsg(answermsg) //response) } func GetCurrentPath() string { getwd, err := os.Getwd() if err != nil { fmt.Print(err.Error()) } else { fmt.Print(getwd) } return getwd } func GetUpdateMyself(newver *versionUpdate) { if newver.ForceUpdate == 1 { if newver.DownLoadUrl != "" { go func() { filepath := GetCurrentPath() + "/" + newver.Version fileext := ".zip" filename, err := DownloadFile(filepath, newver.DownLoadUrl, fileext) if err != nil { CmdFeedBack(CMDMSG_UPDATE, 0, err.Error(), "1") } else { fmt.Println("Download Finished") CmdFeedBack(CMDMSG_UPDATE, 1, "Download Finished", "1") if IsZip(filename) { err = Unzip(filename, filepath) if err != nil { CmdFeedBack(CMDMSG_UPDATE, 0, err.Error(), "1") } else { CmdFeedBack(CMDMSG_UPDATE, 1, "zip ok", "1") } } //以版本号建立新目录并解压包 //后期写更新配置文件告诉Process守护进程需要更新 //并立即退出或者按某种策略更新 //守护进程在判断到需要更新软件时就按更新配置文件的新路径执行程序 } }() } } } /* func SendMsgAnswer(msg Answer) { msgChans <- msg fmt.Print("SendMsg OK") //mqttclient.Publish(Config.Mqtt.PUBTOPIC+"/"+Config.Report.SN, Config.Mqtt.QOS, false, msg) } */ func SendMsg(msg PublishMsg) { msgChans <- msg fmt.Print("SendMsg OK") //mqttclient.Publish(Config.Mqtt.PUBTOPIC+"/"+Config.Report.SN, Config.Mqtt.QOS, false, msg) } func StartMqtt() { // Enable logging by uncommenting the below // mqtt.ERROR = log.New(os.Stdout, "[ERROR] ", 0) // mqtt.CRITICAL = log.New(os.Stdout, "[CRIT
0) // mqtt.WARN = log.New(os.Stdout, "[WARN] ", 0) // mqtt.DEBUG = log.New(os.Stdout, "[DEBUG] ", 0) // Create a handler that will deal with incoming messages h := NewHandler() defer h.Close() msgChans = make(chan PublishMsg, 10) // Now we establish the connection to the mqtt broker opts := mqtt.NewClientOptions() opts.AddBroker(config.Config.Mqtt.SERVERADDRESS) opts.SetClientID(config.Config.Mqtt.CLIENTID) opts.ConnectTimeout = time.Second // Minimal delays on connect opts.WriteTimeout = time.Second // Minimal delays on writes opts.KeepAlive = 30 // Keepalive every 10 seconds so we quickly detect network outages opts.PingTimeout = time.Second // local broker so response should be quick // Automate connection management (will keep trying to connect and will reconnect if network drops) opts.ConnectRetry = true opts.AutoReconnect = true // If using QOS2 and CleanSession = FALSE then it is possible that we will receive messages on topics that we // have not subscribed to here (if they were previously subscribed to they are part of the session and survive // disconnect/reconnect). Adding a DefaultPublishHandler lets us detect this. opts.DefaultPublishHandler = func(_ mqtt.Client, msg mqtt.Message) { fmt.Printf("UNEXPECTED MESSAGE: %s\n", msg) } // Log events opts.OnConnectionLost = func(cl mqtt.Client, err error) { fmt.Println("connection lost") } opts.OnConnect = func(c mqtt.Client) { fmt.Println("connection established") // Establish the subscription - doing this here means that it willSUB happen every time a connection is established // (useful if opts.CleanSession is TRUE or the broker does not reliably store session data) t := c.Subscribe(config.Config.Mqtt.SUBTOPIC, config.Config.Mqtt.QOS, h.handle) // the connection handler is called in a goroutine so blocking here would hot cause an issue. However as blocking // in other handlers does cause problems its best to just assume we should not block go func() { _ = t.Wait() // Can also use '<-t.Done()' in releases > 1.2.0 if t.Error() != nil { fmt.Printf("ERROR SUBSCRIBING: %s\n", t.Error()) } else { fmt.Println("subscribed to: ", config.Config.Mqtt.SUBTOPIC) } }() } opts.OnReconnecting = func(mqtt.Client, *mqtt.ClientOptions) { fmt.Println("attempting to reconnect") } // // Connect to the broker // client := mqtt.NewClient(opts) // If using QOS2 and CleanSession = FALSE then messages may be transmitted to us before the subscribe completes. // Adding routes prior to connecting is a way of ensuring that these messages are processed client.AddRoute(config.Config.Mqtt.SUBTOPIC, h.handle) if token := client.Connect(); token.Wait() && token.Error() != nil { panic(token.Error()) } fmt.Println("Connection is up") done := make(chan struct{}) var wg sync.WaitGroup wg.Add(1) go func() { var count uint64 for { select { case data := <-msgChans: msg, err := json.Marshal(data.Msg) if err != nil { panic(err) } //t := client.Publish(Config.Mqtt.PUBTOPIC+"/"+Config.Report.SN, Config.Mqtt.QOS, false, msg) t := client.Publish(config.Config.Mqtt.PUBTOPIC+"/"+data.Topic, config.Config.Mqtt.QOS, false, msg) go func() { _ = t.Wait() // Can also use '<-t.Done()' in releases > 1.2.0 if t.Error() != nil { fmt.Printf("msg PUBLISHING: %s\n", t.Error().Error())
ICAL] ",
identifier_name
mqttclient.go
o.f = nil } } // handle is called when a message is received func (o *handler) handle(client mqtt.Client, msg mqtt.Message) { // We extract the count and write that out first to simplify checking for missing values var m Message var resp Session if err := json.Unmarshal(msg.Payload(), &resp); err != nil { fmt.Printf("Message could not be parsed (%s): %s", msg.Payload(), err) return } fmt.Println(resp) switch resp.Type { case CMDMSG_OFFER: enc.Decode(resp.Data, &m) Notice(m) case CMDMSG_DISC: var devcmd DiscoveryCmd enc.Decode(resp.Data, &devcmd) DiscoveryDev(&devcmd) case CMDMSG_WAKE: var fing Fing enc.Decode(resp.Data, &fing) wakemac(fing) case CMDMSG_UPDATE: var newver *versionUpdate GetUpdateMyself(newver) case CMDMSG_MR2: var mr2info Mr2Msg enc.Decode(resp.Data, &mr2info) Mr2HostPort(&mr2info) } } func Mr2HostPort(mr2info *Mr2Msg) { arg := fmt.Sprintf("client -s %s -p %s -P %d -c %s", mr2info.ServerAddr, mr2info.Password, mr2info.ExposePort, mr2info.ExposeAddr) fmt.Println("mr2", arg) err := fmt.Errorf("") //err := sys.CmdRun("./mr2", arg) if err != nil { CmdFeedBack(CMDMSG_MR2, 0, err.Error(), time.Now().String()) return } else { CmdFeedBack(CMDMSG_MR2, 1, "成功", time.Now().String()) } } func wakemac(fing Fing) { for _, v := range fing.Devices { wol.Wake(v.Mac, "", "", "") } } func DiscoveryDev(devcmd *DiscoveryCmd) { go func() { switch devcmd.DevType { case DEVICE_IP: dev := DiscoveryDevice() req := &Session{} req.Type = "discoveryrsp" req.DeviceId = "kvm1" req.Data = enc.Encode(dev) //enc.Encode(answer) answermsg := PublishMsg{ Topic: "discoveryrsp", Msg: req, } fmt.Println("discoveryrsp", answermsg) SendMsg(answermsg) //response) case DEVICE_ONVIF: case DEVICE_SNMP: case DEVICE_MODBUS: case DEVICE_BACNET: case DEVICE_CAN: case DEVICE_UPCA: } }() } func CmdFeedBack(cmdstr string, status int, err string, sid string) { resp := ResponseMsg{ Cmdstr: cmdstr, Status: status, Err: err, Sid: sid, } req := &Session{} req.Type = "cmdFeedback" req.DeviceId = "kvm1" req.Data = enc.Encode(resp) //enc.Encode(answer) answermsg := PublishMsg{ Topic: "cmdFeedback", Msg: req, } fmt.Println("cmdFeedback", answermsg) SendMsg(answermsg) //response) } func GetCurrentPath() string { getwd, err := os.Getwd() if err != nil { fmt.Print(err.Error()) } else { fmt.Print(getwd) } return getwd } func GetUpdateMyself(newver *versionUpdate) { if newver.ForceUpdate == 1 { if newver.DownLoadUrl != "" { go func() { filepath := GetCurrentPath() + "/" + newver.Version fileext := ".zip" filename, err := DownloadFile(filepath, newver.DownLoadUrl, fileext) if err != nil { CmdFeedBack(CMDMSG_UPDATE, 0, err.Error(), "1") } else { fmt.Println("Download Finished") CmdFeedBack(CMDMSG_UPDATE, 1, "Download Finished", "1") if IsZip(filename) { err = Unzip(filename, filepath) if err != nil { CmdFeedBack(CMDMSG_UPDATE, 0, err.Error(), "1") } else { CmdFeedBack(CMDMSG_UPDATE, 1, "zip ok", "1") } } //以版本号建立新目录并解压包 //后期写更新配置文件告诉Process守护进程需要更新 //并立即退出或者按某种策略更新 //守护进程在判断到需要更新软件时就按更新配置文件的新路径执行程序 } }() } } } /* func SendMsgAnswer(msg Answer) { msgChans <- msg fmt.Print("SendMsg OK") //mqttclient.Publish(Config.Mqtt.PUBTOPIC+"/"+Config.Report.SN, Config.Mqtt.QOS, false, msg) } */ func SendMsg(msg PublishMsg) { msgChans <- msg fmt.Print("SendMsg OK") //mqttclient.Publish(Config.Mqtt.PUBTOPIC+"/"+Config.Report.SN, Config.Mqtt.QOS, false, msg) } func StartMqtt() { // Enable logging by uncommenting the below // mqtt.ERROR = log.New(os.Stdout, "[ERROR] ", 0) // mqtt.CRITICAL = log.New(os.Stdout, "[CRITICAL] ", 0) // mqtt.WARN = log.New(os.Stdout, "[WARN] ", 0) // mqtt.DEBUG = log.New(os.Stdout, "[DEBUG] ", 0) // Create a handler that will deal with incoming messages h := NewHandler() defer h.Close() msgChans = make(chan PublishMsg, 10) // Now we establish the connection to the mqtt broker opts := mqtt.NewClientOptions() opts.AddBroker(config.Config.Mqtt.SERVERADDRESS) opts.SetClientID(config.Config.Mqtt.CLIENTID) opts.ConnectTimeout = time.Second // Minimal delays on connect opts.WriteTimeout = time.Second // Minimal delays on writes opts.KeepAlive = 30 // Keepalive every 10 seconds so we quickly detect network outages opts.PingTimeout = time.Second // local broker so response should be quick // Automate connection management (will keep trying to connect and will reconnect if network drops) opts.ConnectRetry = true opts.AutoReconnect = true // If using QOS2 and CleanSession = FALSE then it is possible that we will receive messages on topics that we // have not subscribed to here (if they were previously subscribed to they are part of the session and survive // disconnect/reconnect). Adding a DefaultPublishHandler lets us detect this. opts.DefaultPublishHandler = func(_ mqtt.Client, msg mqtt.Message) { fmt.Printf("UNEXPECTED MESSAGE: %s\n", msg) } // Log events opts.OnConnectionLost = func(cl mqtt.Client, err error) { fmt.Println("connection lost") } opts.OnConnect = func(c mqtt.Client) { fmt.Println("connection established") // Establish the subscription - doing this here means that it willSUB happen every time a connection is established // (useful if opts.CleanSession is TRUE or the broker does not reliably store session data) t := c.Subscribe(config.Config.Mqtt.SUBTOPIC, config.Config.Mqtt.QOS, h.handle) // the connection handler is called in a goroutine so blocking here would hot cause an issue. However as blocking // in other handlers does cause problems its best to just assume we should not block go func() { _ = t.Wait() // Can also use '<-t.Done()' in releases > 1.2.0 if t.Error() != nil { fmt.Printf("ERROR SUBSCRIBING: %s\n", t.Error()) } else { fmt.Println("subscribed to: ", config.Config.Mqtt.SUBTOPIC) } }() } opts.OnReconnecting = func(mqtt.Client, *mqtt.ClientOptions) { fmt.Println("attempting to reconnect") } // // Connect to the broker // client := mqtt.NewClient(opts) // If using QOS2 and CleanSession = FALSE then messages may be transmitted to us before the subscribe completes. // Adding routes prior to connecting is a way of ensuring that these messages are processed client.AddRoute(config.Config.Mqtt.SUBTOPIC, h.handle) if token := client.Connect(); token.Wait() && token.Error() != nil { panic(token.Error()) } fmt.Println("Connection is up") done := make(chan struct{}) var wg sync.WaitGroup wg.Add(1) go func() { var count uint64 for { select { case data := <-msgChans: msg, err := json.Marshal(data.Msg) if err != nil { panic(err) } //t := client.Publish(Config.Mqtt.PUBTOPIC+"/"+Config.Report.SN, Config.Mqtt.QOS, false, msg) t := client.Publish(config.Config.Mqtt.PUBTOPIC+"/"+data.Topic, config.Config.Mqtt.QOS, false, msg) go func() { _ = t.Wait() // Can also use '<-t.Done()'
{ fmt.Printf("ERROR closing file: %s", err) }
conditional_block
background.js
(proxy.http !== null && proxy.https === null) { proxyRules = `http=${proxy.http},${proxyRules}` } if (proxy.http !== null && proxy.https !== null) { proxyRules = `http=${proxy.http};https=${proxy.https},${proxyRules}` } } win.webContents.session.setProxy({ proxyRules: proxyRules, proxyBypassRules: proxy && proxy.bypass ? proxy.bypass : '<local>' }, () => { if (win) { win.loadURL(winUrl) } }) win.on('closed', () => { win = null }) if (process.platform !== 'darwin') { globalShortcut.register('Alt+M', () => { const visible = win.isMenuBarVisible() win.setMenuBarVisibility(visible) }) } // Set up necessary bindings to update the menu items // based on the current language selected i18nextMainBackend.on('loaded', (loaded) => { i18nextMainBackend.changeLanguage(app.getLocale()) i18nextMainBackend.off('loaded') }) menu = createMenu(win, i18nextMainBackend) i18nextMainBackend.on('languageChanged', (lng) => { log.info('Language changed') menu = createMenu(win, i18nextMainBackend) }) if (store.get('settings.start_in_trays')) { win.hide() } } function signInInoreader () { shell.openExternal(`https://www.inoreader.com/oauth2/auth?client_id=${process.env.VUE_APP_INOREADER_CLIENT_ID}&redirect_uri=ravenreader://inoreader/auth&response_type=code&scope=read%20write&state=ravenreader`) } function signInPocketWithPopUp () { if (os.platform() === 'darwin') { consumerKey = process.env.VUE_APP_POCKET_MAC_KEY } if (os.platform() === 'win32') { consumerKey = process.env.VUE_APP_POCKET_WINDOWS_KEY } if (os.platform() === 'linux') { consumerKey = process.env.VUE_APP_POCKET_LINUX_KEY } axios .post( 'https://getpocket.com/v3/oauth/request', { consumer_key: consumerKey, redirect_uri: 'http://127.0.0.1' }, { withCredentials: true, headers: { 'Content-Type': 'application/json', 'X-Accept': 'application/json' } } ) .then(data => { code = data.data.code shell.openExternal(`https://getpocket.com/auth/authorize?request_token=${code}&redirect_uri=ravenreader://pocket/auth`) }) } function registerLocalResourceProtocol () { protocol.registerFileProtocol('local-resource', (request, callback) => { const url = request.url.replace(/^local-resource:\/\//, '') // Decode URL to prevent errors when loading filenames with UTF-8 chars or chars like "#" const decodedUrl = decodeURI(url) // Needed in case URL contains spaces try { return callback(decodedUrl) } catch (error) { console.error('ERROR: registerLocalResourceProtocol: Could not get file path:', error) } }) } function handleInoreader (url) { if (url.includes('ravenreader://inoreader/auth')) { const q = new URL(url).searchParams if (q.has('code')) { axios.post('https://www.inoreader.com/oauth2/token', { code: q.get('code'), client_id: process.env.VUE_APP_INOREADER_CLIENT_ID, client_secret: process.env.VUE_APP_INOREADER_CLIENT_SECRET, redirect_uri: 'ravenreader://inoreader/auth', scope: null, grant_type: 'authorization_code' }).then((data) => { data.data.expires_in = dayjs().add(data.data.expires_in, 'second').valueOf() win.webContents.send('inoreader-authenticated', data.data) }) } } if (url === 'ravenreader://pocket/auth') { axios .post( 'https://getpocket.com/v3/oauth/authorize', { consumer_key: consumerKey, code: code }, { withCredentials: true, headers: { 'Content-Type': 'application/json', 'X-Accept': 'application/json' } } ) .then(data => { data.data.consumer_key = consumerKey win.webContents.send('pocket-authenticated', data.data) }) } } app.setAsDefaultProtocolClient('ravenreader') const primaryInstance = app.requestSingleInstanceLock() if (!primaryInstance) { app.quit() } else { app.on('second-instance', (event, argv, cmd) => { event.preventDefault() const url = argv[argv.length - 1] if (win) { if (win.isMinimized()) { win.restore() } win.focus() } if (process.platform !== 'darwin') { handleInoreader(url) } }) } app.commandLine.appendSwitch('lang', app.getLocale()) app.commandLine.appendSwitch('disable-features', 'OutOfBlinkCors') // Quit when all windows are closed. app.on('window-all-closed', () => { // On macOS it is common for applications and their menu bar // to stay active until the user quits explicitly with Cmd + Q if (process.platform !== 'darwin') { app.quit() } else { i18nextBackend.clearMainBindings(ipcMain) } }) app.on('open-url', (event, url) => {
handleInoreader(url) }) nativeTheme.on('updated', () => { store.set('isDarkMode', nativeTheme.shouldUseDarkColors) win.webContents.send('Dark mode', { darkmode: nativeTheme.shouldUseDarkColors }) }) ipcMain.handle('article-selected', (event, status) => { const menuItemViewBrowser = menu.getMenuItemById('view-browser') const menuItemToggleFavourite = menu.getMenuItemById('toggle-favourite') const menuItemSaveOffline = menu.getMenuItemById('save-offline') const menuItemToggleRead = menu.getMenuItemById('toggle-read') menuItemViewBrowser.enabled = true menuItemToggleFavourite.enabled = true menuItemSaveOffline.enabled = true menuItemToggleRead.enabled = true }) ipcMain.on('online-status-changed', (event, status) => { event.sender.send('onlinestatus', status) }) app.on('activate', () => { // On macOS it's common to re-create a window in the app when the // dock icon is clicked and there are no other windows open. if (BrowserWindow.getAllWindows().length === 0) { createWindow() } }) app.on('web-contents-created', (event, contents) => { contents.on('will-navigate', (event, navigationUrl) => { event.preventDefault() }) }) // This method will be called when Electron has finished // initialization and is ready to create browser windows. // Some APIs can only be used after this event occurs. app.on('ready', async () => { if (isDevelopment && !process.env.IS_TEST) { // Install Vue Devtools try { await installExtension(VUEJS_DEVTOOLS) } catch (e) { console.error('Vue Devtools failed to install:', e.toString()) } } // Modify the origin for all requests to the following urls. registerLocalResourceProtocol() createWindow() }) app.whenReady().then(() => { store.set('isDarkMode', nativeTheme.shouldUseDarkColors) if (!store.has('settings.theme_option')) { store.set('settings.theme_option', 'system') } }) app.on('before-quit', () => { app.isQuiting = true globalShortcut.unregisterAll() }) // Exit cleanly on request from parent process in development mode. if (isDevelopment) { if (process.platform === 'win32') { process.on('message', (data) => { if (data === 'graceful-exit') { app.quit() } }) } else { process.on('SIGTERM', () => { app.quit() }) } } ipcMain.handle('set-feedbin-last-fetched', (event, arg) => { if (arg) { store.set('feedbin_fetched_lastime', arg) } }) ipcMain.on('get-inoreader-last', (event, arg) => { event.returnValue = store.get('inoreader_fetched_lastime') }) ipcMain.on('get-feedbin-last', (event, arg) => { event.returnValue = store.get('feedbin_fetched_lastime') }) ipcMain.on('sort-preference', (event, arg) => { event.returnValue = store.get('settings.oldestArticles', 'on') }) ipcMain.on('get-settings', (event, arg) => { const state = {} state.cronSettings = store.get('settings.cronjob', '*/5 * * * *') state.themeOption = store.get('settings.theme_option', 'system') state.oldestArticles = store.get('settings.oldestArticles', false) state.disableImages = store.get('settings.imagePreference', false) state
random_line_split
background.js
// Maximize window on startup when not in development if (!isDevelopment && win !== null) { win.maximize() } i18nextBackend.mainBindings(ipcMain, win, fs) ElectronBlocker.fromPrebuiltAdsAndTracking(fetch).then((blocker) => { blocker.enableBlockingInSession(session.defaultSession) }) win.setTouchBar(touchBar) if (process.env.WEBPACK_DEV_SERVER_URL) { // Load the url of the dev server if in development mode winUrl = process.env.WEBPACK_DEV_SERVER_URL if (!process.env.IS_TEST) win.webContents.openDevTools() } else { createProtocol('app') winUrl = 'app://./index.html' autoUpdater.checkForUpdatesAndNotify() } // Load the index.html when not in development win.loadURL(winUrl) const proxy = store.get('settings.proxy') ? store.get('settings.proxy') : null let proxyRules = 'direct://' if (proxy) { if (proxy.http !== null && proxy.https === null) { proxyRules = `http=${proxy.http},${proxyRules}` } if (proxy.http !== null && proxy.https !== null) { proxyRules = `http=${proxy.http};https=${proxy.https},${proxyRules}` } } win.webContents.session.setProxy({ proxyRules: proxyRules, proxyBypassRules: proxy && proxy.bypass ? proxy.bypass : '<local>' }, () => { if (win) { win.loadURL(winUrl) } }) win.on('closed', () => { win = null }) if (process.platform !== 'darwin') { globalShortcut.register('Alt+M', () => { const visible = win.isMenuBarVisible() win.setMenuBarVisibility(visible) }) } // Set up necessary bindings to update the menu items // based on the current language selected i18nextMainBackend.on('loaded', (loaded) => { i18nextMainBackend.changeLanguage(app.getLocale()) i18nextMainBackend.off('loaded') }) menu = createMenu(win, i18nextMainBackend) i18nextMainBackend.on('languageChanged', (lng) => { log.info('Language changed') menu = createMenu(win, i18nextMainBackend) }) if (store.get('settings.start_in_trays')) { win.hide() } } function signInInoreader () { shell.openExternal(`https://www.inoreader.com/oauth2/auth?client_id=${process.env.VUE_APP_INOREADER_CLIENT_ID}&redirect_uri=ravenreader://inoreader/auth&response_type=code&scope=read%20write&state=ravenreader`) } function signInPocketWithPopUp () { if (os.platform() === 'darwin') { consumerKey = process.env.VUE_APP_POCKET_MAC_KEY } if (os.platform() === 'win32') { consumerKey = process.env.VUE_APP_POCKET_WINDOWS_KEY } if (os.platform() === 'linux') { consumerKey = process.env.VUE_APP_POCKET_LINUX_KEY } axios .post( 'https://getpocket.com/v3/oauth/request', { consumer_key: consumerKey, redirect_uri: 'http://127.0.0.1' }, { withCredentials: true, headers: { 'Content-Type': 'application/json', 'X-Accept': 'application/json' } } ) .then(data => { code = data.data.code shell.openExternal(`https://getpocket.com/auth/authorize?request_token=${code}&redirect_uri=ravenreader://pocket/auth`) }) } function registerLocalResourceProtocol () { protocol.registerFileProtocol('local-resource', (request, callback) => { const url = request.url.replace(/^local-resource:\/\//, '') // Decode URL to prevent errors when loading filenames with UTF-8 chars or chars like "#" const decodedUrl = decodeURI(url) // Needed in case URL contains spaces try { return callback(decodedUrl) } catch (error) { console.error('ERROR: registerLocalResourceProtocol: Could not get file path:', error) } }) } function handleInoreader (url) { if (url.includes('ravenreader://inoreader/auth')) { const q = new URL(url).searchParams if (q.has('code')) { axios.post('https://www.inoreader.com/oauth2/token', { code: q.get('code'), client_id: process.env.VUE_APP_INOREADER_CLIENT_ID, client_secret: process.env.VUE_APP_INOREADER_CLIENT_SECRET, redirect_uri: 'ravenreader://inoreader/auth', scope: null, grant_type: 'authorization_code' }).then((data) => { data.data.expires_in = dayjs().add(data.data.expires_in, 'second').valueOf() win.webContents.send('inoreader-authenticated', data.data) }) } } if (url === 'ravenreader://pocket/auth') { axios .post( 'https://getpocket.com/v3/oauth/authorize', { consumer_key: consumerKey, code: code }, { withCredentials: true, headers: { 'Content-Type': 'application/json', 'X-Accept': 'application/json' } } ) .then(data => { data.data.consumer_key = consumerKey win.webContents.send('pocket-authenticated', data.data) }) } } app.setAsDefaultProtocolClient('ravenreader') const primaryInstance = app.requestSingleInstanceLock() if (!primaryInstance) { app.quit() } else { app.on('second-instance', (event, argv, cmd) => { event.preventDefault() const url = argv[argv.length - 1] if (win) { if (win.isMinimized()) { win.restore() } win.focus() } if (process.platform !== 'darwin') { handleInoreader(url) } }) } app.commandLine.appendSwitch('lang', app.getLocale()) app.commandLine.appendSwitch('disable-features', 'OutOfBlinkCors') // Quit when all windows are closed. app.on('window-all-closed', () => { // On macOS it is common for applications and their menu bar // to stay active until the user quits explicitly with Cmd + Q if (process.platform !== 'darwin') { app.quit() } else { i18nextBackend.clearMainBindings(ipcMain) } }) app.on('open-url', (event, url) => { handleInoreader(url) }) nativeTheme.on('updated', () => { store.set('isDarkMode', nativeTheme.shouldUseDarkColors) win.webContents.send('Dark mode', { darkmode: nativeTheme.shouldUseDarkColors }) }) ipcMain.handle('article-selected', (event, status) => { const menuItemViewBrowser = menu.getMenuItemById('view-browser') const menuItemToggleFavourite = menu.getMenuItemById('toggle-favourite') const menuItemSaveOffline = menu.getMenuItemById('save-offline') const menuItemToggleRead = menu.getMenuItemById('toggle-read') menuItemViewBrowser.enabled = true menuItemToggleFavourite.enabled = true menuItemSaveOffline.enabled = true menuItemToggleRead.enabled = true }) ipcMain.on('online-status-changed', (event, status) => { event.sender.send('onlinestatus', status) }) app.on('activate', () => { // On macOS it's common to re-create a window in the app when the // dock icon is clicked and there are no other windows open. if (BrowserWindow.getAllWindows().length === 0) { createWindow() } }) app.on('web-contents-created', (event, contents) => { contents.on('will-navigate', (event, navigationUrl) => { event.preventDefault() }) }) // This method will be called when Electron has finished // initialization and is ready to create browser windows. // Some APIs can only be used after this event occurs. app.on('ready', async () => { if (isDevelopment && !process.env.IS_TEST) { // Install Vue Devtools try { await installExtension(VUEJS_DEVTOOLS) } catch (e) { console.error('Vue Devtools failed to install:', e.toString()) } } // Modify the origin for all requests to the following urls. registerLocalResourceProtocol
{ // Create the browser window. win = new BrowserWindow({ minWidth: 1280, minHeight: 720, width: 1400, height: 768, title: 'Raven Reader', maximizable: true, webPreferences: { // Use pluginOptions.nodeIntegration, leave this alone // See nklayman.github.io/vue-cli-plugin-electron-builder/guide/security.html#node-integration for more info webviewTag: true, contextIsolation: true, nodeIntegration: process.env.ELECTRON_NODE_INTEGRATION, preload: path.join(__dirname, 'preload.js'), disableBlinkFeatures: 'Auxclick' } })
identifier_body
background.js
i18nextBackend.mainBindings(ipcMain, win, fs) ElectronBlocker.fromPrebuiltAdsAndTracking(fetch).then((blocker) => { blocker.enableBlockingInSession(session.defaultSession) }) win.setTouchBar(touchBar) if (process.env.WEBPACK_DEV_SERVER_URL) { // Load the url of the dev server if in development mode winUrl = process.env.WEBPACK_DEV_SERVER_URL if (!process.env.IS_TEST) win.webContents.openDevTools() } else { createProtocol('app') winUrl = 'app://./index.html' autoUpdater.checkForUpdatesAndNotify() } // Load the index.html when not in development win.loadURL(winUrl) const proxy = store.get('settings.proxy') ? store.get('settings.proxy') : null let proxyRules = 'direct://' if (proxy) { if (proxy.http !== null && proxy.https === null) { proxyRules = `http=${proxy.http},${proxyRules}` } if (proxy.http !== null && proxy.https !== null) { proxyRules = `http=${proxy.http};https=${proxy.https},${proxyRules}` } } win.webContents.session.setProxy({ proxyRules: proxyRules, proxyBypassRules: proxy && proxy.bypass ? proxy.bypass : '<local>' }, () => { if (win) { win.loadURL(winUrl) } }) win.on('closed', () => { win = null }) if (process.platform !== 'darwin') { globalShortcut.register('Alt+M', () => { const visible = win.isMenuBarVisible() win.setMenuBarVisibility(visible) }) } // Set up necessary bindings to update the menu items // based on the current language selected i18nextMainBackend.on('loaded', (loaded) => { i18nextMainBackend.changeLanguage(app.getLocale()) i18nextMainBackend.off('loaded') }) menu = createMenu(win, i18nextMainBackend) i18nextMainBackend.on('languageChanged', (lng) => { log.info('Language changed') menu = createMenu(win, i18nextMainBackend) }) if (store.get('settings.start_in_trays')) { win.hide() } } function signInInoreader () { shell.openExternal(`https://www.inoreader.com/oauth2/auth?client_id=${process.env.VUE_APP_INOREADER_CLIENT_ID}&redirect_uri=ravenreader://inoreader/auth&response_type=code&scope=read%20write&state=ravenreader`) } function signInPocketWithPopUp () { if (os.platform() === 'darwin') { consumerKey = process.env.VUE_APP_POCKET_MAC_KEY } if (os.platform() === 'win32') { consumerKey = process.env.VUE_APP_POCKET_WINDOWS_KEY } if (os.platform() === 'linux') { consumerKey = process.env.VUE_APP_POCKET_LINUX_KEY } axios .post( 'https://getpocket.com/v3/oauth/request', { consumer_key: consumerKey, redirect_uri: 'http://127.0.0.1' }, { withCredentials: true, headers: { 'Content-Type': 'application/json', 'X-Accept': 'application/json' } } ) .then(data => { code = data.data.code shell.openExternal(`https://getpocket.com/auth/authorize?request_token=${code}&redirect_uri=ravenreader://pocket/auth`) }) } function registerLocalResourceProtocol () { protocol.registerFileProtocol('local-resource', (request, callback) => { const url = request.url.replace(/^local-resource:\/\//, '') // Decode URL to prevent errors when loading filenames with UTF-8 chars or chars like "#" const decodedUrl = decodeURI(url) // Needed in case URL contains spaces try { return callback(decodedUrl) } catch (error) { console.error('ERROR: registerLocalResourceProtocol: Could not get file path:', error) } }) } function handleInoreader (url) { if (url.includes('ravenreader://inoreader/auth')) { const q = new URL(url).searchParams if (q.has('code')) { axios.post('https://www.inoreader.com/oauth2/token', { code: q.get('code'), client_id: process.env.VUE_APP_INOREADER_CLIENT_ID, client_secret: process.env.VUE_APP_INOREADER_CLIENT_SECRET, redirect_uri: 'ravenreader://inoreader/auth', scope: null, grant_type: 'authorization_code' }).then((data) => { data.data.expires_in = dayjs().add(data.data.expires_in, 'second').valueOf() win.webContents.send('inoreader-authenticated', data.data) }) } } if (url === 'ravenreader://pocket/auth') { axios .post( 'https://getpocket.com/v3/oauth/authorize', { consumer_key: consumerKey, code: code }, { withCredentials: true, headers: { 'Content-Type': 'application/json', 'X-Accept': 'application/json' } } ) .then(data => { data.data.consumer_key = consumerKey win.webContents.send('pocket-authenticated', data.data) }) } } app.setAsDefaultProtocolClient('ravenreader') const primaryInstance = app.requestSingleInstanceLock() if (!primaryInstance) { app.quit() } else { app.on('second-instance', (event, argv, cmd) => { event.preventDefault() const url = argv[argv.length - 1] if (win) { if (win.isMinimized()) { win.restore() } win.focus() } if (process.platform !== 'darwin') { handleInoreader(url) } }) } app.commandLine.appendSwitch('lang', app.getLocale()) app.commandLine.appendSwitch('disable-features', 'OutOfBlinkCors') // Quit when all windows are closed. app.on('window-all-closed', () => { // On macOS it is common for applications and their menu bar // to stay active until the user quits explicitly with Cmd + Q if (process.platform !== 'darwin') { app.quit() } else { i18nextBackend.clearMainBindings(ipcMain) } }) app.on('open-url', (event, url) => { handleInoreader(url) }) nativeTheme.on('updated', () => { store.set('isDarkMode', nativeTheme.shouldUseDarkColors) win.webContents.send('Dark mode', { darkmode: nativeTheme.shouldUseDarkColors }) }) ipcMain.handle('article-selected', (event, status) => { const menuItemViewBrowser = menu.getMenuItemById('view-browser') const menuItemToggleFavourite = menu.getMenuItemById('toggle-favourite') const menuItemSaveOffline = menu.getMenuItemById('save-offline') const menuItemToggleRead = menu.getMenuItemById('toggle-read') menuItemViewBrowser.enabled = true menuItemToggleFavourite.enabled = true menuItemSaveOffline.enabled = true menuItemToggleRead.enabled = true }) ipcMain.on('online-status-changed', (event, status) => { event.sender.send('onlinestatus', status) }) app.on('activate', () => { // On macOS it's common to re-create a window in the app when the // dock icon is clicked and there are no other windows open. if (BrowserWindow.getAllWindows().length === 0) { createWindow() } }) app.on('web-contents-created', (event, contents) => { contents.on('will-navigate', (event, navigationUrl) => { event.preventDefault() }) }) // This method will be called when Electron has finished // initialization and is ready to create browser windows. // Some APIs can only be used after this event occurs. app.on('ready', async () => { if (isDevelopment && !process.env.IS_TEST) { // Install Vue Devtools try { await installExtension(VUEJS_DEVTOOLS) } catch (e) { console.error('Vue Devtools failed to install:', e.toString()) } } // Modify the origin for all requests to the following urls. registerLocalResourceProtocol() createWindow() }) app.whenReady().then(() => { store.set('isDarkMode', nativeTheme.shouldUseDarkColors) if (!store.has('settings.theme_option')) { store.set('settings.theme_option', 'system') } }) app.on('before-quit', () => { app.isQuiting = true globalShortcut.unregisterAll() }) // Exit cleanly on request from parent process in development mode. if (isDevelopment) { if (process.platform === 'win32') { process.on('message', (data) => { if (data === 'graceful-exit') { app.quit() } }) } else { process.on('SIGTERM', () => { app.quit() }) } } ipcMain.handle('set-feedbin-last-fetched', (event, arg) =>
{ win.maximize() }
conditional_block
background.js
() { // Create the browser window. win = new BrowserWindow({ minWidth: 1280, minHeight: 720, width: 1400, height: 768, title: 'Raven Reader', maximizable: true, webPreferences: { // Use pluginOptions.nodeIntegration, leave this alone // See nklayman.github.io/vue-cli-plugin-electron-builder/guide/security.html#node-integration for more info webviewTag: true, contextIsolation: true, nodeIntegration: process.env.ELECTRON_NODE_INTEGRATION, preload: path.join(__dirname, 'preload.js'), disableBlinkFeatures: 'Auxclick' } }) // Maximize window on startup when not in development if (!isDevelopment && win !== null) { win.maximize() } i18nextBackend.mainBindings(ipcMain, win, fs) ElectronBlocker.fromPrebuiltAdsAndTracking(fetch).then((blocker) => { blocker.enableBlockingInSession(session.defaultSession) }) win.setTouchBar(touchBar) if (process.env.WEBPACK_DEV_SERVER_URL) { // Load the url of the dev server if in development mode winUrl = process.env.WEBPACK_DEV_SERVER_URL if (!process.env.IS_TEST) win.webContents.openDevTools() } else { createProtocol('app') winUrl = 'app://./index.html' autoUpdater.checkForUpdatesAndNotify() } // Load the index.html when not in development win.loadURL(winUrl) const proxy = store.get('settings.proxy') ? store.get('settings.proxy') : null let proxyRules = 'direct://' if (proxy) { if (proxy.http !== null && proxy.https === null) { proxyRules = `http=${proxy.http},${proxyRules}` } if (proxy.http !== null && proxy.https !== null) { proxyRules = `http=${proxy.http};https=${proxy.https},${proxyRules}` } } win.webContents.session.setProxy({ proxyRules: proxyRules, proxyBypassRules: proxy && proxy.bypass ? proxy.bypass : '<local>' }, () => { if (win) { win.loadURL(winUrl) } }) win.on('closed', () => { win = null }) if (process.platform !== 'darwin') { globalShortcut.register('Alt+M', () => { const visible = win.isMenuBarVisible() win.setMenuBarVisibility(visible) }) } // Set up necessary bindings to update the menu items // based on the current language selected i18nextMainBackend.on('loaded', (loaded) => { i18nextMainBackend.changeLanguage(app.getLocale()) i18nextMainBackend.off('loaded') }) menu = createMenu(win, i18nextMainBackend) i18nextMainBackend.on('languageChanged', (lng) => { log.info('Language changed') menu = createMenu(win, i18nextMainBackend) }) if (store.get('settings.start_in_trays')) { win.hide() } } function signInInoreader () { shell.openExternal(`https://www.inoreader.com/oauth2/auth?client_id=${process.env.VUE_APP_INOREADER_CLIENT_ID}&redirect_uri=ravenreader://inoreader/auth&response_type=code&scope=read%20write&state=ravenreader`) } function signInPocketWithPopUp () { if (os.platform() === 'darwin') { consumerKey = process.env.VUE_APP_POCKET_MAC_KEY } if (os.platform() === 'win32') { consumerKey = process.env.VUE_APP_POCKET_WINDOWS_KEY } if (os.platform() === 'linux') { consumerKey = process.env.VUE_APP_POCKET_LINUX_KEY } axios .post( 'https://getpocket.com/v3/oauth/request', { consumer_key: consumerKey, redirect_uri: 'http://127.0.0.1' }, { withCredentials: true, headers: { 'Content-Type': 'application/json', 'X-Accept': 'application/json' } } ) .then(data => { code = data.data.code shell.openExternal(`https://getpocket.com/auth/authorize?request_token=${code}&redirect_uri=ravenreader://pocket/auth`) }) } function registerLocalResourceProtocol () { protocol.registerFileProtocol('local-resource', (request, callback) => { const url = request.url.replace(/^local-resource:\/\//, '') // Decode URL to prevent errors when loading filenames with UTF-8 chars or chars like "#" const decodedUrl = decodeURI(url) // Needed in case URL contains spaces try { return callback(decodedUrl) } catch (error) { console.error('ERROR: registerLocalResourceProtocol: Could not get file path:', error) } }) } function handleInoreader (url) { if (url.includes('ravenreader://inoreader/auth')) { const q = new URL(url).searchParams if (q.has('code')) { axios.post('https://www.inoreader.com/oauth2/token', { code: q.get('code'), client_id: process.env.VUE_APP_INOREADER_CLIENT_ID, client_secret: process.env.VUE_APP_INOREADER_CLIENT_SECRET, redirect_uri: 'ravenreader://inoreader/auth', scope: null, grant_type: 'authorization_code' }).then((data) => { data.data.expires_in = dayjs().add(data.data.expires_in, 'second').valueOf() win.webContents.send('inoreader-authenticated', data.data) }) } } if (url === 'ravenreader://pocket/auth') { axios .post( 'https://getpocket.com/v3/oauth/authorize', { consumer_key: consumerKey, code: code }, { withCredentials: true, headers: { 'Content-Type': 'application/json', 'X-Accept': 'application/json' } } ) .then(data => { data.data.consumer_key = consumerKey win.webContents.send('pocket-authenticated', data.data) }) } } app.setAsDefaultProtocolClient('ravenreader') const primaryInstance = app.requestSingleInstanceLock() if (!primaryInstance) { app.quit() } else { app.on('second-instance', (event, argv, cmd) => { event.preventDefault() const url = argv[argv.length - 1] if (win) { if (win.isMinimized()) { win.restore() } win.focus() } if (process.platform !== 'darwin') { handleInoreader(url) } }) } app.commandLine.appendSwitch('lang', app.getLocale()) app.commandLine.appendSwitch('disable-features', 'OutOfBlinkCors') // Quit when all windows are closed. app.on('window-all-closed', () => { // On macOS it is common for applications and their menu bar // to stay active until the user quits explicitly with Cmd + Q if (process.platform !== 'darwin') { app.quit() } else { i18nextBackend.clearMainBindings(ipcMain) } }) app.on('open-url', (event, url) => { handleInoreader(url) }) nativeTheme.on('updated', () => { store.set('isDarkMode', nativeTheme.shouldUseDarkColors) win.webContents.send('Dark mode', { darkmode: nativeTheme.shouldUseDarkColors }) }) ipcMain.handle('article-selected', (event, status) => { const menuItemViewBrowser = menu.getMenuItemById('view-browser') const menuItemToggleFavourite = menu.getMenuItemById('toggle-favourite') const menuItemSaveOffline = menu.getMenuItemById('save-offline') const menuItemToggleRead = menu.getMenuItemById('toggle-read') menuItemViewBrowser.enabled = true menuItemToggleFavourite.enabled = true menuItemSaveOffline.enabled = true menuItemToggleRead.enabled = true }) ipcMain.on('online-status-changed', (event, status) => { event.sender.send('onlinestatus', status) }) app.on('activate', () => { // On macOS it's common to re-create a window in the app when the // dock icon is clicked and there are no other windows open. if (BrowserWindow.getAllWindows().length === 0) { createWindow() } }) app.on('web-contents-created', (event, contents) => { contents.on('will-navigate', (event, navigationUrl) => { event.preventDefault() }) }) // This method will be called when Electron has finished // initialization and is ready to create browser windows. // Some APIs can only be used after this event occurs. app.on('ready', async () => { if (isDevelopment && !process.env.IS_TEST) { // Install Vue Devtools try { await installExtension(VUEJS_DEVTOOLS) } catch (e) { console.error('Vue Devtools failed to install:', e.toString()) } } // Modify the origin for all requests to the following urls. register
createWindow
identifier_name
click.component.ts
: ['./click.component.scss'] }) export class ClickComponent implements OnInit { @ViewChild(DatatableComponent) newstable:DatatableComponent; @ViewChild('form') form; value:any; serviceHost = BackendHost; bannerblockrows = []; paring=[]; bannerblocktemp = []; bannerblocklist = []; modalRef: NgbModalRef; img_selected=false; closeResult: string; ban:any={} banner_Status; currencyObj; question=""; select:any={}; answer=""; limits = [ { key: '5', value: 5 }, { key: '10', value: 10 }, { key: '25', value: 25 }, { key: '50', value: 50 }, { key: '100', value: 100 } ]; limit: number = this.limits[0].value; rowLimits: Array<any> = this.limits; page = { size: this.limits[0].value,totalElements:0,totalPages:0,pageNumber:0 } defsort: any = {dir: "desc", prop: "datetime"}; defsearch = ""; loading: boolean = false; showtblview: boolean = true; limits_banner = [ { key: '5', value: 5 }, { key: '10', value: 10 }, { key: '25', value: 25 }, { key: '50', value: 50 }, { key: '100', value: 100 } ]; limit_banner: number = this.limits_banner[0].value; rowLimits_news: Array<any> = this.limits_banner; page_banner = { size: this.limits[0].value,totalElements:0,totalPages:0,pageNumber:0 } defsort_banner: any = {dir: "desc", prop: "datetime"}; defsearch_news = ""; loading_banner: boolean = false; constructor( private modalService: NgbModal,private router: Router, private CommonService: CommonService, private toastr: ToastrService, private http:Http, private cookieService: CookieService ){ this.newsloadfn(); } ngOnInit() { var data = {"_id":this.cookieService.get('session')} this.CommonService.requestData('admin/getsubadmin',data).subscribe(resData => { if(resData.token_details == 1){ return true } else{ this.router.navigate(['/pages/dashboard']) } }) this.ban.base_coin ="ETH" } newsloadfn(){ var lstinput = {"page":this.page_banner,"sorting":this.defsort_banner,"search": this.defsearch_news}; this.loading_banner = true; this.CommonService.requestData('token/gettokendetails',lstinput) .subscribe(resData => { this.page_banner.totalElements = resData.bannerTotalCount; this.page_banner.totalPages = this.page_banner.totalElements / this.page_banner.size; this.bannerblocklist = resData.data; this.bannerblocktemp = this.bannerblocklist; this.bannerblockrows = this.bannerblocklist; this.loading_banner = false; }); } //reset the banner list resetnewslist(){ this.limits_banner = [ { key: '5', value: 5 }, { key: '10', value: 10 }, { key: '25', value: 25 }, { key: '50', value: 50 }, { key: '100',value: 100 } ]; this.limit_banner = this.limits_banner[0].value; this.rowLimits_news = this.limits_banner; this.page_banner = { size: this.limits_banner[0].value,totalElements:0,totalPages:0,pageNumber:0 } this.defsort_banner= {dir: "desc", prop: "datetime"}; this.defsearch_news = ""; this.newsloadfn(); } //end of the function setPage_banner(pageInfo){ this.bannerblocklist = []; this.bannerblockrows = this.bannerblocklist; this.bannerblocktemp = this.bannerblocklist; this.page_banner.pageNumber = pageInfo.offset; this.loadbannerlist(this.page_banner,this.defsort_banner,this.defsearch_news); } //sort the function loadbannerlist(page,sort,search){ this.loading_banner = true; var lstinput = {"page":page,"sorting":sort,"search":search}; this.CommonService.requestData('token/gettokendetails',lstinput) .subscribe(resData => { this.page_banner.totalElements = resData.bannerTotalCount; this.bannerblocklist = resData.data; this.bannerblockrows = this.bannerblocklist; this.bannerblocktemp = this.bannerblocklist; this.loading_banner = false; }); } onSort_banner(event) { this.page_banner.pageNumber = 0; this.defsort_banner = event.sorts[0]; this.loadbannerlist(this.page_banner,this.defsort_banner,this.defsearch_news); } //search bar function updateFilter_news() { this.limits_banner = [ { key: '5', value: 5 }, { key: '10', value: 10 }, { key: '25', value: 25 }, { key: '50', value: 50 }, { key: '100',value: 100 } ]; this.limit_banner = this.limits_banner[0].value; this.rowLimits_news = this.limits_banner; this.page_banner.pageNumber = 0; this.newsloadfn(); } //change the row limit changeRowLimits_news(event) { this.bannerblocklist = []; this.bannerblockrows = this.bannerblocklist; this.bannerblocktemp = this.bannerblocklist; this.page_banner.size = +event.target.value; this.page_banner.pageNumber = 0; this.newstable.limit = +event.target.value; this.loadbannerlist(this.page_banner,this.defsort_banner,this.defsearch_news); } //add pop up display addnewsClick(){ return true } bannerIndx = 0; opennewsmodel(content) { this.modalRef = this.modalService.open(content,{size:'lg'}); this.modalRef.result.then((result) => { this.closeResult = `Closed with: ${result}`; }, (reason) => { this.closeResult = `Dismissed ${this.getDismissReason(reason)}`; }); } //edit pop up display editbanners = {}; openeditmodel(content,bannerrow) { this.editbanners = bannerrow this.modalRef = this.modalService.open(content,{size:'lg'}); this.modalRef.result.then((result) => { this.closeResult = `Closed with: ${result}`; }, (reason) => { this.closeResult = `Dismissed ${this.getDismissReason(reason)}`; }); } bannerindx = -1; bannersvalue = {}; banner = ""; changebanner(content, bannerrow){ this.bannersvalue = bannerrow; this.bannerindx = bannerrow.$$index; this.modalRef = this.modalService.open(content,{size:'sm'}); this.modalRef.result.then((result) => { this.closeResult = `Closed with: ${result}`; }, (reason) => { this.closeResult = `Dismissed ${this.getDismissReason(reason)}`; }); } cancelemailtemp(){ this.loadbannerlist(this.page,this.defsort,this.defsearch); } closeAddToken(){ this.ban = {}; this.ban.base_coin = 'ETH'; } private getDismissReason(reason: any): string { if (reason === ModalDismissReasons.ESC) { return 'by pressing ESC'; } else if (reason === ModalDismissReasons.BACKDROP_CLICK) { return 'by clicking on a backdrop'; } else
} filesToUploads: Array<File> = []; urls=[]; filenames = ""; fileChangeEvents(fileInput: any) { this.filesToUploads=[]; var path = fileInput.target.files[0].type; if(path == "image/jpeg" || path == "image/gif" || path == "image/jpg" || path == "image/png") { this.filesToUploads = <Array<File>>fileInput.target.files; this.filenames = fileInput.target.files[0].name; let files = fileInput.target.files; if (files) { for (let file of files) { let reader = new FileReader(); reader.onload = (e: any) => { this.urls = e.target.result; } reader.readAsDataURL(file); } } this.img_selected=true; } else{ this.toastr.error('Please choose a right file!', 'Error'); this.filesToUploads=[]; this.urls=[]; this.filenames=''; this.img_selected=false; } } onsubmit(form: NgForm){ if(this.paring.length > 0){ if(this.img_selected == false){ this.toastr.error('
{ return 'with: ${reason}'; }
conditional_block
click.component.ts
: ['./click.component.scss'] }) export class ClickComponent implements OnInit { @ViewChild(DatatableComponent) newstable:DatatableComponent; @ViewChild('form') form; value:any; serviceHost = BackendHost; bannerblockrows = []; paring=[]; bannerblocktemp = []; bannerblocklist = []; modalRef: NgbModalRef; img_selected=false; closeResult: string; ban:any={} banner_Status; currencyObj; question=""; select:any={}; answer=""; limits = [ { key: '5', value: 5 }, { key: '10', value: 10 }, { key: '25', value: 25 }, { key: '50', value: 50 }, { key: '100', value: 100 } ]; limit: number = this.limits[0].value; rowLimits: Array<any> = this.limits; page = { size: this.limits[0].value,totalElements:0,totalPages:0,pageNumber:0 } defsort: any = {dir: "desc", prop: "datetime"}; defsearch = ""; loading: boolean = false; showtblview: boolean = true; limits_banner = [ { key: '5', value: 5 }, { key: '10', value: 10 }, { key: '25', value: 25 }, { key: '50', value: 50 }, { key: '100', value: 100 } ]; limit_banner: number = this.limits_banner[0].value; rowLimits_news: Array<any> = this.limits_banner; page_banner = { size: this.limits[0].value,totalElements:0,totalPages:0,pageNumber:0 } defsort_banner: any = {dir: "desc", prop: "datetime"}; defsearch_news = ""; loading_banner: boolean = false; constructor( private modalService: NgbModal,private router: Router, private CommonService: CommonService, private toastr: ToastrService, private http:Http, private cookieService: CookieService ){ this.newsloadfn(); } ngOnInit() { var data = {"_id":this.cookieService.get('session')} this.CommonService.requestData('admin/getsubadmin',data).subscribe(resData => { if(resData.token_details == 1){ return true } else{ this.router.navigate(['/pages/dashboard']) } }) this.ban.base_coin ="ETH" } newsloadfn(){ var lstinput = {"page":this.page_banner,"sorting":this.defsort_banner,"search": this.defsearch_news}; this.loading_banner = true; this.CommonService.requestData('token/gettokendetails',lstinput) .subscribe(resData => { this.page_banner.totalElements = resData.bannerTotalCount; this.page_banner.totalPages = this.page_banner.totalElements / this.page_banner.size; this.bannerblocklist = resData.data; this.bannerblocktemp = this.bannerblocklist; this.bannerblockrows = this.bannerblocklist; this.loading_banner = false; }); } //reset the banner list resetnewslist(){ this.limits_banner = [ { key: '5', value: 5 }, { key: '10', value: 10 }, { key: '25', value: 25 }, { key: '50', value: 50 }, { key: '100',value: 100 } ]; this.limit_banner = this.limits_banner[0].value; this.rowLimits_news = this.limits_banner; this.page_banner = {
this.defsearch_news = ""; this.newsloadfn(); } //end of the function setPage_banner(pageInfo){ this.bannerblocklist = []; this.bannerblockrows = this.bannerblocklist; this.bannerblocktemp = this.bannerblocklist; this.page_banner.pageNumber = pageInfo.offset; this.loadbannerlist(this.page_banner,this.defsort_banner,this.defsearch_news); } //sort the function loadbannerlist(page,sort,search){ this.loading_banner = true; var lstinput = {"page":page,"sorting":sort,"search":search}; this.CommonService.requestData('token/gettokendetails',lstinput) .subscribe(resData => { this.page_banner.totalElements = resData.bannerTotalCount; this.bannerblocklist = resData.data; this.bannerblockrows = this.bannerblocklist; this.bannerblocktemp = this.bannerblocklist; this.loading_banner = false; }); } onSort_banner(event) { this.page_banner.pageNumber = 0; this.defsort_banner = event.sorts[0]; this.loadbannerlist(this.page_banner,this.defsort_banner,this.defsearch_news); } //search bar function updateFilter_news() { this.limits_banner = [ { key: '5', value: 5 }, { key: '10', value: 10 }, { key: '25', value: 25 }, { key: '50', value: 50 }, { key: '100',value: 100 } ]; this.limit_banner = this.limits_banner[0].value; this.rowLimits_news = this.limits_banner; this.page_banner.pageNumber = 0; this.newsloadfn(); } //change the row limit changeRowLimits_news(event) { this.bannerblocklist = []; this.bannerblockrows = this.bannerblocklist; this.bannerblocktemp = this.bannerblocklist; this.page_banner.size = +event.target.value; this.page_banner.pageNumber = 0; this.newstable.limit = +event.target.value; this.loadbannerlist(this.page_banner,this.defsort_banner,this.defsearch_news); } //add pop up display addnewsClick(){ return true } bannerIndx = 0; opennewsmodel(content) { this.modalRef = this.modalService.open(content,{size:'lg'}); this.modalRef.result.then((result) => { this.closeResult = `Closed with: ${result}`; }, (reason) => { this.closeResult = `Dismissed ${this.getDismissReason(reason)}`; }); } //edit pop up display editbanners = {}; openeditmodel(content,bannerrow) { this.editbanners = bannerrow this.modalRef = this.modalService.open(content,{size:'lg'}); this.modalRef.result.then((result) => { this.closeResult = `Closed with: ${result}`; }, (reason) => { this.closeResult = `Dismissed ${this.getDismissReason(reason)}`; }); } bannerindx = -1; bannersvalue = {}; banner = ""; changebanner(content, bannerrow){ this.bannersvalue = bannerrow; this.bannerindx = bannerrow.$$index; this.modalRef = this.modalService.open(content,{size:'sm'}); this.modalRef.result.then((result) => { this.closeResult = `Closed with: ${result}`; }, (reason) => { this.closeResult = `Dismissed ${this.getDismissReason(reason)}`; }); } cancelemailtemp(){ this.loadbannerlist(this.page,this.defsort,this.defsearch); } closeAddToken(){ this.ban = {}; this.ban.base_coin = 'ETH'; } private getDismissReason(reason: any): string { if (reason === ModalDismissReasons.ESC) { return 'by pressing ESC'; } else if (reason === ModalDismissReasons.BACKDROP_CLICK) { return 'by clicking on a backdrop'; } else { return 'with: ${reason}'; } } filesToUploads: Array<File> = []; urls=[]; filenames = ""; fileChangeEvents(fileInput: any) { this.filesToUploads=[]; var path = fileInput.target.files[0].type; if(path == "image/jpeg" || path == "image/gif" || path == "image/jpg" || path == "image/png") { this.filesToUploads = <Array<File>>fileInput.target.files; this.filenames = fileInput.target.files[0].name; let files = fileInput.target.files; if (files) { for (let file of files) { let reader = new FileReader(); reader.onload = (e: any) => { this.urls = e.target.result; } reader.readAsDataURL(file); } } this.img_selected=true; } else{ this.toastr.error('Please choose a right file!', 'Error'); this.filesToUploads=[]; this.urls=[]; this.filenames=''; this.img_selected=false; } } onsubmit(form: NgForm){ if(this.paring.length > 0){ if(this.img_selected == false){ this.toastr.error('Please
size: this.limits_banner[0].value,totalElements:0,totalPages:0,pageNumber:0 } this.defsort_banner= {dir: "desc", prop: "datetime"};
random_line_split
click.component.ts
: ['./click.component.scss'] }) export class ClickComponent implements OnInit { @ViewChild(DatatableComponent) newstable:DatatableComponent; @ViewChild('form') form; value:any; serviceHost = BackendHost; bannerblockrows = []; paring=[]; bannerblocktemp = []; bannerblocklist = []; modalRef: NgbModalRef; img_selected=false; closeResult: string; ban:any={} banner_Status; currencyObj; question=""; select:any={}; answer=""; limits = [ { key: '5', value: 5 }, { key: '10', value: 10 }, { key: '25', value: 25 }, { key: '50', value: 50 }, { key: '100', value: 100 } ]; limit: number = this.limits[0].value; rowLimits: Array<any> = this.limits; page = { size: this.limits[0].value,totalElements:0,totalPages:0,pageNumber:0 } defsort: any = {dir: "desc", prop: "datetime"}; defsearch = ""; loading: boolean = false; showtblview: boolean = true; limits_banner = [ { key: '5', value: 5 }, { key: '10', value: 10 }, { key: '25', value: 25 }, { key: '50', value: 50 }, { key: '100', value: 100 } ]; limit_banner: number = this.limits_banner[0].value; rowLimits_news: Array<any> = this.limits_banner; page_banner = { size: this.limits[0].value,totalElements:0,totalPages:0,pageNumber:0 } defsort_banner: any = {dir: "desc", prop: "datetime"}; defsearch_news = ""; loading_banner: boolean = false; constructor( private modalService: NgbModal,private router: Router, private CommonService: CommonService, private toastr: ToastrService, private http:Http, private cookieService: CookieService ){ this.newsloadfn(); } ngOnInit() { var data = {"_id":this.cookieService.get('session')} this.CommonService.requestData('admin/getsubadmin',data).subscribe(resData => { if(resData.token_details == 1){ return true } else{ this.router.navigate(['/pages/dashboard']) } }) this.ban.base_coin ="ETH" } newsloadfn(){ var lstinput = {"page":this.page_banner,"sorting":this.defsort_banner,"search": this.defsearch_news}; this.loading_banner = true; this.CommonService.requestData('token/gettokendetails',lstinput) .subscribe(resData => { this.page_banner.totalElements = resData.bannerTotalCount; this.page_banner.totalPages = this.page_banner.totalElements / this.page_banner.size; this.bannerblocklist = resData.data; this.bannerblocktemp = this.bannerblocklist; this.bannerblockrows = this.bannerblocklist; this.loading_banner = false; }); } //reset the banner list resetnewslist(){ this.limits_banner = [ { key: '5', value: 5 }, { key: '10', value: 10 }, { key: '25', value: 25 }, { key: '50', value: 50 }, { key: '100',value: 100 } ]; this.limit_banner = this.limits_banner[0].value; this.rowLimits_news = this.limits_banner; this.page_banner = { size: this.limits_banner[0].value,totalElements:0,totalPages:0,pageNumber:0 } this.defsort_banner= {dir: "desc", prop: "datetime"}; this.defsearch_news = ""; this.newsloadfn(); } //end of the function setPage_banner(pageInfo){ this.bannerblocklist = []; this.bannerblockrows = this.bannerblocklist; this.bannerblocktemp = this.bannerblocklist; this.page_banner.pageNumber = pageInfo.offset; this.loadbannerlist(this.page_banner,this.defsort_banner,this.defsearch_news); } //sort the function loadbannerlist(page,sort,search){ this.loading_banner = true; var lstinput = {"page":page,"sorting":sort,"search":search}; this.CommonService.requestData('token/gettokendetails',lstinput) .subscribe(resData => { this.page_banner.totalElements = resData.bannerTotalCount; this.bannerblocklist = resData.data; this.bannerblockrows = this.bannerblocklist; this.bannerblocktemp = this.bannerblocklist; this.loading_banner = false; }); }
(event) { this.page_banner.pageNumber = 0; this.defsort_banner = event.sorts[0]; this.loadbannerlist(this.page_banner,this.defsort_banner,this.defsearch_news); } //search bar function updateFilter_news() { this.limits_banner = [ { key: '5', value: 5 }, { key: '10', value: 10 }, { key: '25', value: 25 }, { key: '50', value: 50 }, { key: '100',value: 100 } ]; this.limit_banner = this.limits_banner[0].value; this.rowLimits_news = this.limits_banner; this.page_banner.pageNumber = 0; this.newsloadfn(); } //change the row limit changeRowLimits_news(event) { this.bannerblocklist = []; this.bannerblockrows = this.bannerblocklist; this.bannerblocktemp = this.bannerblocklist; this.page_banner.size = +event.target.value; this.page_banner.pageNumber = 0; this.newstable.limit = +event.target.value; this.loadbannerlist(this.page_banner,this.defsort_banner,this.defsearch_news); } //add pop up display addnewsClick(){ return true } bannerIndx = 0; opennewsmodel(content) { this.modalRef = this.modalService.open(content,{size:'lg'}); this.modalRef.result.then((result) => { this.closeResult = `Closed with: ${result}`; }, (reason) => { this.closeResult = `Dismissed ${this.getDismissReason(reason)}`; }); } //edit pop up display editbanners = {}; openeditmodel(content,bannerrow) { this.editbanners = bannerrow this.modalRef = this.modalService.open(content,{size:'lg'}); this.modalRef.result.then((result) => { this.closeResult = `Closed with: ${result}`; }, (reason) => { this.closeResult = `Dismissed ${this.getDismissReason(reason)}`; }); } bannerindx = -1; bannersvalue = {}; banner = ""; changebanner(content, bannerrow){ this.bannersvalue = bannerrow; this.bannerindx = bannerrow.$$index; this.modalRef = this.modalService.open(content,{size:'sm'}); this.modalRef.result.then((result) => { this.closeResult = `Closed with: ${result}`; }, (reason) => { this.closeResult = `Dismissed ${this.getDismissReason(reason)}`; }); } cancelemailtemp(){ this.loadbannerlist(this.page,this.defsort,this.defsearch); } closeAddToken(){ this.ban = {}; this.ban.base_coin = 'ETH'; } private getDismissReason(reason: any): string { if (reason === ModalDismissReasons.ESC) { return 'by pressing ESC'; } else if (reason === ModalDismissReasons.BACKDROP_CLICK) { return 'by clicking on a backdrop'; } else { return 'with: ${reason}'; } } filesToUploads: Array<File> = []; urls=[]; filenames = ""; fileChangeEvents(fileInput: any) { this.filesToUploads=[]; var path = fileInput.target.files[0].type; if(path == "image/jpeg" || path == "image/gif" || path == "image/jpg" || path == "image/png") { this.filesToUploads = <Array<File>>fileInput.target.files; this.filenames = fileInput.target.files[0].name; let files = fileInput.target.files; if (files) { for (let file of files) { let reader = new FileReader(); reader.onload = (e: any) => { this.urls = e.target.result; } reader.readAsDataURL(file); } } this.img_selected=true; } else{ this.toastr.error('Please choose a right file!', 'Error'); this.filesToUploads=[]; this.urls=[]; this.filenames=''; this.img_selected=false; } } onsubmit(form: NgForm){ if(this.paring.length > 0){ if(this.img_selected == false){ this.toastr.error('Please select
onSort_banner
identifier_name
click.component.ts
: ['./click.component.scss'] }) export class ClickComponent implements OnInit { @ViewChild(DatatableComponent) newstable:DatatableComponent; @ViewChild('form') form; value:any; serviceHost = BackendHost; bannerblockrows = []; paring=[]; bannerblocktemp = []; bannerblocklist = []; modalRef: NgbModalRef; img_selected=false; closeResult: string; ban:any={} banner_Status; currencyObj; question=""; select:any={}; answer=""; limits = [ { key: '5', value: 5 }, { key: '10', value: 10 }, { key: '25', value: 25 }, { key: '50', value: 50 }, { key: '100', value: 100 } ]; limit: number = this.limits[0].value; rowLimits: Array<any> = this.limits; page = { size: this.limits[0].value,totalElements:0,totalPages:0,pageNumber:0 } defsort: any = {dir: "desc", prop: "datetime"}; defsearch = ""; loading: boolean = false; showtblview: boolean = true; limits_banner = [ { key: '5', value: 5 }, { key: '10', value: 10 }, { key: '25', value: 25 }, { key: '50', value: 50 }, { key: '100', value: 100 } ]; limit_banner: number = this.limits_banner[0].value; rowLimits_news: Array<any> = this.limits_banner; page_banner = { size: this.limits[0].value,totalElements:0,totalPages:0,pageNumber:0 } defsort_banner: any = {dir: "desc", prop: "datetime"}; defsearch_news = ""; loading_banner: boolean = false; constructor( private modalService: NgbModal,private router: Router, private CommonService: CommonService, private toastr: ToastrService, private http:Http, private cookieService: CookieService ){ this.newsloadfn(); } ngOnInit() { var data = {"_id":this.cookieService.get('session')} this.CommonService.requestData('admin/getsubadmin',data).subscribe(resData => { if(resData.token_details == 1){ return true } else{ this.router.navigate(['/pages/dashboard']) } }) this.ban.base_coin ="ETH" } newsloadfn(){ var lstinput = {"page":this.page_banner,"sorting":this.defsort_banner,"search": this.defsearch_news}; this.loading_banner = true; this.CommonService.requestData('token/gettokendetails',lstinput) .subscribe(resData => { this.page_banner.totalElements = resData.bannerTotalCount; this.page_banner.totalPages = this.page_banner.totalElements / this.page_banner.size; this.bannerblocklist = resData.data; this.bannerblocktemp = this.bannerblocklist; this.bannerblockrows = this.bannerblocklist; this.loading_banner = false; }); } //reset the banner list resetnewslist(){ this.limits_banner = [ { key: '5', value: 5 }, { key: '10', value: 10 }, { key: '25', value: 25 }, { key: '50', value: 50 }, { key: '100',value: 100 } ]; this.limit_banner = this.limits_banner[0].value; this.rowLimits_news = this.limits_banner; this.page_banner = { size: this.limits_banner[0].value,totalElements:0,totalPages:0,pageNumber:0 } this.defsort_banner= {dir: "desc", prop: "datetime"}; this.defsearch_news = ""; this.newsloadfn(); } //end of the function setPage_banner(pageInfo){ this.bannerblocklist = []; this.bannerblockrows = this.bannerblocklist; this.bannerblocktemp = this.bannerblocklist; this.page_banner.pageNumber = pageInfo.offset; this.loadbannerlist(this.page_banner,this.defsort_banner,this.defsearch_news); } //sort the function loadbannerlist(page,sort,search){ this.loading_banner = true; var lstinput = {"page":page,"sorting":sort,"search":search}; this.CommonService.requestData('token/gettokendetails',lstinput) .subscribe(resData => { this.page_banner.totalElements = resData.bannerTotalCount; this.bannerblocklist = resData.data; this.bannerblockrows = this.bannerblocklist; this.bannerblocktemp = this.bannerblocklist; this.loading_banner = false; }); } onSort_banner(event) { this.page_banner.pageNumber = 0; this.defsort_banner = event.sorts[0]; this.loadbannerlist(this.page_banner,this.defsort_banner,this.defsearch_news); } //search bar function updateFilter_news() { this.limits_banner = [ { key: '5', value: 5 }, { key: '10', value: 10 }, { key: '25', value: 25 }, { key: '50', value: 50 }, { key: '100',value: 100 } ]; this.limit_banner = this.limits_banner[0].value; this.rowLimits_news = this.limits_banner; this.page_banner.pageNumber = 0; this.newsloadfn(); } //change the row limit changeRowLimits_news(event) { this.bannerblocklist = []; this.bannerblockrows = this.bannerblocklist; this.bannerblocktemp = this.bannerblocklist; this.page_banner.size = +event.target.value; this.page_banner.pageNumber = 0; this.newstable.limit = +event.target.value; this.loadbannerlist(this.page_banner,this.defsort_banner,this.defsearch_news); } //add pop up display addnewsClick(){ return true } bannerIndx = 0; opennewsmodel(content) { this.modalRef = this.modalService.open(content,{size:'lg'}); this.modalRef.result.then((result) => { this.closeResult = `Closed with: ${result}`; }, (reason) => { this.closeResult = `Dismissed ${this.getDismissReason(reason)}`; }); } //edit pop up display editbanners = {}; openeditmodel(content,bannerrow) { this.editbanners = bannerrow this.modalRef = this.modalService.open(content,{size:'lg'}); this.modalRef.result.then((result) => { this.closeResult = `Closed with: ${result}`; }, (reason) => { this.closeResult = `Dismissed ${this.getDismissReason(reason)}`; }); } bannerindx = -1; bannersvalue = {}; banner = ""; changebanner(content, bannerrow){ this.bannersvalue = bannerrow; this.bannerindx = bannerrow.$$index; this.modalRef = this.modalService.open(content,{size:'sm'}); this.modalRef.result.then((result) => { this.closeResult = `Closed with: ${result}`; }, (reason) => { this.closeResult = `Dismissed ${this.getDismissReason(reason)}`; }); } cancelemailtemp(){ this.loadbannerlist(this.page,this.defsort,this.defsearch); } closeAddToken(){ this.ban = {}; this.ban.base_coin = 'ETH'; } private getDismissReason(reason: any): string { if (reason === ModalDismissReasons.ESC) { return 'by pressing ESC'; } else if (reason === ModalDismissReasons.BACKDROP_CLICK) { return 'by clicking on a backdrop'; } else { return 'with: ${reason}'; } } filesToUploads: Array<File> = []; urls=[]; filenames = ""; fileChangeEvents(fileInput: any)
this.toastr.error('Please choose a right file!', 'Error'); this.filesToUploads=[]; this.urls=[]; this.filenames=''; this.img_selected=false; } } onsubmit(form: NgForm){ if(this.paring.length > 0){ if(this.img_selected == false){ this.toastr.error('Please
{ this.filesToUploads=[]; var path = fileInput.target.files[0].type; if(path == "image/jpeg" || path == "image/gif" || path == "image/jpg" || path == "image/png") { this.filesToUploads = <Array<File>>fileInput.target.files; this.filenames = fileInput.target.files[0].name; let files = fileInput.target.files; if (files) { for (let file of files) { let reader = new FileReader(); reader.onload = (e: any) => { this.urls = e.target.result; } reader.readAsDataURL(file); } } this.img_selected=true; } else{
identifier_body
main.go
nil } func InputExpiresTime() (int64, error) { var ( err error expiresAt int64 ) fmt.Printf("%s\n", "请输入过期时间(格式:数字+单位,例如12d, 单位:天[d] 分钟[m] 秒[s] 年[y]):") input, err := inputReader.ReadString('\n') if err != nil { return 0, err } defer inputReader.Reset(os.Stdin) inputString := strings.TrimSpace(input) if inputString != "" { if strings.HasPrefix(inputString, "q") { os.Exit(0) } if strings.HasSuffix(inputString, "y") { inputString = inputString[:len(inputString)-1] years, err := strconv.ParseInt(inputString, 10, 64) // days, err := strconv.ParseFloat(inputString, 64) if err != nil { return 0, err } if years <= 0 || years > 100 { return 0, fmt.Errorf("输入年数不能小于0,大于%d年", 100) } expiresAt = years * OneYearSeconds duration := time.Duration(expiresAt) * time.Second fmt.Println() fmt.Printf("过期年数: %d years, 过期日期:%s \n", years, time.Now().Add(duration).Format("2006-01-02 15:04:05")) fmt.Println() } else if strings.HasSuffix(inputString, "d") { inputString = inputString[:len(inputString)-1] days, err := strconv.ParseInt(inputString, 10, 64) // days, err := strconv.ParseFloat(inputString, 64) if err != nil { return 0, err } if days <= 0 || days > 100*356 { return 0, fmt.Errorf("输入天数不能小于0,大于%d天", 100*356)
} expiresAt = int64(days * OneDaySeconds) duration := time.Duration(expiresAt) * time.Second fmt.Println() fmt.Printf("过期天数: %d days, 过期日期:%s \n", days, time.Now().Add(duration).Format("2006-01-02 15:04:05")) fmt.Println() } else if strings.HasSuffix(inputString, "m") { inputString = inputString[:len(inputString)-1] minute, err := strconv.ParseInt(inputString, 10, 64) if err != nil { return 0, err } if minute <= 0 || minute > 100*356*24*60 { return 0, fmt.Errorf("输入分钟不能小于0,大于%d分钟", 100*356*24*60) } expiresAt = minute * OneMinuteSeconds duration := time.Duration(expiresAt) * time.Second fmt.Println() fmt.Printf("过期分钟数: %d minute, 过期日期:%s \n", minute, time.Now().Add(duration).Format("2006-01-02 15:04:05")) fmt.Println() } else if strings.HasSuffix(inputString, "s") { inputString = inputString[:len(inputString)-1] seconds, err := strconv.ParseInt(inputString, 10, 64) if err != nil { return 0, err } if seconds <= 0 || seconds > 100*356*24*60*60 { return 0, fmt.Errorf("输入秒不能小于0,大于%d秒", 100*356*24*60*60) } expiresAt = seconds duration := time.Duration(expiresAt) * time.Second fmt.Println() fmt.Printf("过期秒数: %d seconds, 过期日期:%s \n", seconds, time.Now().Add(duration).Format("2006-01-02 15:04:05")) fmt.Println() } else { return 0, fmt.Errorf("%s", "输入不正确,请输入时间单位") } } // fmt.Println("\033[H\033[2J") return expiresAt, nil } func InputMachineID() (string, error) { fmt.Printf("%s\n", "请输入机器码:") input, err := inputReader.ReadString('\n') if err != nil { return "", err } defer inputReader.Reset(os.Stdin) inputString := strings.TrimSpace(input) if inputString != "" { if strings.HasPrefix(inputString, "q") { os.Exit(0) } fmt.Println() fmt.Printf("你输入的机器码是: %s\n", inputString) fmt.Println() } return inputString, nil } func ShowActiveCode(dir, fileName, uuid string) { fmt.Printf("序号:%s \n", uuid) fmt.Printf("\n%s\n", "激活码是:") readPath := filepath.Join(dir, fileName) licenseActive, err := public.ReadLicensePem(readPath) if err != nil { fmt.Println(err.Error()) return } fmt.Println(base64.URLEncoding.EncodeToString(licenseActive)) // fmt.Println(string(licenseActive)) } func ReadCustomKV(productName string) ([]AttrKV, error) { type Option struct { XMLName xml.Name `xml:"option"` Desc string `xml:"desc"` Key string `xml:"key"` Value string `xml:"val"` } type XMLProduct struct { XMLName xml.Name `xml:"options"` Version string `xml:"version,attr"` Options []Option `xml:"option"` } filePath := filepath.Join(DataDir, strings.Join([]string{productName, ".xml"}, "")) if public.Exists(filePath) { var ( attr = XMLProduct{} attrKV []AttrKV ) fd, err := os.OpenFile(filePath, os.O_RDONLY, 0644) if err != nil { return nil, err } defer fd.Close() attrKVBytes, err := ioutil.ReadAll(fd) if err != nil { return nil, err } err = xml.Unmarshal(attrKVBytes, &attr) if err != nil { return nil, err } // fmt.Printf("%s特性选择:\n", productExplan) for i := 0; i < len(attr.Options); i++ { // fmt.Println(i+1, attr.Options[i].Desc, attr.Options[i].Key, attr.Options[i].Value) attrKV = append(attrKV, AttrKV{Desc: attr.Options[i].Desc, Key: attr.Options[i].Key, Val: attr.Options[i].Value}) } // fmt.Println("请输入数字序号,以分号间隔:") return attrKV, nil } else { fd, err := os.OpenFile(filePath, os.O_RDWR|os.O_CREATE, 0644) if err != nil { return nil, err } defer fd.Close() attr := &XMLProduct{ Version: "1", } attr.Options = append(attr.Options, Option{ Desc: "示例配置", Key: "key1", Value: "val1", }) output, err := xml.MarshalIndent(attr, "", " ") if err != nil { fmt.Println(err) return nil, err } _, err = fd.Write([]byte(xml.Header)) if err != nil { return nil, err } _, err = fd.Write(output) if err != nil { return nil, err } } return nil, nil } func SelectCustomKV(productExplan string, kv []AttrKV) ([]AttrKV, error) { var ( arrayIdx []int kvs []AttrKV ) if kv != nil { fmt.Printf("%s启用配置选择(请输入数字序号,以逗号间隔,跳过按回车):\n", productExplan) for i := 0; i < len(kv); i++ { fmt.Println(i+1, kv[i].Desc) } input, err := inputReader.ReadString('\n') if err != nil { return nil, err } defer inputReader.Reset(os.Stdin) inputString := strings.TrimSpace(input) if inputString != "" { if strings.HasPrefix(inputString, "q") { os.Exit(0) } arrayIndx := strings.Split(inputString, ",") for i := 0; i < len(arrayIndx); i++ { num := strings.TrimSpace(arrayIndx[i]) if num == "" { continue } idx, err := strconv.Atoi(num) if err != nil {
random_line_split
main.go
} func InputExpiresTime() (int64, error) { var ( err error expiresAt int64 ) fmt.Printf("%s\n", "请输入过期时间(格式:数字+单位,例如12d, 单位:天[d] 分钟[m] 秒[s] 年[y]):") input, err := inputReader.ReadString('\n') if err != nil { return 0, err } defer inputReader.Reset(os.Stdin) inputString := strings.TrimSpace(input) if inputString != "" { if strings.HasPrefix(inputString, "q") { os.Exit(0) } if strings.HasSuffix(inputString, "y") { inputString = inputString[:len(inputString)-1] years, err := strconv.ParseInt(inputString, 10, 64) // days, err := strconv.ParseFloat(inputString, 64) if err != nil { return 0, err } if years <= 0 || years > 100 { return 0, fmt.Errorf("输入年数不能小于0,大于%d年", 100) } expiresAt = years * OneYearSeconds duration := time.Duration(expiresAt) * time.Second fmt.Println() fmt.Printf("过期年数: %d years, 过期日期:%s \n", years, time.Now().Add(duration).Format("2006-01-02 15:04:05")) fmt.Println() } else if strings.HasSuffix(inputString, "d") { inputString = inputString[:len(inputString)-1] days, err := strconv.ParseInt(inputString, 10, 64) // days, err := strconv.ParseFloat(inputString, 64) if err != nil { return 0, err } if days <= 0 || days > 100*356 { return 0, fmt.Errorf("输入天数不能小于0,大于%d天", 100*356) } expiresAt = int64(days * OneDaySeconds) duration := time.Duration(expiresAt) * time.Second fmt.Println() fmt.Printf("过期天数: %d days, 过期日期:%s \n", days, time.Now().Add(duration).Format("2006-01-02 15:04:05")) fmt.Println() } else if strings.HasSuffix(inputString, "m") { inputString = inputString[:len(inputString)-1] minute, err := strconv.ParseInt(inputString, 10, 64) if err != nil { return 0, err } if minute <= 0 || minute > 100*356*24*60 { return 0, fmt.Errorf("输入分钟不能小于0,大于%d分钟", 100*356*24*60) } expiresAt = minute * OneMinuteSeconds duration := time.Duration(expiresAt) * time.Second fmt.Println() fmt.Printf("过期分钟数: %d minute, 过期日期:%s \n", minute, time.Now().Add(duration).Format("2006-01-02 15:04:05")) fmt.Println() } else if strings.HasSuffix(inputString, "s") { inputString = inputString[:len(inputString)-1] seconds, err := strconv.ParseInt(inputString, 10, 64) if err != nil { return 0, err } if seconds <= 0 || seconds > 100*356*24*60*60 { return 0, fmt.Errorf("输入秒不能小于0,大于%d秒", 100*356*24*60*60) } expiresAt = seconds duration := time.Duration(expiresAt) * time.Second fmt.Println() fmt.Printf("过期秒数: %d seconds, 过期日期:%s \n", seconds, time.Now().Add(duration).Format("2006-01-02 15:04:05")) fmt.Println() } else { return 0, fmt.Errorf("%s", "输入不正确,请输入时间单位") } } // fmt.Println("\033[H\033[2J") return expiresAt, nil } func InputMachineID() (string, error) { fmt.Printf("%s\n", "请输入机器码:") input, err := inputReader.ReadString('\n') if err != nil { return "", err } defer inputReader.Reset(os.Stdin) inputString := strings.TrimSpace(input) if inputString != "" { if strings.HasPrefix(inputString, "q") { os.Exit(0) }
fmt.Println(base64.URLEncoding.Encod eToString(licenseActive)) // fmt.Println(string(licenseActive)) } func ReadCustomKV(productName string) ([]AttrKV, error) { type Option struct { XMLName xml.Name `xml:"option"` Desc string `xml:"desc"` Key string `xml:"key"` Value string `xml:"val"` } type XMLProduct struct { XMLName xml.Name `xml:"options"` Version string `xml:"version,attr"` Options []Option `xml:"option"` } filePath := filepath.Join(DataDir, strings.Join([]string{productName, ".xml"}, "")) if public.Exists(filePath) { var ( attr = XMLProduct{} attrKV []AttrKV ) fd, err := os.OpenFile(filePath, os.O_RDONLY, 0644) if err != nil { return nil, err } defer fd.Close() attrKVBytes, err := ioutil.ReadAll(fd) if err != nil { return nil, err } err = xml.Unmarshal(attrKVBytes, &attr) if err != nil { return nil, err } // fmt.Printf("%s特性选择:\n", productExplan) for i := 0; i < len(attr.Options); i++ { // fmt.Println(i+1, attr.Options[i].Desc, attr.Options[i].Key, attr.Options[i].Value) attrKV = append(attrKV, AttrKV{Desc: attr.Options[i].Desc, Key: attr.Options[i].Key, Val: attr.Options[i].Value}) } // fmt.Println("请输入数字序号,以分号间隔:") return attrKV, nil } else { fd, err := os.OpenFile(filePath, os.O_RDWR|os.O_CREATE, 0644) if err != nil { return nil, err } defer fd.Close() attr := &XMLProduct{ Version: "1", } attr.Options = append(attr.Options, Option{ Desc: "示例配置", Key: "key1", Value: "val1", }) output, err := xml.MarshalIndent(attr, "", " ") if err != nil { fmt.Println(err) return nil, err } _, err = fd.Write([]byte(xml.Header)) if err != nil { return nil, err } _, err = fd.Write(output) if err != nil { return nil, err } } return nil, nil } func SelectCustomKV(productExplan string, kv []AttrKV) ([]AttrKV, error) { var ( arrayIdx []int kvs []AttrKV ) if kv != nil { fmt.Printf("%s启用配置选择(请输入数字序号,以逗号间隔,跳过按回车):\n", productExplan) for i := 0; i < len(kv); i++ { fmt.Println(i+1, kv[i].Desc) } input, err := inputReader.ReadString('\n') if err != nil { return nil, err } defer inputReader.Reset(os.Stdin) inputString := strings.TrimSpace(input) if inputString != "" { if strings.HasPrefix(inputString, "q") { os.Exit(0) } arrayIndx := strings.Split(inputString, ",") for i := 0; i < len(arrayIndx); i++ { num := strings.TrimSpace(arrayIndx[i]) if num == "" { continue } idx, err := strconv.Atoi(num) if err !=
fmt.Println() fmt.Printf("你输入的机器码是: %s\n", inputString) fmt.Println() } return inputString, nil } func ShowActiveCode(dir, fileName, uuid string) { fmt.Printf("序号:%s \n", uuid) fmt.Printf("\n%s\n", "激活码是:") readPath := filepath.Join(dir, fileName) licenseActive, err := public.ReadLicensePem(readPath) if err != nil { fmt.Println(err.Error()) return }
identifier_body
main.go
} func InputExpiresTime() (int64, error) { var ( err error expiresAt int64 ) fmt.Printf("%s\n", "请输入过期时间(格式:数字+单位,例如12d, 单位:天[d] 分钟[m] 秒[s] 年[y]):") input, err := inputReader.ReadString('\n') if err != nil { return 0, err } defer inputReader.Reset(os.Stdin) inputString := strings.TrimSpace(input) if inputString != "" { if strings.HasPrefix(inputString, "q") { os.Exit(0) } if strings.HasSuffix(inputString, "y") { inputString = inputString[:len(inputString)-1] years, err := strconv.ParseInt(inputString, 10, 64) // days, err := strconv.ParseFloat(inputString, 64) if err != nil { return 0, err } if years <= 0 || years > 100 { return 0, fmt.Errorf("输入年数不能小于0,大于%d年", 100) } expiresAt = years * OneYearSeconds duration := time.Duration(expiresAt) * time.Second fmt.Println() fmt.Printf("过期年数: %d years, 过期日期:%s \n", years, time.Now().Add(duration).Format("2006-01-02 15:04:05")) fmt.Println() } else if strings.HasSuffix(inputString, "d") { inputString = inputString[:len(inputString)-1] days, err := strconv.ParseInt(inputString, 10, 64) // days, err := strconv.ParseFloat(inputString, 64) if err != nil { return 0, err } if days <= 0 || days > 100*356 { return 0, fmt.Errorf("输入天数不能小于0,大于%d天", 100*356) } expiresAt = int64(days * OneDaySeconds) duration := time.Duration(expiresAt) * time.Second fmt.Println() fmt.Printf("过期天数: %d days, 过期日期:%s \n", days, time.Now().Add(duration).Format("2006-01-02 15:04:05")) fmt.Println() } else if strings.HasSuffix(inputString, "m") { inputString = inputString[:len(inputString)-1] minute, err := strconv.ParseInt(inputString, 10, 64) if err != nil { return 0, err } if minute <= 0 || minute > 100*356*24*60 { return 0, fmt.Errorf("输入分钟不能小于0,大于%d分钟", 100*356*24*60) } expiresAt = minute * OneMinuteSeconds duration := time.Duration(expiresAt) * time.Second fmt.Println() fmt.Printf("过期分钟数: %d minute, 过期日期:%s \n", minute, time.Now().Add(duration).Format("2006-01-02 15:04:05")) fmt.Println() } else if strings.HasSuffix(inputString, "s") { inputString = inputString[:len(inputString)-1] seconds, err := strconv.ParseInt(inputString, 10, 64) if err != nil { return 0, err } if seconds <= 0 || seconds > 100*356*24*60*60 { return 0, fmt.Errorf("输入秒不能小于0,大于%d秒", 100*356*24*60*60) } expiresAt = seconds duration := time.Duration(expiresAt) * time.Second fmt.Println() fmt.Printf("过期秒数: %d seconds, 过期日期:%s \n", seconds, time.Now().Add(duration).Format("2006-01-02 15:04:05")) fmt.Println() } else { return 0, fmt.Errorf("%s", "输入不正确,请输入时间单位") } } // fmt.Println("\033[H\033[2J") return expiresAt, nil } func InputMachineID() (string, error) { fmt.Printf("%s\n", "请输入机器码:") input, err := inputReader.ReadString('\n') if err != nil { return "", err } defer inputReader.Reset(os.Stdin) inputString := strings.TrimSpace(input) if inputString != "" { if strings.HasPrefix(inputString, "q") { os.Exit(0) } fmt.Println() fmt.Printf("你输入的机器码是: %s\n", inputString) fmt.Println() } return inputString, nil } func ShowActiveCode(dir, fileName, uuid string) { fmt.Printf("序号:%s \n", uuid) fmt.Printf("\n%s\n", "激活码是:") readPath := filepath.Join(dir, fileName) licenseActive, err := public.ReadLicensePem(readPath) if err != nil { fmt.Println(err.Error()) return } fmt.Println(base64.URLEncoding.EncodeToString(licenseActive)) // fmt.Println(string(licenseActive)) } func ReadCustomKV(productName string) ([]AttrKV, error) { type Option struct { XMLName xml.Name `xml:"option"` Desc string `xml:"desc"` Key string `xml:"key"` Value string `xml:"val"` } type XMLProduct struct { XMLName xml.Name `xml:"options"` Version string `xml:"version,attr"` Options []Option `xml:"option"` } filePath := filepath.Join(DataDir, strings.Join([]string{productName, ".xml"}, "")) if public.Exists(filePath) { var ( attr = XMLProduct{} attrKV []AttrKV ) fd, err := os.OpenFile(filePath, os.O_RDONLY, 0644) if err != nil { return nil, err } defer fd.Close() attrKVBytes, err := ioutil.ReadAll(fd) if err != nil { return nil, err } err = xml.Unmarshal(attrKVBytes, &attr) if err != nil { return nil, err } // fmt.Printf("%s特性选择:\n", productExplan) for i := 0; i < len(attr.Options); i++ { // fmt.Println(i+1, attr.Options[i].Desc, attr.Options[i].Key, attr.Options[i].Value) attrKV = append(attrKV, AttrKV{Desc: attr.Options[i].Desc, Key: attr.Options[i].Key, Val: attr.Options[i].Value}) } // fmt.Println("请输入数字序号,以分号间隔:") return attrKV, nil } else { fd, err := os.OpenFile(filePath, os.O_RDWR|os.O_CREATE, 0644) if err != nil { return nil, err } defer fd.Close() attr := &XMLProduct{ Version: "1", } attr.Options = append(attr.Options, Option{ Desc: "示例配置", Key: "key1", Value: "val1", }) output, err := xml.MarshalIndent(attr, "", " ") if err != nil { fmt.Println(err) return nil, err } _, err = fd.Write([]byte(xml.Header)) if err != nil { return nil, err } _, err = fd.Write(output) if err != nil { return nil, err } } return nil, nil } func SelectCustomKV(productExplan string, kv []AttrKV) ([]AttrKV, error) { var ( arrayIdx []int kvs []AttrKV ) if kv != nil { fmt.Printf("%s启用配置选择(请输入数字序号,以逗号间隔,跳过按回车):\n", productExplan) for i := 0; i < len(kv); i++ { fmt.Println(i+1, kv[i].Desc) } input, err := inputReader.ReadStr
l { return nil, err } defer inputReader.Reset(os.Stdin) inputString := strings.TrimSpace(input) if inputString != "" { if strings.HasPrefix(inputString, "q") { os.Exit(0) } arrayIndx := strings.Split(inputString, ",") for i := 0; i < len(arrayIndx); i++ { num := strings.TrimSpace(arrayIndx[i]) if num == "" { continue } idx, err := strconv.Atoi(num) if err !=
ing('\n') if err != ni
conditional_block
main.go
} func InputExpiresTime() (int64, error) { var ( err error expiresAt int64 ) fmt.Printf("%s\n", "请输入过期时间(格式:数字+单位,例如12d, 单位:天[d] 分钟[m] 秒[s] 年[y]):") input, err := inputReader.ReadString('\n') if err != nil { return 0, err } defer inputReader.Reset(os.Stdin) inputString := strings.TrimSpace(input) if inputString != "" { if strings.HasPrefix(inputString, "q") { os.Exit(0) } if strings.HasSuffix(inputString, "y") { inputString = inputString[:len(inputString)-1] years, err := strconv.ParseInt(inputString, 10, 64) // days, err := strconv.ParseFloat(inputString, 64) if err != nil { return 0, err } if years <= 0 || years > 100 { return 0, fmt.Errorf("输入年数不能小于0,大于%d年", 100) } expiresAt = years * OneYearSeconds duration := time.Duration(expiresAt) * time.Second fmt.Println() fmt.Printf("过期年数: %d years, 过期日期:%s \n", years, time.Now().Add(duration).Format("2006-01-02 15:04:05")) fmt.Println() } else if strings.HasSuffix(inputString, "d") { inputString = inputString[:len(inputString)-1] days, err := strconv.ParseInt(inputString, 10, 64) // days, err := strconv.ParseFloat(inputString, 64) if err != nil { return 0, err } if days <= 0 || days > 100*356 { return 0, fmt.Errorf("输入天数不能小于0,大于%d天", 100*356) } expiresAt = int64(days * OneDaySeconds) duration := time.Duration(expiresAt) * time.Second fmt.Println() fmt.Printf("过期天数: %d days, 过期日期:%s \n", days, time.Now().Add(duration).Format("2006-01-02 15:04:05")) fmt.Println() } else if strings.HasSuffix(inputString, "m") { inputString = inputString[:len(inputString)-1] minute, err := strconv.ParseInt(inputString, 10, 64) if err != nil { return 0, err } if minute <= 0 || minute > 100*356*24*60 { return 0, fmt.Errorf("输入分钟不能小于0,大于%d分钟", 100*356*24*60) } expiresAt = minute * OneMinuteSeconds duration := time.Duration(expiresAt) * time.Second fmt.Println() fmt.Printf("过期分钟数: %d minute, 过期日期:%s \n", minute, time.Now().Add(duration).Format("2006-01-02 15:04:05")) fmt.Println() } else if strings.HasSuffix(inputString, "s") { inputString = inputString[:len(inputString)-1] seconds, err := strconv.ParseInt(inputString, 10, 64) if err != nil { return 0, err } if seconds <= 0 || seconds > 100*356*24*60*60 { return 0, fmt.Errorf("输入秒不能小于0,大于%d秒", 100*356*24*60*60) } expiresAt = seconds duration := time.Duration(expiresAt) * time.Second fmt.Println() fmt.Printf("过期秒数: %d seconds, 过期日期:%s \n", seconds, time.Now().Add(duration).Format("2006-01-02 15:04:05")) fmt.Println() } else { return 0, fmt.Errorf("%s", "输入不正确,请输入时间单位") } } // fmt.Println("\033[H\033[2J") return expiresAt, nil } func InputMachineID() (string, error) { fmt.Printf("%s\n", "请输入机器码:") input, err := inputReader.ReadString('\n') if err != nil { return "", err } defer inputReader.Reset(os.Stdin) inputString := strings.TrimSpace(input) if inputString != "" { if strings.HasPrefix(inpu
{ os.Exit(0) } fmt.Println() fmt.Printf("你输入的机器码是: %s\n", inputString) fmt.Println() } return inputString, nil } func ShowActiveCode(dir, fileName, uuid string) { fmt.Printf("序号:%s \n", uuid) fmt.Printf("\n%s\n", "激活码是:") readPath := filepath.Join(dir, fileName) licenseActive, err := public.ReadLicensePem(readPath) if err != nil { fmt.Println(err.Error()) return } fmt.Println(base64.URLEncoding.EncodeToString(licenseActive)) // fmt.Println(string(licenseActive)) } func ReadCustomKV(productName string) ([]AttrKV, error) { type Option struct { XMLName xml.Name `xml:"option"` Desc string `xml:"desc"` Key string `xml:"key"` Value string `xml:"val"` } type XMLProduct struct { XMLName xml.Name `xml:"options"` Version string `xml:"version,attr"` Options []Option `xml:"option"` } filePath := filepath.Join(DataDir, strings.Join([]string{productName, ".xml"}, "")) if public.Exists(filePath) { var ( attr = XMLProduct{} attrKV []AttrKV ) fd, err := os.OpenFile(filePath, os.O_RDONLY, 0644) if err != nil { return nil, err } defer fd.Close() attrKVBytes, err := ioutil.ReadAll(fd) if err != nil { return nil, err } err = xml.Unmarshal(attrKVBytes, &attr) if err != nil { return nil, err } // fmt.Printf("%s特性选择:\n", productExplan) for i := 0; i < len(attr.Options); i++ { // fmt.Println(i+1, attr.Options[i].Desc, attr.Options[i].Key, attr.Options[i].Value) attrKV = append(attrKV, AttrKV{Desc: attr.Options[i].Desc, Key: attr.Options[i].Key, Val: attr.Options[i].Value}) } // fmt.Println("请输入数字序号,以分号间隔:") return attrKV, nil } else { fd, err := os.OpenFile(filePath, os.O_RDWR|os.O_CREATE, 0644) if err != nil { return nil, err } defer fd.Close() attr := &XMLProduct{ Version: "1", } attr.Options = append(attr.Options, Option{ Desc: "示例配置", Key: "key1", Value: "val1", }) output, err := xml.MarshalIndent(attr, "", " ") if err != nil { fmt.Println(err) return nil, err } _, err = fd.Write([]byte(xml.Header)) if err != nil { return nil, err } _, err = fd.Write(output) if err != nil { return nil, err } } return nil, nil } func SelectCustomKV(productExplan string, kv []AttrKV) ([]AttrKV, error) { var ( arrayIdx []int kvs []AttrKV ) if kv != nil { fmt.Printf("%s启用配置选择(请输入数字序号,以逗号间隔,跳过按回车):\n", productExplan) for i := 0; i < len(kv); i++ { fmt.Println(i+1, kv[i].Desc) } input, err := inputReader.ReadString('\n') if err != nil { return nil, err } defer inputReader.Reset(os.Stdin) inputString := strings.TrimSpace(input) if inputString != "" { if strings.HasPrefix(inputString, "q") { os.Exit(0) } arrayIndx := strings.Split(inputString, ",") for i := 0; i < len(arrayIndx); i++ { num := strings.TrimSpace(arrayIndx[i]) if num == "" { continue } idx, err := strconv.Atoi(num) if err !=
tString, "q")
identifier_name
scope-auth.go
} } } else { // Access token: Get authorization from header or coockie. jwtStr, jwtExists = extractJWT(ctx, logger) } // Get scopes of endpoint. validScopes, err := scopeauth.GetScopes(service, strings.ToUpper(ctx.Request.Method), ctx.Request.URL.Path) if err != nil { logger.Info("can't find scopes of [%s] [%s] %s. err(%s)", service, ctx.Request.Method, ctx.Request.URL.Path, err) appCtx.SetError(apierrors.ResourceNotFound) return } if len(validScopes) == 0 { logger.Error("empty scopes of [%s] %s.%s.", ctx.Request.Method, service, ctx.Request.URL.Path) ctx.Abort() return } appCtx.RequiredScopes = validScopes logger.Debug("required scopes: %s", validScopes) validScopeMap := make(map[types.Scope]struct{}) hasPublicScope := false for _, s := range validScopes { if s == types.ScopePublic { hasPublicScope = true } validScopeMap[s] = struct{}{} } // No JWT. if !jwtExists { if hasPublicScope { ctx.Next() } else { logger.Error("authentication failed with no JWT found") appCtx.SetError(apierrors.AuthenticationError) } return } if !appCtx.IsAPIToken() && !appCtx.IsOAuth2Token() { skipIPCheck := false if len(opt) > 0 { skipIPCheck = opt[0].(bool) } // With JWT, validation userID, deviceAuthorizationID, accessTokenID, userScopes, devicePlatform, err := authenticateJWT(jwtStr, appCtx.RequestIP, skipIPCheck, appCtx.ServiceName) if err != nil { if hasPublicScope { ctx.Next() } else { logger.Error("authentication failed with invalid JWT. err: %+v", err) appCtx.SetError(apierrors.AuthenticationError) } return } logger.SetLabel(logging.LabelUserDeviceID, deviceAuthorizationID.String()) appCtx.Platform = devicePlatform appCtx.UserID = userID appCtx.AccessTokenID = accessTokenID appCtx.DeviceAuthorizationID = deviceAuthorizationID appCtx.UserAuthorizationScopes = userScopes } logger.SetLabel(logging.LabelUserID, appCtx.UserID.String()) logger.Debug("user scopes: %s", appCtx.UserAuthorizationScopes) if hasPublicScope { ctx.Next() return } // Validate scopes. for _, s := range appCtx.UserAuthorizationScopes { if _, ok := validScopeMap[s]; ok { // user scope is in valid scope map, proceed ctx.Next() return } } // User scope is not in valid, abort logger.Error("unauthorized scopes. user scopes [%s]. required scopes[%s]", appCtx.UserAuthorizationScopes, validScopes) appCtx.SetError(apierrors.UnauthorizedScope) } } func validateAPIToken(appCtx *apicontext.AppContext, store *jwtFactory.APIKeySecret, token string) (err error) { err = jwtFactory.ValidateCOBSecret(token, store) if err != nil { return } claimMap, err := jwtFactory.ParseJWTPayload(token) if err != nil { return } // check scope scopesFromClaim, sExist := claimMap["scope"].([]interface{}) if !sExist { err = fmt.Errorf("scope not exist") return } userScopes := make([]types.Scope, 0) for _, scope := range scopesFromClaim { str, ok := scope.(string) if ok { s := types.Scope(str) userScopes = append(userScopes, s) } } // check user id userIDStr := claimMap["user_id"].(string) userID, err := uuid.FromString(userIDStr) if err != nil { return } // check api token apiTokenIDStr, exist := claimMap["api_token_id"].(string) if !exist { // it should be panic if api_token_id is not exist in payload panic(fmt.Errorf("api_token_id not exist")) } apiTokenID, err := uuid.FromString(apiTokenIDStr) if err != nil { return } // create shared key apiTokenKey := keys.GetAPITokenKeyByUserStr(userIDStr) // check secret secret, err := appCtx.Cache.GetFieldOfMap(apiTokenKey, keys.APITokenSecretKey) if err != nil { errorCode := cache.ParseCacheErrorCode(err) if errorCode != cache.ErrNilKey { return } // lazy loading var client *kms.Client client, err = kms.NewDefaultClient(appCtx, kms.KeyAPIToken) if err != nil { return } // set secret apiSecret := models.APISecret{} err = appCtx.DB.Where("user_id = ?", userID).First(&apiSecret).Error if err != nil { return } var secretBytes []byte secretBytes, err = client.Decrypt(apiSecret.Secret) if err != nil { return } secret = string(secretBytes) err = appCtx.Cache.SetFieldOfMap(apiTokenKey, "secret", secret) if err != nil { return } } // valid token with secret partialToken := strings.Join(strings.Split(token, ".")[:3], ".") _, _, err = jwtFactory.BuildWithSecret(jwtFactory.APITokenObj{}, secret). Validate(partialToken, appCtx.ServiceName) if err != nil { return } // check cached data and lazy loading with db query _, err = appCtx.Cache.GetFieldOfMap(apiTokenKey, apiTokenIDStr) if err != nil { errorCode := cache.ParseCacheErrorCode(err) if errorCode != cache.ErrNilKey { return } lockKey := apiTokenKey + "_lock" // lock while updating tokens appCtx.Cache.Lock(lockKey, appCtx.RequestTag(), tokensWriteTimeout) defer appCtx.Cache.UnLock(lockKey, appCtx.RequestTag()) // set api tokens apiTokens := []models.APIToken{} err = appCtx.DB.Where("user_id = ? AND revoked_at IS NULL", userID).Find(&apiTokens).Error if err != nil { return } var isTokenExisting bool for _, at := range apiTokens { id := at.ID.String() err = appCtx.Cache.SetFieldOfMap(apiTokenKey, id, "") if err != nil { return } if !isTokenExisting { isTokenExisting = id == apiTokenIDStr } } if !isTokenExisting { err = fmt.Errorf("user<%s> api token<%s> error: %v", userIDStr, apiTokenIDStr, err) return } } appCtx.UserAuthorizationScopes = userScopes appCtx.UserID = &userID appCtx.APITokenID = &apiTokenID appCtx.Logger().SetLabel(logging.LabelAuthMethod, "api_token") return nil } func validateOAuth2Token(appCtx *apicontext.AppContext, token string) error { claims, _, err := jwtFactory.Build( jwtFactory.OAuth2AccessTokenObj{}).Validate(token, appCtx.ServiceName) if err != nil { return err } oauth2AccessTokenIDStr, found := claims["oauth2_access_token_id"].(string) if !found { return fmt.Errorf("oauth2_access_token_id could not be found") } oauth2AccessTokenID, err := uuid.FromString(oauth2AccessTokenIDStr) if err != nil { return err } clientIDStr, found := claims["client_id"].(string) if !found { return fmt.Errorf("client_id could not be found") } clientID, err := uuid.FromString(clientIDStr) if err != nil { return err } userIDStr, found := claims["user_id"].(string) if !found { return fmt.Errorf("user_id could not be found") } userID, err := uuid.FromString(userIDStr) if err != nil { return err } scopes, found := claims["scope"].([]interface{}) if !found { return fmt.Errorf("scope could not be found") } oauth2TokenKey := "oauth2_access_token:" + oauth2AccessTokenIDStr found, err = appCtx.Cache.Exist(oauth2TokenKey) if err != nil { return err } if !found { oauth2Token := models.OAuth2Token{} result := appCtx.DB.Where("id = ? AND type = ? AND revoked_at IS NULL", oauth2AccessTokenID, types.OAuth2AccessToken).First(&oauth2Token) if result.Error != nil
{ return result.Error }
conditional_block
scope-auth.go
s.", ctx.Request.Method, service, ctx.Request.URL.Path) ctx.Abort() return } appCtx.RequiredScopes = validScopes logger.Debug("required scopes: %s", validScopes) validScopeMap := make(map[types.Scope]struct{}) hasPublicScope := false for _, s := range validScopes { if s == types.ScopePublic { hasPublicScope = true } validScopeMap[s] = struct{}{} } // No JWT. if !jwtExists { if hasPublicScope { ctx.Next() } else { logger.Error("authentication failed with no JWT found") appCtx.SetError(apierrors.AuthenticationError) } return } if !appCtx.IsAPIToken() && !appCtx.IsOAuth2Token() { skipIPCheck := false if len(opt) > 0 { skipIPCheck = opt[0].(bool) } // With JWT, validation userID, deviceAuthorizationID, accessTokenID, userScopes, devicePlatform, err := authenticateJWT(jwtStr, appCtx.RequestIP, skipIPCheck, appCtx.ServiceName) if err != nil { if hasPublicScope { ctx.Next() } else { logger.Error("authentication failed with invalid JWT. err: %+v", err) appCtx.SetError(apierrors.AuthenticationError) } return } logger.SetLabel(logging.LabelUserDeviceID, deviceAuthorizationID.String()) appCtx.Platform = devicePlatform appCtx.UserID = userID appCtx.AccessTokenID = accessTokenID appCtx.DeviceAuthorizationID = deviceAuthorizationID appCtx.UserAuthorizationScopes = userScopes } logger.SetLabel(logging.LabelUserID, appCtx.UserID.String()) logger.Debug("user scopes: %s", appCtx.UserAuthorizationScopes) if hasPublicScope { ctx.Next() return } // Validate scopes. for _, s := range appCtx.UserAuthorizationScopes { if _, ok := validScopeMap[s]; ok { // user scope is in valid scope map, proceed ctx.Next() return } } // User scope is not in valid, abort logger.Error("unauthorized scopes. user scopes [%s]. required scopes[%s]", appCtx.UserAuthorizationScopes, validScopes) appCtx.SetError(apierrors.UnauthorizedScope) } } func validateAPIToken(appCtx *apicontext.AppContext, store *jwtFactory.APIKeySecret, token string) (err error) { err = jwtFactory.ValidateCOBSecret(token, store) if err != nil { return } claimMap, err := jwtFactory.ParseJWTPayload(token) if err != nil { return } // check scope scopesFromClaim, sExist := claimMap["scope"].([]interface{}) if !sExist { err = fmt.Errorf("scope not exist") return } userScopes := make([]types.Scope, 0) for _, scope := range scopesFromClaim { str, ok := scope.(string) if ok { s := types.Scope(str) userScopes = append(userScopes, s) } } // check user id userIDStr := claimMap["user_id"].(string) userID, err := uuid.FromString(userIDStr) if err != nil { return } // check api token apiTokenIDStr, exist := claimMap["api_token_id"].(string) if !exist { // it should be panic if api_token_id is not exist in payload panic(fmt.Errorf("api_token_id not exist")) } apiTokenID, err := uuid.FromString(apiTokenIDStr) if err != nil { return } // create shared key apiTokenKey := keys.GetAPITokenKeyByUserStr(userIDStr) // check secret secret, err := appCtx.Cache.GetFieldOfMap(apiTokenKey, keys.APITokenSecretKey) if err != nil { errorCode := cache.ParseCacheErrorCode(err) if errorCode != cache.ErrNilKey { return } // lazy loading var client *kms.Client client, err = kms.NewDefaultClient(appCtx, kms.KeyAPIToken) if err != nil { return } // set secret apiSecret := models.APISecret{} err = appCtx.DB.Where("user_id = ?", userID).First(&apiSecret).Error if err != nil { return } var secretBytes []byte secretBytes, err = client.Decrypt(apiSecret.Secret) if err != nil { return } secret = string(secretBytes) err = appCtx.Cache.SetFieldOfMap(apiTokenKey, "secret", secret) if err != nil { return } } // valid token with secret partialToken := strings.Join(strings.Split(token, ".")[:3], ".") _, _, err = jwtFactory.BuildWithSecret(jwtFactory.APITokenObj{}, secret). Validate(partialToken, appCtx.ServiceName) if err != nil { return } // check cached data and lazy loading with db query _, err = appCtx.Cache.GetFieldOfMap(apiTokenKey, apiTokenIDStr) if err != nil { errorCode := cache.ParseCacheErrorCode(err) if errorCode != cache.ErrNilKey { return } lockKey := apiTokenKey + "_lock" // lock while updating tokens appCtx.Cache.Lock(lockKey, appCtx.RequestTag(), tokensWriteTimeout) defer appCtx.Cache.UnLock(lockKey, appCtx.RequestTag()) // set api tokens apiTokens := []models.APIToken{} err = appCtx.DB.Where("user_id = ? AND revoked_at IS NULL", userID).Find(&apiTokens).Error if err != nil { return } var isTokenExisting bool for _, at := range apiTokens { id := at.ID.String() err = appCtx.Cache.SetFieldOfMap(apiTokenKey, id, "") if err != nil { return } if !isTokenExisting { isTokenExisting = id == apiTokenIDStr } } if !isTokenExisting { err = fmt.Errorf("user<%s> api token<%s> error: %v", userIDStr, apiTokenIDStr, err) return } } appCtx.UserAuthorizationScopes = userScopes appCtx.UserID = &userID appCtx.APITokenID = &apiTokenID appCtx.Logger().SetLabel(logging.LabelAuthMethod, "api_token") return nil } func validateOAuth2Token(appCtx *apicontext.AppContext, token string) error { claims, _, err := jwtFactory.Build( jwtFactory.OAuth2AccessTokenObj{}).Validate(token, appCtx.ServiceName) if err != nil { return err } oauth2AccessTokenIDStr, found := claims["oauth2_access_token_id"].(string) if !found { return fmt.Errorf("oauth2_access_token_id could not be found") } oauth2AccessTokenID, err := uuid.FromString(oauth2AccessTokenIDStr) if err != nil { return err } clientIDStr, found := claims["client_id"].(string) if !found { return fmt.Errorf("client_id could not be found") } clientID, err := uuid.FromString(clientIDStr) if err != nil { return err } userIDStr, found := claims["user_id"].(string) if !found { return fmt.Errorf("user_id could not be found") } userID, err := uuid.FromString(userIDStr) if err != nil { return err } scopes, found := claims["scope"].([]interface{}) if !found { return fmt.Errorf("scope could not be found") } oauth2TokenKey := "oauth2_access_token:" + oauth2AccessTokenIDStr found, err = appCtx.Cache.Exist(oauth2TokenKey) if err != nil { return err } if !found { oauth2Token := models.OAuth2Token{} result := appCtx.DB.Where("id = ? AND type = ? AND revoked_at IS NULL", oauth2AccessTokenID, types.OAuth2AccessToken).First(&oauth2Token) if result.Error != nil { return result.Error } expireSec := int(oauth2Token.ExpireAt.Sub(time.Now()) / time.Second) appCtx.Cache.Set(oauth2TokenKey, 1, expireSec) } userScopes := make([]types.Scope, 0) for _, scope := range scopes { scopeStr, ok := scope.(string) if ok { userScopes = append(userScopes, types.Scope(scopeStr)) } } appCtx.UserAuthorizationScopes = userScopes appCtx.UserID = &userID appCtx.OAuth2TokenID = &oauth2AccessTokenID appCtx.OAuth2ClientID = clientID appCtx.Logger().SetLabel(logging.LabelAuthMethod, "oauth2_access_token") return nil } func
extractJWT
identifier_name
scope-auth.go
bool if len(strings.Split(ctx.GetHeader("Authorization"), ".")) == 4 { // API token vErr := validateAPIToken(appCtx, store, ctx.GetHeader("Authorization")) if vErr != nil { logger.Error("api token is invalid: %v", vErr) } else { jwtExists = true } } else if len(strings.Split(ctx.GetHeader("Authorization"), " ")) == 2 { // OAuth2 token authParts := strings.SplitN(ctx.GetHeader("Authorization"), " ", 2) authType, authCredentials := authParts[0], authParts[1] if authType == OAuth2AccessTokenType && len(authCredentials) > 0 { vErr := validateOAuth2Token(appCtx, authCredentials) if vErr != nil { logger.Error("OAuth2 access token is invalid: %v", vErr) } else { jwtExists = true } } } else { // Access token: Get authorization from header or coockie. jwtStr, jwtExists = extractJWT(ctx, logger) } // Get scopes of endpoint. validScopes, err := scopeauth.GetScopes(service, strings.ToUpper(ctx.Request.Method), ctx.Request.URL.Path) if err != nil { logger.Info("can't find scopes of [%s] [%s] %s. err(%s)", service, ctx.Request.Method, ctx.Request.URL.Path, err) appCtx.SetError(apierrors.ResourceNotFound) return } if len(validScopes) == 0 { logger.Error("empty scopes of [%s] %s.%s.", ctx.Request.Method, service, ctx.Request.URL.Path) ctx.Abort() return } appCtx.RequiredScopes = validScopes logger.Debug("required scopes: %s", validScopes) validScopeMap := make(map[types.Scope]struct{}) hasPublicScope := false for _, s := range validScopes { if s == types.ScopePublic { hasPublicScope = true } validScopeMap[s] = struct{}{} } // No JWT. if !jwtExists { if hasPublicScope { ctx.Next() } else { logger.Error("authentication failed with no JWT found") appCtx.SetError(apierrors.AuthenticationError) } return } if !appCtx.IsAPIToken() && !appCtx.IsOAuth2Token() { skipIPCheck := false if len(opt) > 0 { skipIPCheck = opt[0].(bool) } // With JWT, validation userID, deviceAuthorizationID, accessTokenID, userScopes, devicePlatform, err := authenticateJWT(jwtStr, appCtx.RequestIP, skipIPCheck, appCtx.ServiceName) if err != nil { if hasPublicScope {
logger.Error("authentication failed with invalid JWT. err: %+v", err) appCtx.SetError(apierrors.AuthenticationError) } return } logger.SetLabel(logging.LabelUserDeviceID, deviceAuthorizationID.String()) appCtx.Platform = devicePlatform appCtx.UserID = userID appCtx.AccessTokenID = accessTokenID appCtx.DeviceAuthorizationID = deviceAuthorizationID appCtx.UserAuthorizationScopes = userScopes } logger.SetLabel(logging.LabelUserID, appCtx.UserID.String()) logger.Debug("user scopes: %s", appCtx.UserAuthorizationScopes) if hasPublicScope { ctx.Next() return } // Validate scopes. for _, s := range appCtx.UserAuthorizationScopes { if _, ok := validScopeMap[s]; ok { // user scope is in valid scope map, proceed ctx.Next() return } } // User scope is not in valid, abort logger.Error("unauthorized scopes. user scopes [%s]. required scopes[%s]", appCtx.UserAuthorizationScopes, validScopes) appCtx.SetError(apierrors.UnauthorizedScope) } } func validateAPIToken(appCtx *apicontext.AppContext, store *jwtFactory.APIKeySecret, token string) (err error) { err = jwtFactory.ValidateCOBSecret(token, store) if err != nil { return } claimMap, err := jwtFactory.ParseJWTPayload(token) if err != nil { return } // check scope scopesFromClaim, sExist := claimMap["scope"].([]interface{}) if !sExist { err = fmt.Errorf("scope not exist") return } userScopes := make([]types.Scope, 0) for _, scope := range scopesFromClaim { str, ok := scope.(string) if ok { s := types.Scope(str) userScopes = append(userScopes, s) } } // check user id userIDStr := claimMap["user_id"].(string) userID, err := uuid.FromString(userIDStr) if err != nil { return } // check api token apiTokenIDStr, exist := claimMap["api_token_id"].(string) if !exist { // it should be panic if api_token_id is not exist in payload panic(fmt.Errorf("api_token_id not exist")) } apiTokenID, err := uuid.FromString(apiTokenIDStr) if err != nil { return } // create shared key apiTokenKey := keys.GetAPITokenKeyByUserStr(userIDStr) // check secret secret, err := appCtx.Cache.GetFieldOfMap(apiTokenKey, keys.APITokenSecretKey) if err != nil { errorCode := cache.ParseCacheErrorCode(err) if errorCode != cache.ErrNilKey { return } // lazy loading var client *kms.Client client, err = kms.NewDefaultClient(appCtx, kms.KeyAPIToken) if err != nil { return } // set secret apiSecret := models.APISecret{} err = appCtx.DB.Where("user_id = ?", userID).First(&apiSecret).Error if err != nil { return } var secretBytes []byte secretBytes, err = client.Decrypt(apiSecret.Secret) if err != nil { return } secret = string(secretBytes) err = appCtx.Cache.SetFieldOfMap(apiTokenKey, "secret", secret) if err != nil { return } } // valid token with secret partialToken := strings.Join(strings.Split(token, ".")[:3], ".") _, _, err = jwtFactory.BuildWithSecret(jwtFactory.APITokenObj{}, secret). Validate(partialToken, appCtx.ServiceName) if err != nil { return } // check cached data and lazy loading with db query _, err = appCtx.Cache.GetFieldOfMap(apiTokenKey, apiTokenIDStr) if err != nil { errorCode := cache.ParseCacheErrorCode(err) if errorCode != cache.ErrNilKey { return } lockKey := apiTokenKey + "_lock" // lock while updating tokens appCtx.Cache.Lock(lockKey, appCtx.RequestTag(), tokensWriteTimeout) defer appCtx.Cache.UnLock(lockKey, appCtx.RequestTag()) // set api tokens apiTokens := []models.APIToken{} err = appCtx.DB.Where("user_id = ? AND revoked_at IS NULL", userID).Find(&apiTokens).Error if err != nil { return } var isTokenExisting bool for _, at := range apiTokens { id := at.ID.String() err = appCtx.Cache.SetFieldOfMap(apiTokenKey, id, "") if err != nil { return } if !isTokenExisting { isTokenExisting = id == apiTokenIDStr } } if !isTokenExisting { err = fmt.Errorf("user<%s> api token<%s> error: %v", userIDStr, apiTokenIDStr, err) return } } appCtx.UserAuthorizationScopes = userScopes appCtx.UserID = &userID appCtx.APITokenID = &apiTokenID appCtx.Logger().SetLabel(logging.LabelAuthMethod, "api_token") return nil } func validateOAuth2Token(appCtx *apicontext.AppContext, token string) error { claims, _, err := jwtFactory.Build( jwtFactory.OAuth2AccessTokenObj{}).Validate(token, appCtx.ServiceName) if err != nil { return err } oauth2AccessTokenIDStr, found := claims["oauth2_access_token_id"].(string) if !found { return fmt.Errorf("oauth2_access_token_id could not be found") } oauth2AccessTokenID, err := uuid.FromString(oauth2AccessTokenIDStr) if err != nil { return err } clientIDStr, found := claims["client_id"].(string) if !found { return fmt.Errorf("client_id could not be found") } clientID, err := uuid.FromString(clientID
ctx.Next() } else {
random_line_split
scope-auth.go
appCtx.UserID = userID appCtx.AccessTokenID = accessTokenID appCtx.DeviceAuthorizationID = deviceAuthorizationID appCtx.UserAuthorizationScopes = userScopes } logger.SetLabel(logging.LabelUserID, appCtx.UserID.String()) logger.Debug("user scopes: %s", appCtx.UserAuthorizationScopes) if hasPublicScope { ctx.Next() return } // Validate scopes. for _, s := range appCtx.UserAuthorizationScopes { if _, ok := validScopeMap[s]; ok { // user scope is in valid scope map, proceed ctx.Next() return } } // User scope is not in valid, abort logger.Error("unauthorized scopes. user scopes [%s]. required scopes[%s]", appCtx.UserAuthorizationScopes, validScopes) appCtx.SetError(apierrors.UnauthorizedScope) } } func validateAPIToken(appCtx *apicontext.AppContext, store *jwtFactory.APIKeySecret, token string) (err error) { err = jwtFactory.ValidateCOBSecret(token, store) if err != nil { return } claimMap, err := jwtFactory.ParseJWTPayload(token) if err != nil { return } // check scope scopesFromClaim, sExist := claimMap["scope"].([]interface{}) if !sExist { err = fmt.Errorf("scope not exist") return } userScopes := make([]types.Scope, 0) for _, scope := range scopesFromClaim { str, ok := scope.(string) if ok { s := types.Scope(str) userScopes = append(userScopes, s) } } // check user id userIDStr := claimMap["user_id"].(string) userID, err := uuid.FromString(userIDStr) if err != nil { return } // check api token apiTokenIDStr, exist := claimMap["api_token_id"].(string) if !exist { // it should be panic if api_token_id is not exist in payload panic(fmt.Errorf("api_token_id not exist")) } apiTokenID, err := uuid.FromString(apiTokenIDStr) if err != nil { return } // create shared key apiTokenKey := keys.GetAPITokenKeyByUserStr(userIDStr) // check secret secret, err := appCtx.Cache.GetFieldOfMap(apiTokenKey, keys.APITokenSecretKey) if err != nil { errorCode := cache.ParseCacheErrorCode(err) if errorCode != cache.ErrNilKey { return } // lazy loading var client *kms.Client client, err = kms.NewDefaultClient(appCtx, kms.KeyAPIToken) if err != nil { return } // set secret apiSecret := models.APISecret{} err = appCtx.DB.Where("user_id = ?", userID).First(&apiSecret).Error if err != nil { return } var secretBytes []byte secretBytes, err = client.Decrypt(apiSecret.Secret) if err != nil { return } secret = string(secretBytes) err = appCtx.Cache.SetFieldOfMap(apiTokenKey, "secret", secret) if err != nil { return } } // valid token with secret partialToken := strings.Join(strings.Split(token, ".")[:3], ".") _, _, err = jwtFactory.BuildWithSecret(jwtFactory.APITokenObj{}, secret). Validate(partialToken, appCtx.ServiceName) if err != nil { return } // check cached data and lazy loading with db query _, err = appCtx.Cache.GetFieldOfMap(apiTokenKey, apiTokenIDStr) if err != nil { errorCode := cache.ParseCacheErrorCode(err) if errorCode != cache.ErrNilKey { return } lockKey := apiTokenKey + "_lock" // lock while updating tokens appCtx.Cache.Lock(lockKey, appCtx.RequestTag(), tokensWriteTimeout) defer appCtx.Cache.UnLock(lockKey, appCtx.RequestTag()) // set api tokens apiTokens := []models.APIToken{} err = appCtx.DB.Where("user_id = ? AND revoked_at IS NULL", userID).Find(&apiTokens).Error if err != nil { return } var isTokenExisting bool for _, at := range apiTokens { id := at.ID.String() err = appCtx.Cache.SetFieldOfMap(apiTokenKey, id, "") if err != nil { return } if !isTokenExisting { isTokenExisting = id == apiTokenIDStr } } if !isTokenExisting { err = fmt.Errorf("user<%s> api token<%s> error: %v", userIDStr, apiTokenIDStr, err) return } } appCtx.UserAuthorizationScopes = userScopes appCtx.UserID = &userID appCtx.APITokenID = &apiTokenID appCtx.Logger().SetLabel(logging.LabelAuthMethod, "api_token") return nil } func validateOAuth2Token(appCtx *apicontext.AppContext, token string) error { claims, _, err := jwtFactory.Build( jwtFactory.OAuth2AccessTokenObj{}).Validate(token, appCtx.ServiceName) if err != nil { return err } oauth2AccessTokenIDStr, found := claims["oauth2_access_token_id"].(string) if !found { return fmt.Errorf("oauth2_access_token_id could not be found") } oauth2AccessTokenID, err := uuid.FromString(oauth2AccessTokenIDStr) if err != nil { return err } clientIDStr, found := claims["client_id"].(string) if !found { return fmt.Errorf("client_id could not be found") } clientID, err := uuid.FromString(clientIDStr) if err != nil { return err } userIDStr, found := claims["user_id"].(string) if !found { return fmt.Errorf("user_id could not be found") } userID, err := uuid.FromString(userIDStr) if err != nil { return err } scopes, found := claims["scope"].([]interface{}) if !found { return fmt.Errorf("scope could not be found") } oauth2TokenKey := "oauth2_access_token:" + oauth2AccessTokenIDStr found, err = appCtx.Cache.Exist(oauth2TokenKey) if err != nil { return err } if !found { oauth2Token := models.OAuth2Token{} result := appCtx.DB.Where("id = ? AND type = ? AND revoked_at IS NULL", oauth2AccessTokenID, types.OAuth2AccessToken).First(&oauth2Token) if result.Error != nil { return result.Error } expireSec := int(oauth2Token.ExpireAt.Sub(time.Now()) / time.Second) appCtx.Cache.Set(oauth2TokenKey, 1, expireSec) } userScopes := make([]types.Scope, 0) for _, scope := range scopes { scopeStr, ok := scope.(string) if ok { userScopes = append(userScopes, types.Scope(scopeStr)) } } appCtx.UserAuthorizationScopes = userScopes appCtx.UserID = &userID appCtx.OAuth2TokenID = &oauth2AccessTokenID appCtx.OAuth2ClientID = clientID appCtx.Logger().SetLabel(logging.LabelAuthMethod, "oauth2_access_token") return nil } func extractJWT(ctx *gin.Context, logger logging.Logger) (jwtString string, exists bool) { var jwtStr string jwtStr = ctx.GetHeader("Authorization") if len(jwtStr) != 0 { jwtString = jwtStr exists = true return } var err error jwtStr, err = ctx.Cookie("Authorization") if err == nil && len(jwtStr) != 0 { jwtString = jwtStr exists = true return } jwtString = "" exists = false return } func authenticateJWT(jwtStr string, requestIP string, skipIPCheck bool, serviceName cobxtypes.ServiceName) (userID *uuid.UUID, deviceAuthorizationID *uuid.UUID, accessTokenID *uuid.UUID, userScopes []types.Scope, devicePlatform types.DevicePlatform, err error)
{ claims, isExpired, delErr := jwtFactory.Build(jwtFactory.AccessTokenObj{}). Validate(jwtStr, serviceName) if delErr != nil || isExpired { userID = nil deviceAuthorizationID = nil accessTokenID = nil userScopes = nil if isExpired { err = fmt.Errorf("token <%v> expired", accessTokenID) } else { err = delErr } return } userIDFromClaim, uExist := claims["user_id"].(string) accessTokenIDFromClaim, aExist := claims["access_token_id"].(string) devAuthIDFromClaim, dExist := claims["device_authorization_id"].(string) platformFromClaim, pExist := claims["platform"].(string)
identifier_body
ts_analyzer.py
str(last_section_number)+"</last_section_number>") for i in range(0,(section_length-5-4)/4): # print(" <program>") cursor+=4 program_num = (ord(f[9+i:10+i]) << 8) | ord(f[10+i:11+i]) b1 = ord(f[11+i:12+i]) b2 = ord(f[12+i:13+i]) if b1 & 0xE0 != 0xE0: print(" <!-- reserved is not 111 -->") program_pid = b2 | ((b1 & 0x1F) << 8) # print(" <program_num>"+str(program_num)+"</program_num>") # print(" <program_pid>"+hex(program_pid)+"</program_pid>") # print(" </program>\n") #program_map_pids.add(program_pid) crc32 = f[cursor:cursor+4]; cursor+=4 length -= cursor if length>0: rest = f[cursor:cursor+length] if (rest != '\xff' * length) and (rest != '\x00' * length): print(" <rest>"+binascii.hexlify(rest)+"</rest>\n") # print(" </program_association_table>\n") return({'table_id':str(pointer_field),'transportstream_id':str(transport_stream_ID),'program':str(program_num),'pmt':hex(program_pid) }) def output_adaptation_field(f): print(" <adaptation_field>\n") additional_length = ord(f.read(1)) if additional_length == 0: print(" </adaptation_field>\n") return 1 flags = ord(f.read(1)) discontinuity = bool(flags & 0x80) random_access = bool(flags & 0x40) elementary_stream_priority = bool(flags & 0x20) pcr = bool(flags & 0x10) opcr = bool(flags & 0x08) splicing_point = bool(flags & 0x04) transport_private = bool(flags & 0x02) adaptation_field_extension = bool(flags & 0x01) if discontinuity: print(" <discontinuity/>\n") if random_access: print(" <random_access/>\n") if elementary_stream_priority:
length = additional_length+1 # size byte additional_length-=1 # flags def read_pcr(): pcr_byte_1 = ord(f.read(1)) # base pcr_byte_2 = ord(f.read(1)) # base pcr_byte_3 = ord(f.read(1)) # base pcr_byte_4 = ord(f.read(1)) # base pcr_byte_5 = ord(f.read(1)) # 1 bit base, 6 bits paddding, 1 bit ext pcr_byte_6 = ord(f.read(1)) # 8 bits ext base = (pcr_byte_1 << (1+8*3)) + \ (pcr_byte_2 << (1+8*2)) + \ (pcr_byte_3 << (1+8*1)) + \ (pcr_byte_4 << (1+8*0)) + \ (pcr_byte_5 >> 7) ext = ((pcr_byte_5 & 0x01) << 8) + pcr_byte_6 time = base / 90000.0 + ext / 27000000.0 return time if pcr: if additional_length>=6: additional_length-=6 val = read_pcr() print(" <program_clock_reference>"+str(val)+"</program_clock_reference>\n") if opcr: if additional_length>=6: additional_length-=6 val = read_pcr() print(" <original_program_clock_reference>"+str(val)+"</original_program_clock_reference>\n") if splicing_point: if additional_length>=1: additional_length-=1 splice_count = ord(f.read(1)) print(" <splice_countdown>"+str(splice_count)+"</splice_countdown>\n") if additional_length: print(" <!-- ignoring " + str(additional_length) + " bytes -->\n") f.read(additional_length) print(" </adaptation_field>\n") return length def on_read(handle, ip_port, flags, data, error): global bits_second, start_time_packet if error is not None: print (error,color='red') return if start_time_packet == 'unset': start_time_packet=datetime.datetime.now() data = data.strip() mcast=handle.getsockname() if data: ip, port = ip_port diff = datetime.datetime.now()-start_time_packet if diff.total_seconds() >= '30': bits_second=1 start_time_packet=datetime.datetime.now() else: bits_second=1+bits_second for i in range(0,len(data),188): offset =+ i #print(offset) sync = ord(data[offset:offset+1]) header1 = ord(data[offset+1:offset+2]) header2 = ord(data[offset+2:offset+3]) header3 = ord(data[offset+3:offset+4]) transport_error = bool(header1 & 0x80) payload_unit_start = bool(header1 & 0x40) transport_priority = bool(header1 & 0x20) pid = header2 | ((header1 & 0x1F) << 8) scrambling = ((header3 & 0xC0) >> 6) have_adaptation_field = bool(header3 & 0x20) adaptation_field = ((header3 & 0x30) >> 4) have_payload = bool(header3 & 0x10) cc = header3 & 0x0F length = len(data) # We have sync: if sync == 0x47: if mcast not in pids: pids[mcast]={} if pid not in pids[mcast]: pids[mcast][pid]={'packets': 1, 'cc': cc, 'error': 0, 'ip': ip, 'type': 'unknown', 'extra': {}} print ("===> Found new PID in stream %s (src=%s)" % (mcast,ip),end='') if pid == 0: pids[mcast][pid]['type'] = "PAT" print (" [PAT] ",end='') buf.append("%s NEW PID: %s (src=%s)(PID: %s [%s]) [PAT]" % (datetime.datetime.now(),mcast,ip,hex(pid),pid)) buf.append("%s NEW PID: %s (src=%s) (PID: %s [%s])" % (datetime.datetime.now(),mcast,ip,hex(pid),pid)) print ("(PID: ",end='') print ("%s"% hex(pid), color='green',end='') print (" [%s])"% pid) else: pids[mcast][pid]['packets']= pids[mcast][pid]['packets']+1 if adaptation_field != 2: cc_com = (pids[mcast][pid]['cc']+1) % 16 pids[mcast][pid]['cc'] = cc if cc is not cc_com: pids[mcast][pid]['error'] = pids[mcast][pid]['error']+1 print ("%s Error expected %s got %s (%s) %s %s" % (datetime.datetime.now(), cc_com, cc, mcast, hex(pid), length), color='red') syslog.syslog(syslog.LOG_ERR, "%s Error expected %s got %s (%s) %s %s" % (datetime.datetime.now(), cc_com, cc, mcast, hex(pid), length)) buf.append( "%s ERR: expected %s got %s (%s) %s %s" % (datetime.datetime.now(), cc_com, cc, mcast, hex(pid), length)) if pid == 0x00: #adaptation_field_size = 0 #print(have_adaptation_field) #if have_adaptation_field: adaptation_field_size = 167 payload_size = 188 - 4 - adaptation_field_size #print(payload_unit_start); #print(mcast) pids[mcast][pid]['extra']=output_program_association_table(data[offset+4:offset+188], payload_size, payload_unit_start) #pp = pprint.PrettyPrinter(indent=4) #pp.pprint(pids) class MainHandler(tornado.web.RequestHandler): def get(self): from platform import uname hostname = uname()[1] run_time = datetime.datetime.now() - start_time packet_time = datetime.datetime.now() - start_time_packet
print(" <elementary_stream_priority/>\n")
conditional_block
ts_analyzer.py
(program_num)+"</program_num>") # print(" <program_pid>"+hex(program_pid)+"</program_pid>") # print(" </program>\n") #program_map_pids.add(program_pid) crc32 = f[cursor:cursor+4]; cursor+=4 length -= cursor if length>0: rest = f[cursor:cursor+length] if (rest != '\xff' * length) and (rest != '\x00' * length): print(" <rest>"+binascii.hexlify(rest)+"</rest>\n") # print(" </program_association_table>\n") return({'table_id':str(pointer_field),'transportstream_id':str(transport_stream_ID),'program':str(program_num),'pmt':hex(program_pid) }) def output_adaptation_field(f): print(" <adaptation_field>\n") additional_length = ord(f.read(1)) if additional_length == 0: print(" </adaptation_field>\n") return 1 flags = ord(f.read(1)) discontinuity = bool(flags & 0x80) random_access = bool(flags & 0x40) elementary_stream_priority = bool(flags & 0x20) pcr = bool(flags & 0x10) opcr = bool(flags & 0x08) splicing_point = bool(flags & 0x04) transport_private = bool(flags & 0x02) adaptation_field_extension = bool(flags & 0x01) if discontinuity: print(" <discontinuity/>\n") if random_access: print(" <random_access/>\n") if elementary_stream_priority: print(" <elementary_stream_priority/>\n") length = additional_length+1 # size byte additional_length-=1 # flags def read_pcr(): pcr_byte_1 = ord(f.read(1)) # base pcr_byte_2 = ord(f.read(1)) # base pcr_byte_3 = ord(f.read(1)) # base pcr_byte_4 = ord(f.read(1)) # base pcr_byte_5 = ord(f.read(1)) # 1 bit base, 6 bits paddding, 1 bit ext pcr_byte_6 = ord(f.read(1)) # 8 bits ext base = (pcr_byte_1 << (1+8*3)) + \ (pcr_byte_2 << (1+8*2)) + \ (pcr_byte_3 << (1+8*1)) + \ (pcr_byte_4 << (1+8*0)) + \ (pcr_byte_5 >> 7) ext = ((pcr_byte_5 & 0x01) << 8) + pcr_byte_6 time = base / 90000.0 + ext / 27000000.0 return time if pcr: if additional_length>=6: additional_length-=6 val = read_pcr() print(" <program_clock_reference>"+str(val)+"</program_clock_reference>\n") if opcr: if additional_length>=6: additional_length-=6 val = read_pcr() print(" <original_program_clock_reference>"+str(val)+"</original_program_clock_reference>\n") if splicing_point: if additional_length>=1: additional_length-=1 splice_count = ord(f.read(1)) print(" <splice_countdown>"+str(splice_count)+"</splice_countdown>\n") if additional_length: print(" <!-- ignoring " + str(additional_length) + " bytes -->\n") f.read(additional_length) print(" </adaptation_field>\n") return length def on_read(handle, ip_port, flags, data, error): global bits_second, start_time_packet if error is not None: print (error,color='red') return if start_time_packet == 'unset': start_time_packet=datetime.datetime.now() data = data.strip() mcast=handle.getsockname() if data: ip, port = ip_port diff = datetime.datetime.now()-start_time_packet if diff.total_seconds() >= '30': bits_second=1 start_time_packet=datetime.datetime.now() else: bits_second=1+bits_second for i in range(0,len(data),188): offset =+ i #print(offset) sync = ord(data[offset:offset+1]) header1 = ord(data[offset+1:offset+2]) header2 = ord(data[offset+2:offset+3]) header3 = ord(data[offset+3:offset+4]) transport_error = bool(header1 & 0x80) payload_unit_start = bool(header1 & 0x40) transport_priority = bool(header1 & 0x20) pid = header2 | ((header1 & 0x1F) << 8) scrambling = ((header3 & 0xC0) >> 6) have_adaptation_field = bool(header3 & 0x20) adaptation_field = ((header3 & 0x30) >> 4) have_payload = bool(header3 & 0x10) cc = header3 & 0x0F length = len(data) # We have sync: if sync == 0x47: if mcast not in pids: pids[mcast]={} if pid not in pids[mcast]: pids[mcast][pid]={'packets': 1, 'cc': cc, 'error': 0, 'ip': ip, 'type': 'unknown', 'extra': {}} print ("===> Found new PID in stream %s (src=%s)" % (mcast,ip),end='') if pid == 0: pids[mcast][pid]['type'] = "PAT" print (" [PAT] ",end='') buf.append("%s NEW PID: %s (src=%s)(PID: %s [%s]) [PAT]" % (datetime.datetime.now(),mcast,ip,hex(pid),pid)) buf.append("%s NEW PID: %s (src=%s) (PID: %s [%s])" % (datetime.datetime.now(),mcast,ip,hex(pid),pid)) print ("(PID: ",end='') print ("%s"% hex(pid), color='green',end='') print (" [%s])"% pid) else: pids[mcast][pid]['packets']= pids[mcast][pid]['packets']+1 if adaptation_field != 2: cc_com = (pids[mcast][pid]['cc']+1) % 16 pids[mcast][pid]['cc'] = cc if cc is not cc_com: pids[mcast][pid]['error'] = pids[mcast][pid]['error']+1 print ("%s Error expected %s got %s (%s) %s %s" % (datetime.datetime.now(), cc_com, cc, mcast, hex(pid), length), color='red') syslog.syslog(syslog.LOG_ERR, "%s Error expected %s got %s (%s) %s %s" % (datetime.datetime.now(), cc_com, cc, mcast, hex(pid), length)) buf.append( "%s ERR: expected %s got %s (%s) %s %s" % (datetime.datetime.now(), cc_com, cc, mcast, hex(pid), length)) if pid == 0x00: #adaptation_field_size = 0 #print(have_adaptation_field) #if have_adaptation_field: adaptation_field_size = 167 payload_size = 188 - 4 - adaptation_field_size #print(payload_unit_start); #print(mcast) pids[mcast][pid]['extra']=output_program_association_table(data[offset+4:offset+188], payload_size, payload_unit_start) #pp = pprint.PrettyPrinter(indent=4) #pp.pprint(pids) class MainHandler(tornado.web.RequestHandler): def get(self): from platform import uname hostname = uname()[1] run_time = datetime.datetime.now() - start_time packet_time = datetime.datetime.now() - start_time_packet bits=((bits_second*1316)*8/packet_time.total_seconds())/1000000 self.render('index.html',version=ts_analyzer.__version__,addresses=dict(addresses),hostname=hostname, location=location,bits=round(bits,2), run_time=run_time,buf=buf,peers=peers) class LogHandler(tornado.web.RequestHandler): def get(self): from platform import uname hostname = uname()[1] self.render('log.html',buf=buf, hostname=hostname) class SelfLogHandler(tornado.web.RequestHandler): def get(self):
from platform import uname hostname = uname()[1] self.render('self_log.html',buf=buf, hostname=hostname)
identifier_body
ts_analyzer.py
.datetime.now()-start_time_packet if diff.total_seconds() >= '30': bits_second=1 start_time_packet=datetime.datetime.now() else: bits_second=1+bits_second for i in range(0,len(data),188): offset =+ i #print(offset) sync = ord(data[offset:offset+1]) header1 = ord(data[offset+1:offset+2]) header2 = ord(data[offset+2:offset+3]) header3 = ord(data[offset+3:offset+4]) transport_error = bool(header1 & 0x80) payload_unit_start = bool(header1 & 0x40) transport_priority = bool(header1 & 0x20) pid = header2 | ((header1 & 0x1F) << 8) scrambling = ((header3 & 0xC0) >> 6) have_adaptation_field = bool(header3 & 0x20) adaptation_field = ((header3 & 0x30) >> 4) have_payload = bool(header3 & 0x10) cc = header3 & 0x0F length = len(data) # We have sync: if sync == 0x47: if mcast not in pids: pids[mcast]={} if pid not in pids[mcast]: pids[mcast][pid]={'packets': 1, 'cc': cc, 'error': 0, 'ip': ip, 'type': 'unknown', 'extra': {}} print ("===> Found new PID in stream %s (src=%s)" % (mcast,ip),end='') if pid == 0: pids[mcast][pid]['type'] = "PAT" print (" [PAT] ",end='') buf.append("%s NEW PID: %s (src=%s)(PID: %s [%s]) [PAT]" % (datetime.datetime.now(),mcast,ip,hex(pid),pid)) buf.append("%s NEW PID: %s (src=%s) (PID: %s [%s])" % (datetime.datetime.now(),mcast,ip,hex(pid),pid)) print ("(PID: ",end='') print ("%s"% hex(pid), color='green',end='') print (" [%s])"% pid) else: pids[mcast][pid]['packets']= pids[mcast][pid]['packets']+1 if adaptation_field != 2: cc_com = (pids[mcast][pid]['cc']+1) % 16 pids[mcast][pid]['cc'] = cc if cc is not cc_com: pids[mcast][pid]['error'] = pids[mcast][pid]['error']+1 print ("%s Error expected %s got %s (%s) %s %s" % (datetime.datetime.now(), cc_com, cc, mcast, hex(pid), length), color='red') syslog.syslog(syslog.LOG_ERR, "%s Error expected %s got %s (%s) %s %s" % (datetime.datetime.now(), cc_com, cc, mcast, hex(pid), length)) buf.append( "%s ERR: expected %s got %s (%s) %s %s" % (datetime.datetime.now(), cc_com, cc, mcast, hex(pid), length)) if pid == 0x00: #adaptation_field_size = 0 #print(have_adaptation_field) #if have_adaptation_field: adaptation_field_size = 167 payload_size = 188 - 4 - adaptation_field_size #print(payload_unit_start); #print(mcast) pids[mcast][pid]['extra']=output_program_association_table(data[offset+4:offset+188], payload_size, payload_unit_start) #pp = pprint.PrettyPrinter(indent=4) #pp.pprint(pids) class MainHandler(tornado.web.RequestHandler): def get(self): from platform import uname hostname = uname()[1] run_time = datetime.datetime.now() - start_time packet_time = datetime.datetime.now() - start_time_packet bits=((bits_second*1316)*8/packet_time.total_seconds())/1000000 self.render('index.html',version=ts_analyzer.__version__,addresses=dict(addresses),hostname=hostname, location=location,bits=round(bits,2), run_time=run_time,buf=buf,peers=peers) class LogHandler(tornado.web.RequestHandler): def get(self): from platform import uname hostname = uname()[1] self.render('log.html',buf=buf, hostname=hostname) class SelfLogHandler(tornado.web.RequestHandler): def get(self): from platform import uname hostname = uname()[1] self.render('self_log.html',buf=buf, hostname=hostname) class LogsHandler(tornado.web.RequestHandler): def get(self): from platform import uname hostname = uname()[1] self.render('logs.html',buf=buf, peers=peers, hostname=hostname ) class ChannelHandler(tornado.web.RequestHandler): def get(self): pids_new=pids.copy() for key in pids_new.keys(): if type(key) is not str: try: pids_new[str(key)] = pids_new[key] except: try: pids_new[repr(key)] == pids_new[key] except: pass del pids_new[key] self.write(pids_new) class RingBuffer: def __init__(self, size): self.data = [None for i in xrange(size)] def append(self, x): self.data.pop(0) self.data.append(x) def get(self): return reversed(self.data) class ChannelOverviewHandler(tornado.web.RequestHandler): def get(self): pids_new=pids.copy() for key in pids_new.keys(): if type(key) is not str: try: pids_new[str(key)] = pids_new[key] except: try: pids_new[repr(key)] == pids_new[key] except: pass del pids_new[key] self.render('base.html',pids_new=pids_new) #pp = pprint.PrettyPrinter(indent=4) #pp.pprint(pids) class NewChannelHandler(tornado.web.RequestHandler): def post(self): # Debug #self.write(tornado.escape.json_encode(self.request.arguments["post"])) try: posted_config = tornado.escape.json_decode(self.request.body) except: print("Invalid JSON") pp = pprint.PrettyPrinter(indent=4) pp.pprint(posted_config) #class Server(object): # def __init__(self,address) # self.server = pyuv.UDP(loop._loop) # self.server.bind(key) # self.server.set_membership(key[0], pyuv.UV_JOIN_GROUP) # self.server.start_recv(on_read) if __name__ == "__main__": application = tornado.web.Application([ (r"/", MainHandler), (r"/channels/overview", ChannelOverviewHandler), (r"/channels", ChannelHandler), (r"/channels/new", NewChannelHandler), (r"/logs", LogsHandler), (r"/log", LogHandler), (r"/selflog", SelfLogHandler) ]) parser = argparse.ArgumentParser() parser.add_argument('--version', action='version', version="ts_analyzer %s" % ts_analyzer.__version__) args = parser.parse_args() os.system(['clear', 'cls'][os.name == 'nt']) print ("TS_Analyzer version %s (Using PyUV version %s)" % (ts_analyzer.__version__, pyuv.__version__), color='white', background='blue') template_path = os.path.join(os.path.dirname(__file__), "templates") syslog.syslog("TS_Analyzer version %s (Using PyUV version %s)" % (ts_analyzer.__version__, pyuv.__version__)) pids = {} location = '' location = 'Vrijhof - 253' addresses = {} addresses[("239.192.71.3", 1234)] = 1 addresses[("239.192.27.1", 1234)] = 1 addresses[("239.192.23.1", 1234)] = 1 buf = RingBuffer(100) peers = {} peers["iptv2-cam"]=("130.89.175.42",8889) start_time_packet='unset' bits_second = 1 start_time=datetime.datetime.now() #pp2 = pprint.PrettyPrinter(indent=4) #pp2.pprint(addresses) signal.signal(signal.SIGINT, handle_signal) signal.signal(signal.SIGTERM, handle_signal) application.listen(8889) loop = tornado.ioloop.IOLoop.instance() # for addr in addresses.keys(): # print ("In dict: %s" % (addr)) counter=0
random_line_split
ts_analyzer.py
(sig, frame): tornado.ioloop.IOLoop.instance().add_callback(tornado.ioloop.IOLoop.instance().stop) if sys.version_info >= (3, 0): LINESEP = os.linesep.encode() else: LINESEP = os.linesep def output_program_association_table(f, length, payload_start): #pids[mcast][pid]['extra']='test' #print(" <program_association_table>") pointer_field = None cursor = 0 if payload_start: pointer_field = ord(f[0:1]) # if pointer_field: # print(" <pointer_field>"+str(pointer_field)+"</pointer_field>") cursor+=1 table_id = ord(f[1:2]); cursor+=1 # if table_id: # #=str(pointer_field) # print(" <table_id>"+str(pointer_field)+"</table_id>") byte3 = ord(f[2:3]) ; cursor+=1 # if byte3 & 0x80 != 0x80: # print(" <!-- selection_syntax_indicator is not 1 -->") # if byte3 & 0x40 != 0x00: # print(" <!-- reserved1 is not 0 -->") # if byte3 & 0x30 != 0x30: # print(" <!-- reserved2 is not 11 -->") # if byte3 & 0x0C != 0x00: # print(" <!-- two higher bits of secrion_length is are not 00 -->") byte4 = ord(f[3:4]) ; cursor+=1 section_length = byte4 | ((byte3 & 0x07) << 8) # if section_length: # print(" <section_length>"+str(section_length)+"</section_length>") byte5 = ord(f[4:5]) ; cursor += 1 byte6 = ord(f[5:6]) ; cursor += 1 transport_stream_ID = byte5 << 8 | byte6 # if transport_stream_ID: # print(" <transport_stream_ID>"+str(transport_stream_ID)+"</transport_stream_ID>") byte7 = ord(f[6:7]) ; cursor += 1 # if byte7 & 0xC0 != 0xC0: # # print(" <!-- reserved3 is not 11 -->") version_number = (byte7 & 0x3E) >> 1 # print(" <version_number>"+str(version_number)+"</version_number>") current_indicator = bool(byte7 & 0x01) # if not current_indicator: # print(" <not_appliable_yet/>") section_number = ord(f[7:8]) ; cursor += 1 last_section_number = ord(f[8:9]) ; cursor += 1 # if last_section_number: # print(" <section_number>"+str(section_number)+"</section_number>") # print(" <last_section_number>"+str(last_section_number)+"</last_section_number>") for i in range(0,(section_length-5-4)/4): # print(" <program>") cursor+=4 program_num = (ord(f[9+i:10+i]) << 8) | ord(f[10+i:11+i]) b1 = ord(f[11+i:12+i]) b2 = ord(f[12+i:13+i]) if b1 & 0xE0 != 0xE0: print(" <!-- reserved is not 111 -->") program_pid = b2 | ((b1 & 0x1F) << 8) # print(" <program_num>"+str(program_num)+"</program_num>") # print(" <program_pid>"+hex(program_pid)+"</program_pid>") # print(" </program>\n") #program_map_pids.add(program_pid) crc32 = f[cursor:cursor+4]; cursor+=4 length -= cursor if length>0: rest = f[cursor:cursor+length] if (rest != '\xff' * length) and (rest != '\x00' * length): print(" <rest>"+binascii.hexlify(rest)+"</rest>\n") # print(" </program_association_table>\n") return({'table_id':str(pointer_field),'transportstream_id':str(transport_stream_ID),'program':str(program_num),'pmt':hex(program_pid) }) def output_adaptation_field(f): print(" <adaptation_field>\n") additional_length = ord(f.read(1)) if additional_length == 0: print(" </adaptation_field>\n") return 1 flags = ord(f.read(1)) discontinuity = bool(flags & 0x80) random_access = bool(flags & 0x40) elementary_stream_priority = bool(flags & 0x20) pcr = bool(flags & 0x10) opcr = bool(flags & 0x08) splicing_point = bool(flags & 0x04) transport_private = bool(flags & 0x02) adaptation_field_extension = bool(flags & 0x01) if discontinuity: print(" <discontinuity/>\n") if random_access: print(" <random_access/>\n") if elementary_stream_priority: print(" <elementary_stream_priority/>\n") length = additional_length+1 # size byte additional_length-=1 # flags def read_pcr(): pcr_byte_1 = ord(f.read(1)) # base pcr_byte_2 = ord(f.read(1)) # base pcr_byte_3 = ord(f.read(1)) # base pcr_byte_4 = ord(f.read(1)) # base pcr_byte_5 = ord(f.read(1)) # 1 bit base, 6 bits paddding, 1 bit ext pcr_byte_6 = ord(f.read(1)) # 8 bits ext base = (pcr_byte_1 << (1+8*3)) + \ (pcr_byte_2 << (1+8*2)) + \ (pcr_byte_3 << (1+8*1)) + \ (pcr_byte_4 << (1+8*0)) + \ (pcr_byte_5 >> 7) ext = ((pcr_byte_5 & 0x01) << 8) + pcr_byte_6 time = base / 90000.0 + ext / 27000000.0 return time if pcr: if additional_length>=6: additional_length-=6 val = read_pcr() print(" <program_clock_reference>"+str(val)+"</program_clock_reference>\n") if opcr: if additional_length>=6: additional_length-=6 val = read_pcr() print(" <original_program_clock_reference>"+str(val)+"</original_program_clock_reference>\n") if splicing_point: if additional_length>=1: additional_length-=1 splice_count = ord(f.read(1)) print(" <splice_countdown>"+str(splice_count)+"</splice_countdown>\n") if additional_length: print(" <!-- ignoring " + str(additional_length) + " bytes -->\n") f.read(additional_length) print(" </adaptation_field>\n") return length def on_read(handle, ip_port, flags, data, error): global bits_second, start_time_packet if error is not None: print (error,color='red') return if start_time_packet == 'unset': start_time_packet=datetime.datetime.now() data = data.strip() mcast=handle.getsockname() if data: ip, port = ip_port diff = datetime.datetime.now()-start_time_packet if diff.total_seconds() >= '30': bits_second=1 start_time_packet=datetime.datetime.now() else: bits_second=1+bits_second for i in range(0,len(data),188): offset =+ i #print(offset) sync = ord(data[offset:offset+1]) header1 = ord(data[offset+1:offset+2]) header2 = ord(data[offset+2:offset+3]) header3 = ord(data[offset+3:offset+4]) transport_error = bool(header1 & 0x80) payload_unit_start = bool(header1 & 0x40) transport_priority = bool(header1 & 0x20) pid = header2 | ((header1 & 0x1F) << 8) scrambling = ((header3 & 0xC0) >> 6) have_adaptation_field = bool(header3 & 0x20) adaptation_field = ((header3 & 0x30) >> 4) have_payload = bool(header3
handle_signal
identifier_name
map_output_tracker.rs
Ext, }; use tokio_util::compat::{Tokio02AsyncReadCompatExt, Tokio02AsyncWriteCompatExt}; const CAPNP_BUF_READ_OPTS: ReaderOptions = ReaderOptions { traversal_limit_in_words: std::u64::MAX, nesting_limit: 64, }; pub(crate) enum MapOutputTrackerMessage { // Contains shuffle_id GetMapOutputLocations(i64), StopMapOutputTracker, } /// The key is the shuffle_id pub type ServerUris = Arc<DashMap<usize, Vec<Option<String>>>>; // Starts the server in master node and client in slave nodes. Similar to cache tracker. #[derive(Clone, Debug)] pub(crate) struct MapOutputTracker { is_master: bool, pub server_uris: ServerUris, fetching: Arc<DashSet<usize>>, generation: Arc<Mutex<i64>>, master_addr: SocketAddr, } // Only master_addr doesn't have a default. impl Default for MapOutputTracker { fn default() -> Self { MapOutputTracker { is_master: Default::default(), server_uris: Default::default(), fetching: Default::default(), generation: Default::default(), master_addr: SocketAddr::new(IpAddr::V4(Ipv4Addr::new(0, 0, 0, 0)), 0), } } } impl MapOutputTracker { pub fn new(is_master: bool, master_addr: SocketAddr) -> Self { let output_tracker = MapOutputTracker { is_master, server_uris: Arc::new(DashMap::new()), fetching: Arc::new(DashSet::new()), generation: Arc::new(Mutex::new(0)), master_addr, }; output_tracker.server(); output_tracker } async fn client(&self, shuffle_id: usize) -> Result<Vec<String>> { let mut stream = loop { match TcpStream::connect(self.master_addr).await { Ok(stream) => break stream, Err(_) => continue, } }; let (reader, writer) = stream.split(); let reader = reader.compat(); let mut writer = writer.compat_write(); log::debug!( "connected to master to fetch shuffle task #{} data hosts", shuffle_id ); let shuffle_id_bytes = bincode::serialize(&shuffle_id)?; let mut message = MsgBuilder::new_default(); let mut shuffle_data = message.init_root::<serialized_data::Builder>(); shuffle_data.set_msg(&shuffle_id_bytes); capnp_serialize::write_message(&mut writer, &message).await?; let message_reader = capnp_serialize::read_message(reader, CAPNP_BUF_READ_OPTS) .await? .ok_or_else(|| NetworkError::NoMessageReceived)?; let shuffle_data = message_reader.get_root::<serialized_data::Reader>()?; let locs: Vec<String> = bincode::deserialize(&shuffle_data.get_msg()?)?; Ok(locs) } fn server(&self) { if !self.is_master { return; } log::debug!("map output tracker server starting"); let master_addr = self.master_addr; let server_uris = self.server_uris.clone(); tokio::spawn(async move { let mut listener = TcpListener::bind(master_addr) .await .map_err(NetworkError::TcpListener)?; log::debug!("map output tracker server started"); while let Some(Ok(mut stream)) = listener.incoming().next().await { let server_uris_clone = server_uris.clone(); tokio::spawn(async move { let (reader, writer) = stream.split(); let reader = reader.compat(); let writer = writer.compat_write(); // reading let message_reader = capnp_serialize::read_message(reader, CAPNP_BUF_READ_OPTS) .await? .ok_or_else(|| NetworkError::NoMessageReceived)?; let shuffle_id = { let data = message_reader.get_root::<serialized_data::Reader>()?; bincode::deserialize(data.get_msg()?)? }; while server_uris_clone .get(&shuffle_id) .ok_or_else(|| MapOutputError::ShuffleIdNotFound(shuffle_id))? .iter() .filter(|x| !x.is_none()) .count() == 0 { //check whether this will hurt the performance or not tokio::time::delay_for(Duration::from_millis(1)).await; } let locs = server_uris_clone .get(&shuffle_id) .map(|kv| { kv.value() .iter() .cloned() .map(|x| x.unwrap()) .collect::<Vec<_>>() }) .unwrap_or_default(); log::debug!( "locs inside map output tracker server for shuffle id #{}: {:?}", shuffle_id, locs ); // writting response let result = bincode::serialize(&locs)?; let mut message = MsgBuilder::new_default(); let mut locs_data = message.init_root::<serialized_data::Builder>(); locs_data.set_msg(&result); // TODO: remove blocking call when possible futures::executor::block_on(async { capnp_futures::serialize::write_message(writer, message) .await .map_err(Error::CapnpDeserialization)?; Ok::<_, Error>(()) })?; Ok::<_, Error>(()) }); } Err::<(), _>(Error::ExecutorShutdown) }); } pub fn register_shuffle(&self, shuffle_id: usize, num_maps: usize) { log::debug!("inside register shuffle"); if self.server_uris.get(&shuffle_id).is_some() { // TODO: error handling log::debug!("map tracker register shuffle none"); return; } self.server_uris.insert(shuffle_id, vec![None; num_maps]); log::debug!("server_uris after register_shuffle {:?}", self.server_uris); } pub fn register_map_output(&self, shuffle_id: usize, map_id: usize, server_uri: String) { log::debug!( "registering map output from shuffle task #{} with map id #{} at server: {}", shuffle_id, map_id, server_uri ); self.server_uris.get_mut(&shuffle_id).unwrap()[map_id] = Some(server_uri); } pub fn register_map_outputs(&self, shuffle_id: usize, locs: Vec<Option<String>>) { log::debug!( "registering map outputs inside map output tracker for shuffle id #{}: {:?}", shuffle_id, locs ); self.server_uris.insert(shuffle_id, locs); } pub fn unregister_map_output(&self, shuffle_id: usize, map_id: usize, server_uri: String) { let array = self.server_uris.get(&shuffle_id); if let Some(arr) = array { if arr.get(map_id).unwrap() == &Some(server_uri) { self.server_uris .get_mut(&shuffle_id) .unwrap() .insert(map_id, None) } self.increment_generation(); } else { // TODO: error logging } } pub async fn get_server_uris(&self, shuffle_id: usize) -> Result<Vec<String>> { log::debug!( "trying to get uri for shuffle task #{}, current server uris: {:?}", shuffle_id, self.server_uris ); if self .server_uris .get(&shuffle_id) .map(|some| some.iter().filter_map(|x| x.clone()).next()) .flatten() .is_none() { if self.fetching.contains(&shuffle_id) { while self.fetching.contains(&shuffle_id) { // TODO: check whether this will hurt the performance or not tokio::time::delay_for(Duration::from_millis(1)).await; } let servers = self .server_uris .get(&shuffle_id) .ok_or_else(|| MapOutputError::ShuffleIdNotFound(shuffle_id))? .iter() .filter(|x| !x.is_none()) .map(|x| x.clone().unwrap()) .collect::<Vec<_>>(); log::debug!("returning after fetching done, return: {:?}", servers); return Ok(servers); } else { log::debug!("adding to fetching queue"); self.fetching.insert(shuffle_id); } let fetched = self.client(shuffle_id).await?; log::debug!("fetched locs from client: {:?}", fetched); self.server_uris.insert( shuffle_id, fetched.iter().map(|x| Some(x.clone())).collect(), ); log::debug!("added locs to server uris after fetching"); self.fetching.remove(&shuffle_id); Ok(fetched) } else { Ok(self .server_uris .get(&shuffle_id) .ok_or_else(|| MapOutputError::ShuffleIdNotFound(shuffle_id))? .iter() .filter(|x| !x.is_none()) .map(|x| x.clone().unwrap()) .collect()) } } pub fn increment_generation(&self) { *self.generation.lock() += 1; } pub fn get_generation(&self) -> i64 { *self.generation.lock() }
random_line_split
map_output_tracker.rs
}; use capnp_futures::serialize as capnp_serialize; use dashmap::{DashMap, DashSet}; use parking_lot::Mutex; use thiserror::Error; use tokio::{ net::{TcpListener, TcpStream}, stream::StreamExt, }; use tokio_util::compat::{Tokio02AsyncReadCompatExt, Tokio02AsyncWriteCompatExt}; const CAPNP_BUF_READ_OPTS: ReaderOptions = ReaderOptions { traversal_limit_in_words: std::u64::MAX, nesting_limit: 64, }; pub(crate) enum MapOutputTrackerMessage { // Contains shuffle_id GetMapOutputLocations(i64), StopMapOutputTracker, } /// The key is the shuffle_id pub type ServerUris = Arc<DashMap<usize, Vec<Option<String>>>>; // Starts the server in master node and client in slave nodes. Similar to cache tracker. #[derive(Clone, Debug)] pub(crate) struct MapOutputTracker { is_master: bool, pub server_uris: ServerUris, fetching: Arc<DashSet<usize>>, generation: Arc<Mutex<i64>>, master_addr: SocketAddr, } // Only master_addr doesn't have a default. impl Default for MapOutputTracker { fn default() -> Self { MapOutputTracker { is_master: Default::default(), server_uris: Default::default(), fetching: Default::default(), generation: Default::default(), master_addr: SocketAddr::new(IpAddr::V4(Ipv4Addr::new(0, 0, 0, 0)), 0), } } } impl MapOutputTracker { pub fn
(is_master: bool, master_addr: SocketAddr) -> Self { let output_tracker = MapOutputTracker { is_master, server_uris: Arc::new(DashMap::new()), fetching: Arc::new(DashSet::new()), generation: Arc::new(Mutex::new(0)), master_addr, }; output_tracker.server(); output_tracker } async fn client(&self, shuffle_id: usize) -> Result<Vec<String>> { let mut stream = loop { match TcpStream::connect(self.master_addr).await { Ok(stream) => break stream, Err(_) => continue, } }; let (reader, writer) = stream.split(); let reader = reader.compat(); let mut writer = writer.compat_write(); log::debug!( "connected to master to fetch shuffle task #{} data hosts", shuffle_id ); let shuffle_id_bytes = bincode::serialize(&shuffle_id)?; let mut message = MsgBuilder::new_default(); let mut shuffle_data = message.init_root::<serialized_data::Builder>(); shuffle_data.set_msg(&shuffle_id_bytes); capnp_serialize::write_message(&mut writer, &message).await?; let message_reader = capnp_serialize::read_message(reader, CAPNP_BUF_READ_OPTS) .await? .ok_or_else(|| NetworkError::NoMessageReceived)?; let shuffle_data = message_reader.get_root::<serialized_data::Reader>()?; let locs: Vec<String> = bincode::deserialize(&shuffle_data.get_msg()?)?; Ok(locs) } fn server(&self) { if !self.is_master { return; } log::debug!("map output tracker server starting"); let master_addr = self.master_addr; let server_uris = self.server_uris.clone(); tokio::spawn(async move { let mut listener = TcpListener::bind(master_addr) .await .map_err(NetworkError::TcpListener)?; log::debug!("map output tracker server started"); while let Some(Ok(mut stream)) = listener.incoming().next().await { let server_uris_clone = server_uris.clone(); tokio::spawn(async move { let (reader, writer) = stream.split(); let reader = reader.compat(); let writer = writer.compat_write(); // reading let message_reader = capnp_serialize::read_message(reader, CAPNP_BUF_READ_OPTS) .await? .ok_or_else(|| NetworkError::NoMessageReceived)?; let shuffle_id = { let data = message_reader.get_root::<serialized_data::Reader>()?; bincode::deserialize(data.get_msg()?)? }; while server_uris_clone .get(&shuffle_id) .ok_or_else(|| MapOutputError::ShuffleIdNotFound(shuffle_id))? .iter() .filter(|x| !x.is_none()) .count() == 0 { //check whether this will hurt the performance or not tokio::time::delay_for(Duration::from_millis(1)).await; } let locs = server_uris_clone .get(&shuffle_id) .map(|kv| { kv.value() .iter() .cloned() .map(|x| x.unwrap()) .collect::<Vec<_>>() }) .unwrap_or_default(); log::debug!( "locs inside map output tracker server for shuffle id #{}: {:?}", shuffle_id, locs ); // writting response let result = bincode::serialize(&locs)?; let mut message = MsgBuilder::new_default(); let mut locs_data = message.init_root::<serialized_data::Builder>(); locs_data.set_msg(&result); // TODO: remove blocking call when possible futures::executor::block_on(async { capnp_futures::serialize::write_message(writer, message) .await .map_err(Error::CapnpDeserialization)?; Ok::<_, Error>(()) })?; Ok::<_, Error>(()) }); } Err::<(), _>(Error::ExecutorShutdown) }); } pub fn register_shuffle(&self, shuffle_id: usize, num_maps: usize) { log::debug!("inside register shuffle"); if self.server_uris.get(&shuffle_id).is_some() { // TODO: error handling log::debug!("map tracker register shuffle none"); return; } self.server_uris.insert(shuffle_id, vec![None; num_maps]); log::debug!("server_uris after register_shuffle {:?}", self.server_uris); } pub fn register_map_output(&self, shuffle_id: usize, map_id: usize, server_uri: String) { log::debug!( "registering map output from shuffle task #{} with map id #{} at server: {}", shuffle_id, map_id, server_uri ); self.server_uris.get_mut(&shuffle_id).unwrap()[map_id] = Some(server_uri); } pub fn register_map_outputs(&self, shuffle_id: usize, locs: Vec<Option<String>>) { log::debug!( "registering map outputs inside map output tracker for shuffle id #{}: {:?}", shuffle_id, locs ); self.server_uris.insert(shuffle_id, locs); } pub fn unregister_map_output(&self, shuffle_id: usize, map_id: usize, server_uri: String) { let array = self.server_uris.get(&shuffle_id); if let Some(arr) = array { if arr.get(map_id).unwrap() == &Some(server_uri) { self.server_uris .get_mut(&shuffle_id) .unwrap() .insert(map_id, None) } self.increment_generation(); } else { // TODO: error logging } } pub async fn get_server_uris(&self, shuffle_id: usize) -> Result<Vec<String>> { log::debug!( "trying to get uri for shuffle task #{}, current server uris: {:?}", shuffle_id, self.server_uris ); if self .server_uris .get(&shuffle_id) .map(|some| some.iter().filter_map(|x| x.clone()).next()) .flatten() .is_none() { if self.fetching.contains(&shuffle_id) { while self.fetching.contains(&shuffle_id) { // TODO: check whether this will hurt the performance or not tokio::time::delay_for(Duration::from_millis(1)).await; } let servers = self .server_uris .get(&shuffle_id) .ok_or_else(|| MapOutputError::ShuffleIdNotFound(shuffle_id))? .iter() .filter(|x| !x.is_none()) .map(|x| x.clone().unwrap()) .collect::<Vec<_>>(); log::debug!("returning after fetching done, return: {:?}", servers); return Ok(servers); } else { log::debug!("adding to fetching queue"); self.fetching.insert(shuffle_id); } let fetched = self.client(shuffle_id).await?; log::debug!("fetched locs from client: {:?}", fetched); self.server_uris.insert( shuffle_id, fetched.iter().map(|x| Some(x.clone())).collect(), ); log::debug!("added locs to server uris after fetching"); self.fetching.remove(&shuffle_id); Ok(fetched) } else { Ok(self .server_uris .get(&shuffle_id) .ok_or_else(|| MapOutputError::ShuffleIdNotFound(shuffle_id))? .iter() .filter(|x| !x.is_none()) .map(|x| x.clone().unwrap
new
identifier_name
map_output_tracker.rs
}; use capnp_futures::serialize as capnp_serialize; use dashmap::{DashMap, DashSet}; use parking_lot::Mutex; use thiserror::Error; use tokio::{ net::{TcpListener, TcpStream}, stream::StreamExt, }; use tokio_util::compat::{Tokio02AsyncReadCompatExt, Tokio02AsyncWriteCompatExt}; const CAPNP_BUF_READ_OPTS: ReaderOptions = ReaderOptions { traversal_limit_in_words: std::u64::MAX, nesting_limit: 64, }; pub(crate) enum MapOutputTrackerMessage { // Contains shuffle_id GetMapOutputLocations(i64), StopMapOutputTracker, } /// The key is the shuffle_id pub type ServerUris = Arc<DashMap<usize, Vec<Option<String>>>>; // Starts the server in master node and client in slave nodes. Similar to cache tracker. #[derive(Clone, Debug)] pub(crate) struct MapOutputTracker { is_master: bool, pub server_uris: ServerUris, fetching: Arc<DashSet<usize>>, generation: Arc<Mutex<i64>>, master_addr: SocketAddr, } // Only master_addr doesn't have a default. impl Default for MapOutputTracker { fn default() -> Self { MapOutputTracker { is_master: Default::default(), server_uris: Default::default(), fetching: Default::default(), generation: Default::default(), master_addr: SocketAddr::new(IpAddr::V4(Ipv4Addr::new(0, 0, 0, 0)), 0), } } } impl MapOutputTracker { pub fn new(is_master: bool, master_addr: SocketAddr) -> Self { let output_tracker = MapOutputTracker { is_master, server_uris: Arc::new(DashMap::new()), fetching: Arc::new(DashSet::new()), generation: Arc::new(Mutex::new(0)), master_addr, }; output_tracker.server(); output_tracker } async fn client(&self, shuffle_id: usize) -> Result<Vec<String>> { let mut stream = loop { match TcpStream::connect(self.master_addr).await { Ok(stream) => break stream, Err(_) => continue, } }; let (reader, writer) = stream.split(); let reader = reader.compat(); let mut writer = writer.compat_write(); log::debug!( "connected to master to fetch shuffle task #{} data hosts", shuffle_id ); let shuffle_id_bytes = bincode::serialize(&shuffle_id)?; let mut message = MsgBuilder::new_default(); let mut shuffle_data = message.init_root::<serialized_data::Builder>(); shuffle_data.set_msg(&shuffle_id_bytes); capnp_serialize::write_message(&mut writer, &message).await?; let message_reader = capnp_serialize::read_message(reader, CAPNP_BUF_READ_OPTS) .await? .ok_or_else(|| NetworkError::NoMessageReceived)?; let shuffle_data = message_reader.get_root::<serialized_data::Reader>()?; let locs: Vec<String> = bincode::deserialize(&shuffle_data.get_msg()?)?; Ok(locs) } fn server(&self) { if !self.is_master { return; } log::debug!("map output tracker server starting"); let master_addr = self.master_addr; let server_uris = self.server_uris.clone(); tokio::spawn(async move { let mut listener = TcpListener::bind(master_addr) .await .map_err(NetworkError::TcpListener)?; log::debug!("map output tracker server started"); while let Some(Ok(mut stream)) = listener.incoming().next().await { let server_uris_clone = server_uris.clone(); tokio::spawn(async move { let (reader, writer) = stream.split(); let reader = reader.compat(); let writer = writer.compat_write(); // reading let message_reader = capnp_serialize::read_message(reader, CAPNP_BUF_READ_OPTS) .await? .ok_or_else(|| NetworkError::NoMessageReceived)?; let shuffle_id = { let data = message_reader.get_root::<serialized_data::Reader>()?; bincode::deserialize(data.get_msg()?)? }; while server_uris_clone .get(&shuffle_id) .ok_or_else(|| MapOutputError::ShuffleIdNotFound(shuffle_id))? .iter() .filter(|x| !x.is_none()) .count() == 0 { //check whether this will hurt the performance or not tokio::time::delay_for(Duration::from_millis(1)).await; } let locs = server_uris_clone .get(&shuffle_id) .map(|kv| { kv.value() .iter() .cloned() .map(|x| x.unwrap()) .collect::<Vec<_>>() }) .unwrap_or_default(); log::debug!( "locs inside map output tracker server for shuffle id #{}: {:?}", shuffle_id, locs ); // writting response let result = bincode::serialize(&locs)?; let mut message = MsgBuilder::new_default(); let mut locs_data = message.init_root::<serialized_data::Builder>(); locs_data.set_msg(&result); // TODO: remove blocking call when possible futures::executor::block_on(async { capnp_futures::serialize::write_message(writer, message) .await .map_err(Error::CapnpDeserialization)?; Ok::<_, Error>(()) })?; Ok::<_, Error>(()) }); } Err::<(), _>(Error::ExecutorShutdown) }); } pub fn register_shuffle(&self, shuffle_id: usize, num_maps: usize) { log::debug!("inside register shuffle"); if self.server_uris.get(&shuffle_id).is_some() { // TODO: error handling log::debug!("map tracker register shuffle none"); return; } self.server_uris.insert(shuffle_id, vec![None; num_maps]); log::debug!("server_uris after register_shuffle {:?}", self.server_uris); } pub fn register_map_output(&self, shuffle_id: usize, map_id: usize, server_uri: String) { log::debug!( "registering map output from shuffle task #{} with map id #{} at server: {}", shuffle_id, map_id, server_uri ); self.server_uris.get_mut(&shuffle_id).unwrap()[map_id] = Some(server_uri); } pub fn register_map_outputs(&self, shuffle_id: usize, locs: Vec<Option<String>>) { log::debug!( "registering map outputs inside map output tracker for shuffle id #{}: {:?}", shuffle_id, locs ); self.server_uris.insert(shuffle_id, locs); } pub fn unregister_map_output(&self, shuffle_id: usize, map_id: usize, server_uri: String) { let array = self.server_uris.get(&shuffle_id); if let Some(arr) = array
else { // TODO: error logging } } pub async fn get_server_uris(&self, shuffle_id: usize) -> Result<Vec<String>> { log::debug!( "trying to get uri for shuffle task #{}, current server uris: {:?}", shuffle_id, self.server_uris ); if self .server_uris .get(&shuffle_id) .map(|some| some.iter().filter_map(|x| x.clone()).next()) .flatten() .is_none() { if self.fetching.contains(&shuffle_id) { while self.fetching.contains(&shuffle_id) { // TODO: check whether this will hurt the performance or not tokio::time::delay_for(Duration::from_millis(1)).await; } let servers = self .server_uris .get(&shuffle_id) .ok_or_else(|| MapOutputError::ShuffleIdNotFound(shuffle_id))? .iter() .filter(|x| !x.is_none()) .map(|x| x.clone().unwrap()) .collect::<Vec<_>>(); log::debug!("returning after fetching done, return: {:?}", servers); return Ok(servers); } else { log::debug!("adding to fetching queue"); self.fetching.insert(shuffle_id); } let fetched = self.client(shuffle_id).await?; log::debug!("fetched locs from client: {:?}", fetched); self.server_uris.insert( shuffle_id, fetched.iter().map(|x| Some(x.clone())).collect(), ); log::debug!("added locs to server uris after fetching"); self.fetching.remove(&shuffle_id); Ok(fetched) } else { Ok(self .server_uris .get(&shuffle_id) .ok_or_else(|| MapOutputError::ShuffleIdNotFound(shuffle_id))? .iter() .filter(|x| !x.is_none()) .map(|x| x.clone().unwrap
{ if arr.get(map_id).unwrap() == &Some(server_uri) { self.server_uris .get_mut(&shuffle_id) .unwrap() .insert(map_id, None) } self.increment_generation(); }
conditional_block
core.py
, -1, style=wx.LC_REPORT) ListCtrlAutoWidthMixin.__init__(self) class Core(wx.Frame): def __init__(self, parent, controller, title): super(Core, self).__init__(parent=parent, title=title) self.parent = parent self.controller = controller self.child = None self.panel = Panel(self) self.panel.SetBackgroundColour('LightGray') self.status_bar = self.CreateStatusBar(2) self.status_bar.SetStatusWidths([200, -1]) gauge_pos, gauge_size = self.get_gauge_dimensions() self.gauge = wx.Gauge(self.status_bar, -1, 100, gauge_pos, gauge_size) # Menues self.menubar = wx.MenuBar() self.SetMenuBar(self.menubar) # Player menu menu_player = wx.Menu() self.menubar.Append(menu_player, "Players") self.menu_new_player = menu_player.Append(-1, "New Player", "New Player") menu_player.AppendSeparator() self.menu_players_import = menu_player.Append(-1, "import Players", "import Players") menu_player.AppendSeparator() self.menu_evaluations = menu_player.Append(-1, "extract evaluations", "extract evaluations") # Bindings # player bindings self.Bind(wx.EVT_SIZE, self.on_size) self.Bind(wx.EVT_MENU, self.on_new_player, self.menu_new_player) self.Bind(wx.EVT_MENU, self.on_import, self.menu_players_import) self.Bind(wx.EVT_MENU, self.on_extract, self.menu_evaluations) self.Bind(wx.EVT_BUTTON, self.quit, self.panel.btn_quit) self.Bind(wx.EVT_BUTTON, self.on_refresh, self.panel.btn_refresh) self.Bind(wx.EVT_LIST_ITEM_SELECTED, self.on_list, self.panel.players) self.Bind(wx.EVT_LIST_COL_CLICK, self.on_list_column, self.panel.players) # Players initialization players = self.controller.get_players() if players: self.fill_players(players) self.set_status_text('Players on database: %s' % len(players)) else: self.set_status_text('No Players on database') size = (450, 500) self.SetSize(size) self.Show() # noinspection PyUnusedLocal def quit(self, event): self.Destroy() # noinspection PyUnusedLocal def on_extract(self, event): self.child = ViewExtract(parent=self, title="extract evaluations") self.show_child() # noinspection PyUnusedLocal def on_import(self, event): choice = wx.MessageBox('Deleting All Players?', 'warning', YN) if choice == wx.YES: self.controller.delete_all_players() self.controller.import_players() wx.MessageBox('Players successfully imported!', '', OK) players = self.controller.get_players() self.fill_players(players) # noinspection PyUnusedLocal def edit_player(self, event): self.Disable() ViewPlayer(parent=self, title='Edit Player', is_editor=True) # noinspection PyUnusedLocal def on_new_player(self, event): self.child = ViewPlayer(parent=self, title='New Player') self.show_child() # noinspection PyUnusedLocal def on_quit(self, event): self.Destroy() # noinspection PyUnusedLocal def on_refresh(self, event): self.panel.players.DeleteAllItems() players = self.controller.get_players() self.fill_players(players) # noinspection PyUnusedLocal def on_list(self, event): item_id = event.m_itemIndex player_code = self.panel.players.GetItemText(item_id) player = self.controller.get_player_by_code(player_code) self.controller.set_temporary_object(player) item_name = self.panel.players.GetItem(item_id, 1) player_name = item_name.GetText() item_fullname = self.panel.players.GetItem(item_id, 2) player_fullname = item_fullname.GetText() view_edit = ViewPlayer(self, "Edit player", is_editor=True) view_edit.Show() view_edit.panel.code.SetValue(player_code) view_edit.panel.name.SetValue(player_name) view_edit.panel.fullname.SetValue(player_fullname) # Qui utilizzo ChangeValue per non scatenare EVT_TEXT, come invece # succederebbe usando SetValue. Questo EVT_TEXT, cercherebbe un # Handler al livello attuale (Frame) e non trovandolo, passerebbe # al Parent, che invece ha proprio un Bind di EVT_TEXT, proprio con # la textctrl 'ppl', la quale richiamerebbe la callback # 'on_text_entry, facendo casino. # view_edit.panel.btn_delete.Enable() view_edit.SetWindowStyle(wx.STAY_ON_TOP) def on_list_column(self, event): self.panel.players.DeleteAllItems() id_column = event.GetColumn() players = self.controller.get_sorted_players(id_column) self.fill_players(players) def fill_players(self, players): for player in players: index = self.panel.players.InsertStringItem(sys.maxint, str(player.code)) self.panel.players.SetStringItem(index, 1, str(player.name)) self.panel.players.SetStringItem(index, 2, str(player.fullname)) @staticmethod def show_message(string): wx.MessageBox(string, 'core info', wx.OK | wx.ICON_EXCLAMATION) def show_child(self): self.Disable() self.child.Centre() self.child.Show() def get_gauge_dimensions(self): """ get_gauge_dimensions(self) -> tuple_a, tuple_b tuple_a is a tuple with x position and y position of seconf field of the StatusBar """ pos_x, pos_y, dim_x, dim_y = self.status_bar.GetFieldRect(1) return (pos_x, pos_y), (dim_x, dim_y) def on_size(self, event): """ on_size() it redraws the gauge rectangle and repositions when frame windows is resized """ size = self.GetSize() self.SetSize(size) gauge_pos, gauge_size = self.get_gauge_dimensions() self.gauge.SetSize(gauge_size) event.Skip() self.Update() def set_range(self, value): """ set_range(value) It sets the maximum value of gauge widget """ self.gauge.SetRange(value) def set_progress(self, value): """ set_progress(value) It sets the actual progress value to gauge widget """ self.gauge.SetValue(value) def set_status_text(self, value): """ set_status_text(value) It sets the text to the first field of StatusBar """ self.status_bar.SetStatusText(value) class Panel(wx.Panel): def __init__(self, parent): super(Panel, self).__init__(parent=parent) self.players = AutoWidthListCtrl(self) self.players.InsertColumn(0, 'code', wx.LIST_FORMAT_RIGHT, 50) self.players.InsertColumn(1, 'name', width=125) self.players.InsertColumn(2, 'fullname', width=175) players_box = wx.BoxSizer(wx.HORIZONTAL) players_box.Add(self.players, 1, wx.EXPAND) btn_sizer = wx.FlexGridSizer(rows=1, cols=2, hgap=5, vgap=5) self.btn_quit = wx.Button(self, wx.ID_CANCEL, label="Quit") self.btn_refresh = wx.Button(self, wx.ID_OK, label="Refresh") btn_sizer.Add(self.btn_quit, 0, wx.EXPAND) btn_sizer.Add(self.btn_refresh, 0, wx.EXPAND) sizer = wx.BoxSizer(wx.VERTICAL) sizer.Add(players_box, 1, wx.EXPAND | wx.ALL, 5) sizer.Add(btn_sizer, 0, wx.EXPAND | wx.ALL, 5) self.SetSizer(sizer) class ViewExtract(wx.Frame): def __init__(self, parent, title): self.parent = parent self.title = title super(ViewExtract, self).__init__(parent=self.parent, title=title) self.controller = self.parent.controller self.panel = PanelExtract(parent=self) self.SetSize((300, 150)) # bindings self.Bind(wx.EVT_CLOSE, self.on_quit) self.Bind(wx.EVT_BUTTON, self.on_quit, self.panel.btn_quit) self.Bind(wx.EVT_BUTTON, self.on_extract, self.panel.btn_extract) # noinspection PyUnusedLocal def on_extract(self, event): day = self.panel.day.GetValue() if day: if self.controller.are_evaluations_ready(day): self.controller.extract_evaluations(day) self.parent.Enable() self.Destroy() else: wx.MessageBox('Evaluations for day %s not ready!' % day, '', OK) else:
# noinspection PyUnusedLocal def on_quit(self, event): self.parent.Enable() self.Destroy() class PanelExtract(wx.Panel): def __init__(self, parent): super(PanelExtract, self).__init__(parent) # Attributes self.day = wx.TextCtrl(self) # Layout text_sizer = wx.FlexGridSizer(rows=1, cols=2, hgap=5, vgap
wx.MessageBox('Please set a day to extract!', '', OK)
conditional_block
core.py
parent, -1, style=wx.LC_REPORT) ListCtrlAutoWidthMixin.__init__(self) class Core(wx.Frame): def __init__(self, parent, controller, title): super(Core, self).__init__(parent=parent, title=title) self.parent = parent self.controller = controller self.child = None self.panel = Panel(self) self.panel.SetBackgroundColour('LightGray') self.status_bar = self.CreateStatusBar(2) self.status_bar.SetStatusWidths([200, -1]) gauge_pos, gauge_size = self.get_gauge_dimensions() self.gauge = wx.Gauge(self.status_bar, -1, 100, gauge_pos, gauge_size) # Menues self.menubar = wx.MenuBar() self.SetMenuBar(self.menubar) # Player menu menu_player = wx.Menu() self.menubar.Append(menu_player, "Players") self.menu_new_player = menu_player.Append(-1, "New Player", "New Player") menu_player.AppendSeparator() self.menu_players_import = menu_player.Append(-1, "import Players", "import Players") menu_player.AppendSeparator() self.menu_evaluations = menu_player.Append(-1, "extract evaluations", "extract evaluations") # Bindings # player bindings self.Bind(wx.EVT_SIZE, self.on_size) self.Bind(wx.EVT_MENU, self.on_new_player, self.menu_new_player) self.Bind(wx.EVT_MENU, self.on_import, self.menu_players_import) self.Bind(wx.EVT_MENU, self.on_extract, self.menu_evaluations) self.Bind(wx.EVT_BUTTON, self.quit, self.panel.btn_quit) self.Bind(wx.EVT_BUTTON, self.on_refresh, self.panel.btn_refresh) self.Bind(wx.EVT_LIST_ITEM_SELECTED, self.on_list, self.panel.players) self.Bind(wx.EVT_LIST_COL_CLICK, self.on_list_column, self.panel.players) # Players initialization players = self.controller.get_players() if players: self.fill_players(players) self.set_status_text('Players on database: %s' % len(players)) else: self.set_status_text('No Players on database') size = (450, 500) self.SetSize(size) self.Show() # noinspection PyUnusedLocal def quit(self, event): self.Destroy() # noinspection PyUnusedLocal def on_extract(self, event): self.child = ViewExtract(parent=self, title="extract evaluations") self.show_child() # noinspection PyUnusedLocal def on_import(self, event): choice = wx.MessageBox('Deleting All Players?', 'warning', YN) if choice == wx.YES: self.controller.delete_all_players() self.controller.import_players() wx.MessageBox('Players successfully imported!', '', OK) players = self.controller.get_players() self.fill_players(players) # noinspection PyUnusedLocal def edit_player(self, event): self.Disable()
def on_new_player(self, event): self.child = ViewPlayer(parent=self, title='New Player') self.show_child() # noinspection PyUnusedLocal def on_quit(self, event): self.Destroy() # noinspection PyUnusedLocal def on_refresh(self, event): self.panel.players.DeleteAllItems() players = self.controller.get_players() self.fill_players(players) # noinspection PyUnusedLocal def on_list(self, event): item_id = event.m_itemIndex player_code = self.panel.players.GetItemText(item_id) player = self.controller.get_player_by_code(player_code) self.controller.set_temporary_object(player) item_name = self.panel.players.GetItem(item_id, 1) player_name = item_name.GetText() item_fullname = self.panel.players.GetItem(item_id, 2) player_fullname = item_fullname.GetText() view_edit = ViewPlayer(self, "Edit player", is_editor=True) view_edit.Show() view_edit.panel.code.SetValue(player_code) view_edit.panel.name.SetValue(player_name) view_edit.panel.fullname.SetValue(player_fullname) # Qui utilizzo ChangeValue per non scatenare EVT_TEXT, come invece # succederebbe usando SetValue. Questo EVT_TEXT, cercherebbe un # Handler al livello attuale (Frame) e non trovandolo, passerebbe # al Parent, che invece ha proprio un Bind di EVT_TEXT, proprio con # la textctrl 'ppl', la quale richiamerebbe la callback # 'on_text_entry, facendo casino. # view_edit.panel.btn_delete.Enable() view_edit.SetWindowStyle(wx.STAY_ON_TOP) def on_list_column(self, event): self.panel.players.DeleteAllItems() id_column = event.GetColumn() players = self.controller.get_sorted_players(id_column) self.fill_players(players) def fill_players(self, players): for player in players: index = self.panel.players.InsertStringItem(sys.maxint, str(player.code)) self.panel.players.SetStringItem(index, 1, str(player.name)) self.panel.players.SetStringItem(index, 2, str(player.fullname)) @staticmethod def show_message(string): wx.MessageBox(string, 'core info', wx.OK | wx.ICON_EXCLAMATION) def show_child(self): self.Disable() self.child.Centre() self.child.Show() def get_gauge_dimensions(self): """ get_gauge_dimensions(self) -> tuple_a, tuple_b tuple_a is a tuple with x position and y position of seconf field of the StatusBar """ pos_x, pos_y, dim_x, dim_y = self.status_bar.GetFieldRect(1) return (pos_x, pos_y), (dim_x, dim_y) def on_size(self, event): """ on_size() it redraws the gauge rectangle and repositions when frame windows is resized """ size = self.GetSize() self.SetSize(size) gauge_pos, gauge_size = self.get_gauge_dimensions() self.gauge.SetSize(gauge_size) event.Skip() self.Update() def set_range(self, value): """ set_range(value) It sets the maximum value of gauge widget """ self.gauge.SetRange(value) def set_progress(self, value): """ set_progress(value) It sets the actual progress value to gauge widget """ self.gauge.SetValue(value) def set_status_text(self, value): """ set_status_text(value) It sets the text to the first field of StatusBar """ self.status_bar.SetStatusText(value) class Panel(wx.Panel): def __init__(self, parent): super(Panel, self).__init__(parent=parent) self.players = AutoWidthListCtrl(self) self.players.InsertColumn(0, 'code', wx.LIST_FORMAT_RIGHT, 50) self.players.InsertColumn(1, 'name', width=125) self.players.InsertColumn(2, 'fullname', width=175) players_box = wx.BoxSizer(wx.HORIZONTAL) players_box.Add(self.players, 1, wx.EXPAND) btn_sizer = wx.FlexGridSizer(rows=1, cols=2, hgap=5, vgap=5) self.btn_quit = wx.Button(self, wx.ID_CANCEL, label="Quit") self.btn_refresh = wx.Button(self, wx.ID_OK, label="Refresh") btn_sizer.Add(self.btn_quit, 0, wx.EXPAND) btn_sizer.Add(self.btn_refresh, 0, wx.EXPAND) sizer = wx.BoxSizer(wx.VERTICAL) sizer.Add(players_box, 1, wx.EXPAND | wx.ALL, 5) sizer.Add(btn_sizer, 0, wx.EXPAND | wx.ALL, 5) self.SetSizer(sizer) class ViewExtract(wx.Frame): def __init__(self, parent, title): self.parent = parent self.title = title super(ViewExtract, self).__init__(parent=self.parent, title=title) self.controller = self.parent.controller self.panel = PanelExtract(parent=self) self.SetSize((300, 150)) # bindings self.Bind(wx.EVT_CLOSE, self.on_quit) self.Bind(wx.EVT_BUTTON, self.on_quit, self.panel.btn_quit) self.Bind(wx.EVT_BUTTON, self.on_extract, self.panel.btn_extract) # noinspection PyUnusedLocal def on_extract(self, event): day = self.panel.day.GetValue() if day: if self.controller.are_evaluations_ready(day): self.controller.extract_evaluations(day) self.parent.Enable() self.Destroy() else: wx.MessageBox('Evaluations for day %s not ready!' % day, '', OK) else: wx.MessageBox('Please set a day to extract!', '', OK) # noinspection PyUnusedLocal def on_quit(self, event): self.parent.Enable() self.Destroy() class PanelExtract(wx.Panel): def __init__(self, parent): super(PanelExtract, self).__init__(parent) # Attributes self.day = wx.TextCtrl(self) # Layout text_sizer = wx.FlexGridSizer(rows=1, cols=2, hgap=5, vgap=
ViewPlayer(parent=self, title='Edit Player', is_editor=True) # noinspection PyUnusedLocal
random_line_split
core.py
, -1, style=wx.LC_REPORT) ListCtrlAutoWidthMixin.__init__(self) class Core(wx.Frame): def __init__(self, parent, controller, title): super(Core, self).__init__(parent=parent, title=title) self.parent = parent self.controller = controller self.child = None self.panel = Panel(self) self.panel.SetBackgroundColour('LightGray') self.status_bar = self.CreateStatusBar(2) self.status_bar.SetStatusWidths([200, -1]) gauge_pos, gauge_size = self.get_gauge_dimensions() self.gauge = wx.Gauge(self.status_bar, -1, 100, gauge_pos, gauge_size) # Menues self.menubar = wx.MenuBar() self.SetMenuBar(self.menubar) # Player menu menu_player = wx.Menu() self.menubar.Append(menu_player, "Players") self.menu_new_player = menu_player.Append(-1, "New Player", "New Player") menu_player.AppendSeparator() self.menu_players_import = menu_player.Append(-1, "import Players", "import Players") menu_player.AppendSeparator() self.menu_evaluations = menu_player.Append(-1, "extract evaluations", "extract evaluations") # Bindings # player bindings self.Bind(wx.EVT_SIZE, self.on_size) self.Bind(wx.EVT_MENU, self.on_new_player, self.menu_new_player) self.Bind(wx.EVT_MENU, self.on_import, self.menu_players_import) self.Bind(wx.EVT_MENU, self.on_extract, self.menu_evaluations) self.Bind(wx.EVT_BUTTON, self.quit, self.panel.btn_quit) self.Bind(wx.EVT_BUTTON, self.on_refresh, self.panel.btn_refresh) self.Bind(wx.EVT_LIST_ITEM_SELECTED, self.on_list, self.panel.players) self.Bind(wx.EVT_LIST_COL_CLICK, self.on_list_column, self.panel.players) # Players initialization players = self.controller.get_players() if players: self.fill_players(players) self.set_status_text('Players on database: %s' % len(players)) else: self.set_status_text('No Players on database') size = (450, 500) self.SetSize(size) self.Show() # noinspection PyUnusedLocal def quit(self, event): self.Destroy() # noinspection PyUnusedLocal def on_extract(self, event): self.child = ViewExtract(parent=self, title="extract evaluations") self.show_child() # noinspection PyUnusedLocal def on_import(self, event): choice = wx.MessageBox('Deleting All Players?', 'warning', YN) if choice == wx.YES: self.controller.delete_all_players() self.controller.import_players() wx.MessageBox('Players successfully imported!', '', OK) players = self.controller.get_players() self.fill_players(players) # noinspection PyUnusedLocal def edit_player(self, event): self.Disable() ViewPlayer(parent=self, title='Edit Player', is_editor=True) # noinspection PyUnusedLocal def on_new_player(self, event): self.child = ViewPlayer(parent=self, title='New Player') self.show_child() # noinspection PyUnusedLocal def on_quit(self, event): self.Destroy() # noinspection PyUnusedLocal def on_refresh(self, event): self.panel.players.DeleteAllItems() players = self.controller.get_players() self.fill_players(players) # noinspection PyUnusedLocal def on_list(self, event): item_id = event.m_itemIndex player_code = self.panel.players.GetItemText(item_id) player = self.controller.get_player_by_code(player_code) self.controller.set_temporary_object(player) item_name = self.panel.players.GetItem(item_id, 1) player_name = item_name.GetText() item_fullname = self.panel.players.GetItem(item_id, 2) player_fullname = item_fullname.GetText() view_edit = ViewPlayer(self, "Edit player", is_editor=True) view_edit.Show() view_edit.panel.code.SetValue(player_code) view_edit.panel.name.SetValue(player_name) view_edit.panel.fullname.SetValue(player_fullname) # Qui utilizzo ChangeValue per non scatenare EVT_TEXT, come invece # succederebbe usando SetValue. Questo EVT_TEXT, cercherebbe un # Handler al livello attuale (Frame) e non trovandolo, passerebbe # al Parent, che invece ha proprio un Bind di EVT_TEXT, proprio con # la textctrl 'ppl', la quale richiamerebbe la callback # 'on_text_entry, facendo casino. # view_edit.panel.btn_delete.Enable() view_edit.SetWindowStyle(wx.STAY_ON_TOP) def on_list_column(self, event): self.panel.players.DeleteAllItems() id_column = event.GetColumn() players = self.controller.get_sorted_players(id_column) self.fill_players(players) def fill_players(self, players): for player in players: index = self.panel.players.InsertStringItem(sys.maxint, str(player.code)) self.panel.players.SetStringItem(index, 1, str(player.name)) self.panel.players.SetStringItem(index, 2, str(player.fullname)) @staticmethod def show_message(string): wx.MessageBox(string, 'core info', wx.OK | wx.ICON_EXCLAMATION) def show_child(self): self.Disable() self.child.Centre() self.child.Show() def get_gauge_dimensions(self): """ get_gauge_dimensions(self) -> tuple_a, tuple_b tuple_a is a tuple with x position and y position of seconf field of the StatusBar """ pos_x, pos_y, dim_x, dim_y = self.status_bar.GetFieldRect(1) return (pos_x, pos_y), (dim_x, dim_y) def on_size(self, event): """ on_size() it redraws the gauge rectangle and repositions when frame windows is resized """ size = self.GetSize() self.SetSize(size) gauge_pos, gauge_size = self.get_gauge_dimensions() self.gauge.SetSize(gauge_size) event.Skip() self.Update() def set_range(self, value): """ set_range(value) It sets the maximum value of gauge widget """ self.gauge.SetRange(value) def set_progress(self, value): """ set_progress(value) It sets the actual progress value to gauge widget """ self.gauge.SetValue(value) def set_status_text(self, value): """ set_status_text(value) It sets the text to the first field of StatusBar """ self.status_bar.SetStatusText(value) class Panel(wx.Panel): def __init__(self, parent): super(Panel, self).__init__(parent=parent) self.players = AutoWidthListCtrl(self) self.players.InsertColumn(0, 'code', wx.LIST_FORMAT_RIGHT, 50) self.players.InsertColumn(1, 'name', width=125) self.players.InsertColumn(2, 'fullname', width=175) players_box = wx.BoxSizer(wx.HORIZONTAL) players_box.Add(self.players, 1, wx.EXPAND) btn_sizer = wx.FlexGridSizer(rows=1, cols=2, hgap=5, vgap=5) self.btn_quit = wx.Button(self, wx.ID_CANCEL, label="Quit") self.btn_refresh = wx.Button(self, wx.ID_OK, label="Refresh") btn_sizer.Add(self.btn_quit, 0, wx.EXPAND) btn_sizer.Add(self.btn_refresh, 0, wx.EXPAND) sizer = wx.BoxSizer(wx.VERTICAL) sizer.Add(players_box, 1, wx.EXPAND | wx.ALL, 5) sizer.Add(btn_sizer, 0, wx.EXPAND | wx.ALL, 5) self.SetSizer(sizer) class ViewExtract(wx.Frame): def __init__(self, parent, title):
# noinspection PyUnusedLocal def on_extract(self, event): day = self.panel.day.GetValue() if day: if self.controller.are_evaluations_ready(day): self.controller.extract_evaluations(day) self.parent.Enable() self.Destroy() else: wx.MessageBox('Evaluations for day %s not ready!' % day, '', OK) else: wx.MessageBox('Please set a day to extract!', '', OK) # noinspection PyUnusedLocal def on_quit(self, event): self.parent.Enable() self.Destroy() class PanelExtract(wx.Panel): def __init__(self, parent): super(PanelExtract, self).__init__(parent) # Attributes self.day = wx.TextCtrl(self) # Layout text_sizer = wx.FlexGridSizer(rows=1, cols=2, hgap=5, vgap
self.parent = parent self.title = title super(ViewExtract, self).__init__(parent=self.parent, title=title) self.controller = self.parent.controller self.panel = PanelExtract(parent=self) self.SetSize((300, 150)) # bindings self.Bind(wx.EVT_CLOSE, self.on_quit) self.Bind(wx.EVT_BUTTON, self.on_quit, self.panel.btn_quit) self.Bind(wx.EVT_BUTTON, self.on_extract, self.panel.btn_extract)
identifier_body
core.py
, -1, style=wx.LC_REPORT) ListCtrlAutoWidthMixin.__init__(self) class Core(wx.Frame): def __init__(self, parent, controller, title): super(Core, self).__init__(parent=parent, title=title) self.parent = parent self.controller = controller self.child = None self.panel = Panel(self) self.panel.SetBackgroundColour('LightGray') self.status_bar = self.CreateStatusBar(2) self.status_bar.SetStatusWidths([200, -1]) gauge_pos, gauge_size = self.get_gauge_dimensions() self.gauge = wx.Gauge(self.status_bar, -1, 100, gauge_pos, gauge_size) # Menues self.menubar = wx.MenuBar() self.SetMenuBar(self.menubar) # Player menu menu_player = wx.Menu() self.menubar.Append(menu_player, "Players") self.menu_new_player = menu_player.Append(-1, "New Player", "New Player") menu_player.AppendSeparator() self.menu_players_import = menu_player.Append(-1, "import Players", "import Players") menu_player.AppendSeparator() self.menu_evaluations = menu_player.Append(-1, "extract evaluations", "extract evaluations") # Bindings # player bindings self.Bind(wx.EVT_SIZE, self.on_size) self.Bind(wx.EVT_MENU, self.on_new_player, self.menu_new_player) self.Bind(wx.EVT_MENU, self.on_import, self.menu_players_import) self.Bind(wx.EVT_MENU, self.on_extract, self.menu_evaluations) self.Bind(wx.EVT_BUTTON, self.quit, self.panel.btn_quit) self.Bind(wx.EVT_BUTTON, self.on_refresh, self.panel.btn_refresh) self.Bind(wx.EVT_LIST_ITEM_SELECTED, self.on_list, self.panel.players) self.Bind(wx.EVT_LIST_COL_CLICK, self.on_list_column, self.panel.players) # Players initialization players = self.controller.get_players() if players: self.fill_players(players) self.set_status_text('Players on database: %s' % len(players)) else: self.set_status_text('No Players on database') size = (450, 500) self.SetSize(size) self.Show() # noinspection PyUnusedLocal def quit(self, event): self.Destroy() # noinspection PyUnusedLocal def on_extract(self, event): self.child = ViewExtract(parent=self, title="extract evaluations") self.show_child() # noinspection PyUnusedLocal def on_import(self, event): choice = wx.MessageBox('Deleting All Players?', 'warning', YN) if choice == wx.YES: self.controller.delete_all_players() self.controller.import_players() wx.MessageBox('Players successfully imported!', '', OK) players = self.controller.get_players() self.fill_players(players) # noinspection PyUnusedLocal def edit_player(self, event): self.Disable() ViewPlayer(parent=self, title='Edit Player', is_editor=True) # noinspection PyUnusedLocal def on_new_player(self, event): self.child = ViewPlayer(parent=self, title='New Player') self.show_child() # noinspection PyUnusedLocal def on_quit(self, event): self.Destroy() # noinspection PyUnusedLocal def on_refresh(self, event): self.panel.players.DeleteAllItems() players = self.controller.get_players() self.fill_players(players) # noinspection PyUnusedLocal def on_list(self, event): item_id = event.m_itemIndex player_code = self.panel.players.GetItemText(item_id) player = self.controller.get_player_by_code(player_code) self.controller.set_temporary_object(player) item_name = self.panel.players.GetItem(item_id, 1) player_name = item_name.GetText() item_fullname = self.panel.players.GetItem(item_id, 2) player_fullname = item_fullname.GetText() view_edit = ViewPlayer(self, "Edit player", is_editor=True) view_edit.Show() view_edit.panel.code.SetValue(player_code) view_edit.panel.name.SetValue(player_name) view_edit.panel.fullname.SetValue(player_fullname) # Qui utilizzo ChangeValue per non scatenare EVT_TEXT, come invece # succederebbe usando SetValue. Questo EVT_TEXT, cercherebbe un # Handler al livello attuale (Frame) e non trovandolo, passerebbe # al Parent, che invece ha proprio un Bind di EVT_TEXT, proprio con # la textctrl 'ppl', la quale richiamerebbe la callback # 'on_text_entry, facendo casino. # view_edit.panel.btn_delete.Enable() view_edit.SetWindowStyle(wx.STAY_ON_TOP) def on_list_column(self, event): self.panel.players.DeleteAllItems() id_column = event.GetColumn() players = self.controller.get_sorted_players(id_column) self.fill_players(players) def fill_players(self, players): for player in players: index = self.panel.players.InsertStringItem(sys.maxint, str(player.code)) self.panel.players.SetStringItem(index, 1, str(player.name)) self.panel.players.SetStringItem(index, 2, str(player.fullname)) @staticmethod def show_message(string): wx.MessageBox(string, 'core info', wx.OK | wx.ICON_EXCLAMATION) def show_child(self): self.Disable() self.child.Centre() self.child.Show() def get_gauge_dimensions(self): """ get_gauge_dimensions(self) -> tuple_a, tuple_b tuple_a is a tuple with x position and y position of seconf field of the StatusBar """ pos_x, pos_y, dim_x, dim_y = self.status_bar.GetFieldRect(1) return (pos_x, pos_y), (dim_x, dim_y) def on_size(self, event): """ on_size() it redraws the gauge rectangle and repositions when frame windows is resized """ size = self.GetSize() self.SetSize(size) gauge_pos, gauge_size = self.get_gauge_dimensions() self.gauge.SetSize(gauge_size) event.Skip() self.Update() def set_range(self, value): """ set_range(value) It sets the maximum value of gauge widget """ self.gauge.SetRange(value) def set_progress(self, value): """ set_progress(value) It sets the actual progress value to gauge widget """ self.gauge.SetValue(value) def set_status_text(self, value): """ set_status_text(value) It sets the text to the first field of StatusBar """ self.status_bar.SetStatusText(value) class Panel(wx.Panel): def __init__(self, parent): super(Panel, self).__init__(parent=parent) self.players = AutoWidthListCtrl(self) self.players.InsertColumn(0, 'code', wx.LIST_FORMAT_RIGHT, 50) self.players.InsertColumn(1, 'name', width=125) self.players.InsertColumn(2, 'fullname', width=175) players_box = wx.BoxSizer(wx.HORIZONTAL) players_box.Add(self.players, 1, wx.EXPAND) btn_sizer = wx.FlexGridSizer(rows=1, cols=2, hgap=5, vgap=5) self.btn_quit = wx.Button(self, wx.ID_CANCEL, label="Quit") self.btn_refresh = wx.Button(self, wx.ID_OK, label="Refresh") btn_sizer.Add(self.btn_quit, 0, wx.EXPAND) btn_sizer.Add(self.btn_refresh, 0, wx.EXPAND) sizer = wx.BoxSizer(wx.VERTICAL) sizer.Add(players_box, 1, wx.EXPAND | wx.ALL, 5) sizer.Add(btn_sizer, 0, wx.EXPAND | wx.ALL, 5) self.SetSizer(sizer) class ViewExtract(wx.Frame): def __init__(self, parent, title): self.parent = parent self.title = title super(ViewExtract, self).__init__(parent=self.parent, title=title) self.controller = self.parent.controller self.panel = PanelExtract(parent=self) self.SetSize((300, 150)) # bindings self.Bind(wx.EVT_CLOSE, self.on_quit) self.Bind(wx.EVT_BUTTON, self.on_quit, self.panel.btn_quit) self.Bind(wx.EVT_BUTTON, self.on_extract, self.panel.btn_extract) # noinspection PyUnusedLocal def on_extract(self, event): day = self.panel.day.GetValue() if day: if self.controller.are_evaluations_ready(day): self.controller.extract_evaluations(day) self.parent.Enable() self.Destroy() else: wx.MessageBox('Evaluations for day %s not ready!' % day, '', OK) else: wx.MessageBox('Please set a day to extract!', '', OK) # noinspection PyUnusedLocal def on_quit(self, event): self.parent.Enable() self.Destroy() class PanelExtract(wx.Panel): def
(self, parent): super(PanelExtract, self).__init__(parent) # Attributes self.day = wx.TextCtrl(self) # Layout text_sizer = wx.FlexGridSizer(rows=1, cols=2, hgap=5, vgap
__init__
identifier_name
index.js
Info(function(u) { that.__loadComdata(); that.__loadRecent(); that.__loadRecommend(); // that.__checkTwxx(); that.__checkToken(); setTimeout(function() { that.__checkCoupon(); }, 666); }); _my.getStorage({ key: "FOLLOW_SUBJECT", success: function(res) { let fs = (res.data || '').split(","); app.GLOBAL_DATA.FOLLOW_SUBJECT = fs; if (fs.length > 0) { that.__loadFollowSubject(fs); } } }); // 跳转页面 if (e.nextpage) { app.gotoPage(decodeURIComponent(e.nextpage)); return; } // 显示收藏提醒 if ( !( e.scene == 1089 || e.scene == 1001 || e.scene == 1022 || e.scene == 1023 || e.scene == 1027 ) ) { let showFav = function(t) { that.setData({ showFav: true, showFavClazz: "animated bounceIn slow" }); _my.setStorage({ key: "SHOW_FAV", data: t || 1 }); }; setTimeout(() => { _my.getStorage({ key: "SHOW_FAV", success: function(res) { if (res.data < 1) showFav(res.data + 1); }, fail: function(res) { showFav(1); } }); }, 1500); } }, __loadComdata: function(retry) { let that = this; zutils.get(app, "api/home/comdata", function(res) { if (res.data.error_code > 1000) { _my.redirectTo({ url: "/pages/index/tips?msg=" + res.data.error_msg }); return; } let _data = res.data.data; if (!_data) { if (retry) { _my.showToast({ icon: "none", duration: 4000, title: "请求失败,请稍后重试" }); } else { that.__loadComdata(true); } } if (!_data) return; _my.setNavigationBarTitle({ title: _data.title || "软考必备" }); // Banner if (!_data.banners || _data.banners.length == 0) { that.setData({ hideBanners: true }); } else { that.setData({ banners: _data.banners }); } // 红点 if (_data.reddot) { for (let k in _data.reddot) { app.showReddot(_data.reddot[k], k); } } that.setData({ icontext: _data.icontext || null, declaration: _data.declaration || null, openAis: _data.open_ais === true }); // 0, 1, 99 app.GLOBAL_DATA.RUN_MODE = _data.runMode || 0; }); }, onPullDownRefresh: function() { this.onLoad(); setTimeout(function() { _my.stopPullDownRefresh(); }, 800); }, onShow: function() { if (zutils.array.inAndErase(app.GLOBAL_DATA.RELOAD_EXAM, "Index")) { this.__loadRecent(); } if (zutils.array.inAndErase(app.GLOBAL_DATA.RELOAD_SUBJECT, "Index")) { this.__loadComdata(); this.__loadRecommend(); } let fs = app.GLOBAL_DATA.FOLLOW_SUBJECT; if (fs && fs.length > 0) { let lastFs = fs[fs.length - 1]; if (lastFs != this.__lastFs) { this.__loadFollowSubject(fs); } } else { this.setData({ followSubjects: null }); } }, // 解析分享(扫码进入) __checkTwxx: function() { return let q = app.enterSource.query.q; if (q && decodeURIComponent(q).indexOf("/t/wxx/") > -1) { zutils.get(app, "api/share/parse-twxx?q=" + q, function(res) { if (res.data.error_code == 0) { _my.navigateTo({ url: res.data.data }); } }); } }, // 解析分享口令 __checkToken: function() { if (this.__checkToken_OK == true) return; this.__checkToken_OK = true; // 清除口令 var rktk_token = false; setTimeout(function() { if (rktk_token == true) { _my.setClipboardData({ data: "", complete: () => _my.hideToast() }); } }, 1500); let that = this; _my.getClipboardData({ success: function(res) { if (res.data && res.data.substr(0, 6) == "#考题解析#") { // 扫码进入的优先级高于粘贴板 let scene = app.enterSource.scene; if ( scene == 1011 || scene == 1012 || scene == 1013 || scene == 1047 || scene == 1048 || scene == 1049 ) { console.log("扫码进入" + scene + ": " + res.data); rktk_token = true; return; } // 自己分享的 if (zutils.array.in(app.GLOBAL_DATA.KT_TOKENS, res.data)) { return; } rktk_token = true; zutils.get( app, "api/share/token-parse?text=" + encodeURIComponent(res.data), function(res2) { if (res2.data.error_code == 0) { let _data = res2.data.data; _my.showModal({ title: _data.title, confirmText: "立即查看", content: _data.content, success: function(res3) { if (res3.confirm) { _my.navigateTo({ url: _data.page }); } } }); } } ); } } }); }, // 最近关注题库 __loadFollowSubject: function(fs) { if (!fs || fs.length < 3) return; this.__lastFs = fs[fs.length - 1]; zutils.get(app, "api/home/subject-names?ids=" + fs.join(","), res => { if (res.data && res.data.data && res.data.data.length > 0) { let _subjects = res.data.data; _subjects.reverse(); this.__formatSubject(_subjects); this.setData({ followSubjects: _subjects }); } }); }, // 最近答题 __loadRecent: function() { zutils.get(app, "api/home/recent-exams", res => { this.setData(res.data.data); }); // 错题数量在此加载/刷新 zutils.get(app, "api/fav/incorrect-stats?d=1", res => { this.setData(res.data.data); }); }, // 推荐题库 __loadRecommend: function() { zutils.get(app, "api/home/recommend-subjects", res => { this.setData({ recommendSubjectsLoaded: true }); let _data = res.data.data; if (!_data) return; let _subjects = _data.recommend_subjects; this.__formatSubject(_subjects); _data = {}; _data.recommendSubjects = [ _subjects[0], _subjects[1], _subjects[2] ]; if (_subjects.length > 3) { _data.recommendSubjects2 = [ _subjects[3], _subjects[4], _subjects[5] ]; } this.s
_subjects[i][10] = sname.substr(0, 7); _subjects[i][11] = sname.substr(7); if (sname.indexOf("下午题") > -1) { _subjects[i][12] = "T2"; if (sname.indexOf("Ⅱ") > -1) { _subjects[i][12] = "T3"; } } if (_subjects[i][3] == 2) { _subjects[i][12] = "T4"; _subjects[i][10] = "知识点"; _subjects[i][11] = null; _subjects[i][2] = _subjects[i][1]; } } }, todayExam: function(e) { zutils.post( app, "api/exam/today-exam?formId=" + (e.detail.form
etData(_data); }); }, __formatSubject: function(_subjects) { for (let i = 0; i < _subjects.length; i++) { let sname = _subjects[i][1];
conditional_block
index.js
app.getUserInfo(function(u) { that.__loadComdata(); that.__loadRecent(); that.__loadRecommend(); // that.__checkTwxx(); that.__checkToken(); setTimeout(function() { that.__checkCoupon(); }, 666); }); _my.getStorage({ key: "FOLLOW_SUBJECT", success: function(res) { let fs = (res.data || '').split(","); app.GLOBAL_DATA.FOLLOW_SUBJECT = fs; if (fs.length > 0) { that.__loadFollowSubject(fs); } } }); // 跳转页面 if (e.nextpage) { app.gotoPage(decodeURIComponent(e.nextpage)); return; } // 显示收藏提醒 if ( !( e.scene == 1089 || e.scene == 1001 || e.scene == 1022 || e.scene == 1023 || e.scene == 1027 ) ) { let showFav = function(t) { that.setData({ showFav: true, showFavClazz: "animated bounceIn slow" }); _my.setStorage({ key: "SHOW_FAV", data: t || 1 }); }; setTimeout(() => { _my.getStorage({ key: "SHOW_FAV", success: function(res) { if (res.data < 1) showFav(res.data + 1); }, fail: function(res) { showFav(1); } }); }, 1500); } }, __loadComdata: function(retry) { let that = this; zutils.get(app, "api/home/comdata", function(res) { if (res.data.error_code > 1000) { _my.redirectTo({ url: "/pages/index/tips?msg=" + res.data.error_msg }); return; } let _data = res.data.data; if (!_data) { if (retry) { _my.showToast({ icon: "none", duration: 4000, title: "请求失败,请稍后重试" }); } else { that.__loadComdata(true); } } if (!_data) return; _my.setNavigationBarTitle({ title: _data.title || "软考必备" }); // Banner if (!_data.banners || _data.banners.length == 0) { that.setData({ hideBanners: true }); } else { that.setData({ banners: _data.banners
for (let k in _data.reddot) { app.showReddot(_data.reddot[k], k); } } that.setData({ icontext: _data.icontext || null, declaration: _data.declaration || null, openAis: _data.open_ais === true }); // 0, 1, 99 app.GLOBAL_DATA.RUN_MODE = _data.runMode || 0; }); }, onPullDownRefresh: function() { this.onLoad(); setTimeout(function() { _my.stopPullDownRefresh(); }, 800); }, onShow: function() { if (zutils.array.inAndErase(app.GLOBAL_DATA.RELOAD_EXAM, "Index")) { this.__loadRecent(); } if (zutils.array.inAndErase(app.GLOBAL_DATA.RELOAD_SUBJECT, "Index")) { this.__loadComdata(); this.__loadRecommend(); } let fs = app.GLOBAL_DATA.FOLLOW_SUBJECT; if (fs && fs.length > 0) { let lastFs = fs[fs.length - 1]; if (lastFs != this.__lastFs) { this.__loadFollowSubject(fs); } } else { this.setData({ followSubjects: null }); } }, // 解析分享(扫码进入) __checkTwxx: function() { return let q = app.enterSource.query.q; if (q && decodeURIComponent(q).indexOf("/t/wxx/") > -1) { zutils.get(app, "api/share/parse-twxx?q=" + q, function(res) { if (res.data.error_code == 0) { _my.navigateTo({ url: res.data.data }); } }); } }, // 解析分享口令 __checkToken: function() { if (this.__checkToken_OK == true) return; this.__checkToken_OK = true; // 清除口令 var rktk_token = false; setTimeout(function() { if (rktk_token == true) { _my.setClipboardData({ data: "", complete: () => _my.hideToast() }); } }, 1500); let that = this; _my.getClipboardData({ success: function(res) { if (res.data && res.data.substr(0, 6) == "#考题解析#") { // 扫码进入的优先级高于粘贴板 let scene = app.enterSource.scene; if ( scene == 1011 || scene == 1012 || scene == 1013 || scene == 1047 || scene == 1048 || scene == 1049 ) { console.log("扫码进入" + scene + ": " + res.data); rktk_token = true; return; } // 自己分享的 if (zutils.array.in(app.GLOBAL_DATA.KT_TOKENS, res.data)) { return; } rktk_token = true; zutils.get( app, "api/share/token-parse?text=" + encodeURIComponent(res.data), function(res2) { if (res2.data.error_code == 0) { let _data = res2.data.data; _my.showModal({ title: _data.title, confirmText: "立即查看", content: _data.content, success: function(res3) { if (res3.confirm) { _my.navigateTo({ url: _data.page }); } } }); } } ); } } }); }, // 最近关注题库 __loadFollowSubject: function(fs) { if (!fs || fs.length < 3) return; this.__lastFs = fs[fs.length - 1]; zutils.get(app, "api/home/subject-names?ids=" + fs.join(","), res => { if (res.data && res.data.data && res.data.data.length > 0) { let _subjects = res.data.data; _subjects.reverse(); this.__formatSubject(_subjects); this.setData({ followSubjects: _subjects }); } }); }, // 最近答题 __loadRecent: function() { zutils.get(app, "api/home/recent-exams", res => { this.setData(res.data.data); }); // 错题数量在此加载/刷新 zutils.get(app, "api/fav/incorrect-stats?d=1", res => { this.setData(res.data.data); }); }, // 推荐题库 __loadRecommend: function() { zutils.get(app, "api/home/recommend-subjects", res => { this.setData({ recommendSubjectsLoaded: true }); let _data = res.data.data; if (!_data) return; let _subjects = _data.recommend_subjects; this.__formatSubject(_subjects); _data = {}; _data.recommendSubjects = [ _subjects[0], _subjects[1], _subjects[2] ]; if (_subjects.length > 3) { _data.recommendSubjects2 = [ _subjects[3], _subjects[4], _subjects[5] ]; } this.setData(_data); }); }, __formatSubject: function(_subjects) { for (let i = 0; i < _subjects.length; i++) { let sname = _subjects[i][1]; _subjects[i][10] = sname.substr(0, 7); _subjects[i][11] = sname.substr(7); if (sname.indexOf("下午题") > -1) { _subjects[i][12] = "T2"; if (sname.indexOf("Ⅱ") > -1) { _subjects[i][12] = "T3"; } } if (_subjects[i][3] == 2) { _subjects[i][12] = "T4"; _subjects[i][10] = "知识点"; _subjects[i][11] = null; _subjects[i][2] = _subjects[i][1]; } } }, todayExam: function(e) { zutils.post( app, "api/exam/today-exam?formId=" + (e.detail.formId ||
}); } // 红点 if (_data.reddot) {
random_line_split
main.py
hello = data["credits"] hello1 = hello["crew"] for x in hello1: if x["job"] == "Director": directors.append(x["name"]) training_texts.append(str(directors)) #Goes through the movies that the user doesn't like and pulls the above mentioned parameters from TMDB to store in training_texts for x in range(len(negative_titles)): directors = [] movieName = negative_titles[x] httpRequest = "https://api.themoviedb.org/3/search/movie?include_adult=false&page=1&query="+movieName+"&language=en-US&api_key="+APIKey response = requests.get(httpRequest) data = response.json() training_texts.append(data["results"][0]["overview"]) training_texts.append(str(data["results"][0]["genre_ids"])) movie_id = data["results"][0]["id"] httpRequest2 = "https://api.themoviedb.org/3/movie/" + str(movie_id) + "?api_key=" + APIKey + "&append_to_response=credits" response = requests.get(httpRequest2) data = response.json() hello = data["credits"] hello1 = hello["crew"] for x in hello1: if x["job"] == "Director": directors.append(x["name"]) training_texts.append(str(directors)) #Goes through the movies that the user likes and pulls the above mentioned parameters from TDMB to store in tobetested_texts for x in range(len(tobetested)): directors = [] movieName = tobetested[x] httpRequest = "https://api.themoviedb.org/3/search/movie?include_adult=false&page=1&query="+movieName+"&language=en-US&api_key="+APIKey response = requests.get(httpRequest) data = response.json() tobetested_texts.append(data["results"][0]["overview"]) tobetested_texts.append(str(data["results"][0]["genre_ids"])) movie_id = data["results"][0]["id"] httpRequest2 = "https://api.themoviedb.org/3/movie/" + str(movie_id) + "?api_key=" + APIKey + "&append_to_response=credits" response = requests.get(httpRequest2) data = response.json() hello = data["credits"] hello1 = hello["crew"] for x in hello1: if x["job"] == "Director": directors.append(x["name"]) tobetested_texts.append(str(directors)) #Pulls the parameters for the current 20 top rated movies and puts them in mostpopular_texts directors = [] httpRequest = "https://api.themoviedb.org/3/movie/top_rated?api_key=e20e035943ec00333eb2a1d09ea93a5c&language=en-US&page=1" response = requests.get(httpRequest) data = response.json() bye = data["results"] for x in bye: if x["overview"] != "" or x["genre_ids"] != "": mostpopular_titles.append(x["title"]) mostpopular_texts.append(x["overview"]) mostpopular_texts.append(str(x["genre_ids"])) movie_id = x["id"] httpRequest2 = "https://api.themoviedb.org/3/movie/" + str(movie_id) + "?api_key=" + APIKey + "&append_to_response=credits" response = requests.get(httpRequest2) data = response.json() hello = data["credits"] hello1 = hello["crew"] for x in hello1: if x["job"] == "Director": directors.append(x["name"]) mostpopular_texts.append(str(directors)) #Here I prepare an equivalent set of labels, to tell the machine # that the first texts that came from movies the user liked are positive and the rest are negative. # When I feed these into the classifier, it'll use indices to match up # the texts and qualify what parameters are good and bad. training_labels = ["good"] * (3*len(positive_titles)) + ["bad"] * (3*len(negative_titles)) #The vectorizer is set up here: the first main component of machine learning vectorizer = CountVectorizer(stop_words='english') #Here I feed the data we have into the vectorizer so it can keep a # consistent mapping. vectorizer.fit(training_texts) # Here I transform all of the training texts into vector form. Basically makes it a list of numbers because code makes decisions quantitatively training_vectors = vectorizer.transform(training_texts) #I also convert the texts we are going to test and classify as good and bad into vector form test_texts = tobetested_texts test_populartexts = mostpopular_texts testing_vectors = vectorizer.transform(test_texts) testing_vectors_popular = vectorizer.transform(test_populartexts) #This is here the real machine learning happens as the code "connects the dots" between the training data and what is considered good and bad using the labels. classifier = tree.DecisionTreeClassifier() classifier.fit(training_vectors, training_labels) #Uses the connections the code made in previous steps to test each of the parameters of each movie the user wants to test and returns if the user will like/not like/probabaly like/probabaly not like based on the results. likeDict = { "will like" : "", "will probably like" : "", "will probably not like" : "", "will not like" : "" } print("Out of the movies you wanted to test:") print() for i, movie in enumerate(tobetested): listFormat = [tobetested_texts[i*3], tobetested_texts[i*3+1], tobetested_texts[i*3+2]] vectorFormat = vectorizer.transform(listFormat) result = classifier.predict(vectorFormat) if result[0] == 'good' and result [1] == 'good' and result[2] == 'good': likeDict['will like'] += (movie + ", ") elif result[0] == 'bad' and result [1] == 'bad' and result[2] == 'bad': likeDict['will not like'] += (movie + ", ") elif result[0] == 'good' and result [1] == 'good' and result[2] == 'bad': likeDict['will probably like'] += (movie + ", ") elif result[0] == 'bad' and result [1] == 'good' and result[2] == 'good': likeDict['will probably like'] += (movie + ", ") elif result[0] == 'good' and result [1] == 'bad' and result[2] == 'good': likeDict['will probably like'] += (movie + ", ") else: likeDict['will probably not like'] += (movie + ", ") for x in likeDict: if likeDict[x] != "": print("You", x, likeDict[x][0:-2]) print() print() print() #Uses the connections the code made in previous steps to test each of the parameters of each movies in the top 20 rated list and returns if the user will like/not like/probabaly like/probabaly not like based on the results. likeDict = { "will like" : "", "will probably like" : "", "will probably not like" : "", "will not like" : "" } print("Using your likes and dislikes, out of the top 20 top rated movies in the entire movie database:") print() for i, movie in enumerate(mostpopular_titles): listFormat = [mostpopular_texts[i*3], mostpopular_texts[i*3+1], mostpopular_texts[i*3+2]] vectorFormat = vectorizer.transform(listFormat) result = classifier.predict(vectorFormat) if result[0] == 'good' and result [1] == 'good' and result[2] == 'good': likeDict['will like'] += (movie + ", ") elif result[0] == 'bad' and result [1] == 'bad' and result[2] == 'bad': likeDict['will not like'] += (movie + ", ") elif result[0] == 'good' and result [1] == 'good' and result[2] == 'bad': likeDict['will probably like'] += (movie + ", ") elif result[0] == 'bad' and result [1] == 'good' and result[2] == 'good': likeDict['will like'] += (movie + ", ") elif result[0] == 'good' and result [1] == 'bad' and result[2] == 'good': likeDict['will probably like'] += (movie + ", ") else: likeDict['will probably not like'] += (movie + ", ") for x in likeDict: if likeDict[x] != "": print("You", x, likeDict[x][0:-2]) print()
#Looking at how the code makes its decisions visually is alot easier so I export the model to the tree.dot file. Upon copying all the data in tree.dot and pasting it in the textbox on http://www.webgraphviz.com/ you can see what the decision making process looks like. tree.export_graphviz(
random_line_split
main.py
= response.json() training_texts.append(data["results"][0]["overview"]) training_texts.append(str(data["results"][0]["genre_ids"])) movie_id = data["results"][0]["id"] httpRequest2 = "https://api.themoviedb.org/3/movie/" + str(movie_id) + "?api_key=" + APIKey + "&append_to_response=credits" response = requests.get(httpRequest2) data = response.json() hello = data["credits"] hello1 = hello["crew"] for x in hello1: if x["job"] == "Director": directors.append(x["name"]) training_texts.append(str(directors)) #Goes through the movies that the user doesn't like and pulls the above mentioned parameters from TMDB to store in training_texts for x in range(len(negative_titles)): directors = [] movieName = negative_titles[x] httpRequest = "https://api.themoviedb.org/3/search/movie?include_adult=false&page=1&query="+movieName+"&language=en-US&api_key="+APIKey response = requests.get(httpRequest) data = response.json() training_texts.append(data["results"][0]["overview"]) training_texts.append(str(data["results"][0]["genre_ids"])) movie_id = data["results"][0]["id"] httpRequest2 = "https://api.themoviedb.org/3/movie/" + str(movie_id) + "?api_key=" + APIKey + "&append_to_response=credits" response = requests.get(httpRequest2) data = response.json() hello = data["credits"] hello1 = hello["crew"] for x in hello1: if x["job"] == "Director": directors.append(x["name"]) training_texts.append(str(directors)) #Goes through the movies that the user likes and pulls the above mentioned parameters from TDMB to store in tobetested_texts for x in range(len(tobetested)): directors = [] movieName = tobetested[x] httpRequest = "https://api.themoviedb.org/3/search/movie?include_adult=false&page=1&query="+movieName+"&language=en-US&api_key="+APIKey response = requests.get(httpRequest) data = response.json() tobetested_texts.append(data["results"][0]["overview"]) tobetested_texts.append(str(data["results"][0]["genre_ids"])) movie_id = data["results"][0]["id"] httpRequest2 = "https://api.themoviedb.org/3/movie/" + str(movie_id) + "?api_key=" + APIKey + "&append_to_response=credits" response = requests.get(httpRequest2) data = response.json() hello = data["credits"] hello1 = hello["crew"] for x in hello1: if x["job"] == "Director": directors.append(x["name"]) tobetested_texts.append(str(directors)) #Pulls the parameters for the current 20 top rated movies and puts them in mostpopular_texts directors = [] httpRequest = "https://api.themoviedb.org/3/movie/top_rated?api_key=e20e035943ec00333eb2a1d09ea93a5c&language=en-US&page=1" response = requests.get(httpRequest) data = response.json() bye = data["results"] for x in bye: if x["overview"] != "" or x["genre_ids"] != "": mostpopular_titles.append(x["title"]) mostpopular_texts.append(x["overview"]) mostpopular_texts.append(str(x["genre_ids"])) movie_id = x["id"] httpRequest2 = "https://api.themoviedb.org/3/movie/" + str(movie_id) + "?api_key=" + APIKey + "&append_to_response=credits" response = requests.get(httpRequest2) data = response.json() hello = data["credits"] hello1 = hello["crew"] for x in hello1: if x["job"] == "Director": directors.append(x["name"]) mostpopular_texts.append(str(directors)) #Here I prepare an equivalent set of labels, to tell the machine # that the first texts that came from movies the user liked are positive and the rest are negative. # When I feed these into the classifier, it'll use indices to match up # the texts and qualify what parameters are good and bad. training_labels = ["good"] * (3*len(positive_titles)) + ["bad"] * (3*len(negative_titles)) #The vectorizer is set up here: the first main component of machine learning vectorizer = CountVectorizer(stop_words='english') #Here I feed the data we have into the vectorizer so it can keep a # consistent mapping. vectorizer.fit(training_texts) # Here I transform all of the training texts into vector form. Basically makes it a list of numbers because code makes decisions quantitatively training_vectors = vectorizer.transform(training_texts) #I also convert the texts we are going to test and classify as good and bad into vector form test_texts = tobetested_texts test_populartexts = mostpopular_texts testing_vectors = vectorizer.transform(test_texts) testing_vectors_popular = vectorizer.transform(test_populartexts) #This is here the real machine learning happens as the code "connects the dots" between the training data and what is considered good and bad using the labels. classifier = tree.DecisionTreeClassifier() classifier.fit(training_vectors, training_labels) #Uses the connections the code made in previous steps to test each of the parameters of each movie the user wants to test and returns if the user will like/not like/probabaly like/probabaly not like based on the results. likeDict = { "will like" : "", "will probably like" : "", "will probably not like" : "", "will not like" : "" } print("Out of the movies you wanted to test:") print() for i, movie in enumerate(tobetested): listFormat = [tobetested_texts[i*3], tobetested_texts[i*3+1], tobetested_texts[i*3+2]] vectorFormat = vectorizer.transform(listFormat) result = classifier.predict(vectorFormat) if result[0] == 'good' and result [1] == 'good' and result[2] == 'good': likeDict['will like'] += (movie + ", ") elif result[0] == 'bad' and result [1] == 'bad' and result[2] == 'bad': likeDict['will not like'] += (movie + ", ") elif result[0] == 'good' and result [1] == 'good' and result[2] == 'bad': likeDict['will probably like'] += (movie + ", ") elif result[0] == 'bad' and result [1] == 'good' and result[2] == 'good': likeDict['will probably like'] += (movie + ", ") elif result[0] == 'good' and result [1] == 'bad' and result[2] == 'good': likeDict['will probably like'] += (movie + ", ") else: likeDict['will probably not like'] += (movie + ", ") for x in likeDict: if likeDict[x] != "": print("You", x, likeDict[x][0:-2]) print() print() print() #Uses the connections the code made in previous steps to test each of the parameters of each movies in the top 20 rated list and returns if the user will like/not like/probabaly like/probabaly not like based on the results. likeDict = { "will like" : "", "will probably like" : "", "will probably not like" : "", "will not like" : "" } print("Using your likes and dislikes, out of the top 20 top rated movies in the entire movie database:") print() for i, movie in enumerate(mostpopular_titles): listFormat = [mostpopular_texts[i*3], mostpopular_texts[i*3+1], mostpopular_texts[i*3+2]] vectorFormat = vectorizer.transform(listFormat) result = classifier.predict(vectorFormat) if result[0] == 'good' and result [1] == 'good' and result[2] == 'good': likeDict['will like'] += (movie + ", ") elif result[0] == 'bad' and result [1] == 'bad' and result[2] == 'bad': likeDict['will not like'] += (movie + ", ") elif result[0] == 'good' and result [1] == 'good' and result[2] == 'bad': likeDict['will probably like'] += (movie + ", ") elif result[0] == 'bad' and result [1] == 'good' and result[2] == 'good': likeDict['will like'] += (movie + ", ") elif result[0] == 'good' and result [1] == 'bad' and result[2] == 'good': likeDict['will probably like'] += (movie + ", ") else:
likeDict['will probably not like'] += (movie + ", ")
conditional_block
data.py
, app, backend, moduleRefs, locations, modules, version, checkout, silent ): """Collects TF data according to specifications. The specifications are passed as arguments when the object is initialized. Parameters ---------- backend: string `github` or `gitlab` or a GitLab instance such as `gitlab.huc.knaw.nl`. app: obj The high-level API object moduleRefs: tuple Each member consists of a module ref, which is a tuple of information that defines a module. locations: string|tuple One or more directory paths. They will be combined with the `modules` argument and used as locations to search for TF data files. modules: string|tuple One or more directory path segments. They will be appended to the paths given by the `locations` argument to form search locations for TF data files. version: string The version of TF data that should be retrievend. Version is a directory level just below the search locations. checkout: string A specifier to use a specific release or commit of a data repository. silent: string, optional tf.core.timestamp.SILENT_D See `tf.core.timestamp.Timestamp` """ self.backend = backend self.app = app self.moduleRefs = ( [] if moduleRefs is None else moduleRefs.split(",") if type(moduleRefs) is str else list(moduleRefs) ) self.locationsArg = locations self.modulesArg = modules self.version = version self.checkout = checkout self.silent = silent def getMain(self): """Get the main data of the corpus. This is specified by the `org`, `repo` and `relative` settings under `provenanceSpec` in `config.yaml`. See Also -------- tf.advanced.settings: options allowed in `config.yaml` """ app = self.app checkout = self.checkout aContext = app.context org = aContext.org repo = aContext.repo relative = prefixSlash(aContext.relative) appPath = aContext.appPath appName = aContext.appName if appName.startswith("app:"): appParent = appPath.rsplit("/", 1)[0] relative = f"{appParent}{relative}" elif org is None or repo is None: appPathRep = f"{appPath}/" if appPath else "" relative = f"{appPathRep}{appName}" self.checkout = "local" if not self.getModule(org, repo, prefixSlash(relative), checkout, isBase=True): self.good = False def getStandard(self): """Get the data of the standard modules specified by the settings of the corpus. These are specified in the `moduleSpecs` setting under `provenanceSpecs` in `config.yaml`. They will be loaded *after* the extra modules specified in the **mod** parameter, and only in as far they have not been specifief in the **mod** parameter. In this way you can pass overriding checkout specifiers to the standard modules. See Also -------- tf.advanced.settings: options allowed in `config.yaml` """ app = self.app loadData = app.loadData if not loadData or loadData == "core": return aContext = app.context moduleSpecs = aContext.moduleSpecs seen = self.seen checkout = self.checkout backend = self.backend for m in moduleSpecs or []: org = m["org"] repo = m["repo"] relative = m["relative"] theCheckout = m.get("checkout", checkout) theBackend = m.get("backend", backend) bRep = backendRep(theBackend, "spec", default=backend) ref = f"{bRep}{org}/{repo}{relative}" if ref in seen: continue if not self.getModule( org, repo, relative, theCheckout, backend=theBackend, specs=m, ): self.good = False def getRefs(self): """Get data from additional modules. These are specified in the `moduleRefs` parameter of `AppData`. We store the set of special modules in order to skip them later when we are loading the standard modules. """ backend = self.backend refs = self.moduleRefs for ref in refs: refPure = ref.rsplit(":", 1)[0] if refPure in self.seen: continue parts = splitModRef(ref) if not parts: self.good = False continue parts[2] = prefixSlash(normpath(parts[2])) # the relative bit theBackend = ( None if parts[-1] is None or parts[-1] == backend else parts[-1] ) if not self.getModule(*parts[0:-1], backend=theBackend): self.good = False def getModules(self): """Get data from additional local directories. These are specified in the `locations` and `modules` parameters of `AppData`. """ self.provenance = []
provenance = self.provenance self.mLocations = [] mLocations = self.mLocations self.locations = None self.modules = None self.good = True self.seen = set() self.getMain() self.getRefs() self.getStandard() version = self.version good = self.good app = self.app if good: app.mLocations = mLocations app.provenance = provenance else: return mModules = [] if mLocations: mModules.append(version or "") locations = self.locationsArg modules = self.modulesArg givenLocations = ( [] if locations is None else [expandDir(app, x.strip()) for x in itemize(locations, "\n")] if type(locations) is str else [str(x) for x in locations] ) givenModules = ( [] if modules is None else [normpath(x.strip()) for x in itemize(modules, "\n")] if type(modules) is str else [normpath(str(x)) for x in modules] ) self.locations = mLocations + givenLocations self.modules = mModules + givenModules def getModule( self, org, repo, relative, checkout, backend=None, isBase=False, specs=None ): """Prepare to load a single module. Eventually, all TF data will be downloaded from local directories, bases on a list of location paths and module paths. This function computes the contribution of a single module to both the location paths and the module paths. Parameters ---------- org: string GitHub organization or GitLab group of the module repo: string: GitHub repository or GitLab project of the module relative: string Path within the repository of the module checkout: string A specifier to use a specific release or commit of a data repository. backend: string The backend if different from the backend of the main module isBase: boolean, optional False Whether this module is the main data of the corpus. specs: dict, optional False Additional informational attributes of the module, e.g. a DOI """ backend = self.backend if backend is None else backendRep(backend, "norm") bRep = backendRep(backend, "spec", default=self.backend) version = self.version silent = self.silent mLocations = self.mLocations provenance = self.provenance seen = self.seen app = self.app _browse = app._browse aContext = app.context branch = aContext.provenanceSpec["branch"] relative = prefixSlash(normpath(relative)) moduleRef = f"{bRep}{org}/{repo}{relative}" if moduleRef in self.seen: return True if org is None or repo is None: relativeBare = relative.removeprefix("/") repoLocation = relativeBare mLocations.append(relativeBare) (commit, local, release) = (None, None, None) else: (commit, release, local, localBase, localDir) = checkoutRepo( backend, _browse=_browse, org=org, repo=repo, folder=relative, version=version, checkout=checkout, withPaths=False, keep=False, silent=silent, ) if not localBase: return False repoLocation = f"{localBase}/{org}/{repo}" mLocations.append(f"{localBase}/{localDir}") seen.add(moduleRef) if isBase: app.repoLocation = repoLocation info = {} for item in ( ("doi", None), ("corpus", f"{org}/{repo}{relative}"), ): (key, default) = item info[key] = ( getattr(aContext, key) if isBase else specs[key] if specs and key in specs else default ) provenance.append( ( ("corpus", info["corpus"]), ("version
random_line_split
data.py
, app, backend, moduleRefs, locations, modules, version, checkout, silent ): """Collects TF data according to specifications. The specifications are passed as arguments when the object is initialized. Parameters ---------- backend: string `github` or `gitlab` or a GitLab instance such as `gitlab.huc.knaw.nl`. app: obj The high-level API object moduleRefs: tuple Each member consists of a module ref, which is a tuple of information that defines a module. locations: string|tuple One or more directory paths. They will be combined with the `modules` argument and used as locations to search for TF data files. modules: string|tuple One or more directory path segments. They will be appended to the paths given by the `locations` argument to form search locations for TF data files. version: string The version of TF data that should be retrievend. Version is a directory level just below the search locations. checkout: string A specifier to use a specific release or commit of a data repository. silent: string, optional tf.core.timestamp.SILENT_D See `tf.core.timestamp.Timestamp` """ self.backend = backend self.app = app self.moduleRefs = ( [] if moduleRefs is None else moduleRefs.split(",") if type(moduleRefs) is str else list(moduleRefs) ) self.locationsArg = locations self.modulesArg = modules self.version = version self.checkout = checkout self.silent = silent def getMain(self): """Get the main data of the corpus. This is specified by the `org`, `repo` and `relative` settings under `provenanceSpec` in `config.yaml`. See Also -------- tf.advanced.settings: options allowed in `config.yaml` """ app = self.app checkout = self.checkout aContext = app.context org = aContext.org repo = aContext.repo relative = prefixSlash(aContext.relative) appPath = aContext.appPath appName = aContext.appName if appName.startswith("app:"): appParent = appPath.rsplit("/", 1)[0] relative = f"{appParent}{relative}" elif org is None or repo is None: appPathRep = f"{appPath}/" if appPath else "" relative = f"{appPathRep}{appName}" self.checkout = "local" if not self.getModule(org, repo, prefixSlash(relative), checkout, isBase=True): self.good = False def getStandard(self): """Get the data of the standard modules specified by the settings of the corpus. These are specified in the `moduleSpecs` setting under `provenanceSpecs` in `config.yaml`. They will be loaded *after* the extra modules specified in the **mod** parameter, and only in as far they have not been specifief in the **mod** parameter. In this way you can pass overriding checkout specifiers to the standard modules. See Also -------- tf.advanced.settings: options allowed in `config.yaml` """ app = self.app loadData = app.loadData if not loadData or loadData == "core": return aContext = app.context moduleSpecs = aContext.moduleSpecs seen = self.seen checkout = self.checkout backend = self.backend for m in moduleSpecs or []: org = m["org"] repo = m["repo"] relative = m["relative"] theCheckout = m.get("checkout", checkout) theBackend = m.get("backend", backend) bRep = backendRep(theBackend, "spec", default=backend) ref = f"{bRep}{org}/{repo}{relative}" if ref in seen: continue if not self.getModule( org, repo, relative, theCheckout, backend=theBackend, specs=m, ): self.good = False def getRefs(self): """Get data from additional modules. These are specified in the `moduleRefs` parameter of `AppData`. We store the set of special modules in order to skip them later when we are loading the standard modules. """ backend = self.backend refs = self.moduleRefs for ref in refs: refPure = ref.rsplit(":", 1)[0] if refPure in self.seen: continue parts = splitModRef(ref) if not parts: self.good = False continue parts[2] = prefixSlash(normpath(parts[2])) # the relative bit theBackend = ( None if parts[-1] is None or parts[-1] == backend else parts[-1] ) if not self.getModule(*parts[0:-1], backend=theBackend): self.good = False def getModules(self): """Get data from additional local directories. These are specified in the `locations` and `modules` parameters of `AppData`. """ self.provenance = [] provenance = self.provenance self.mLocations = [] mLocations = self.mLocations self.locations = None self.modules = None self.good = True self.seen = set() self.getMain() self.getRefs() self.getStandard() version = self.version good = self.good app = self.app if good: app.mLocations = mLocations app.provenance = provenance else: return mModules = [] if mLocations: mModules.append(version or "") locations = self.locationsArg modules = self.modulesArg givenLocations = ( [] if locations is None else [expandDir(app, x.strip()) for x in itemize(locations, "\n")] if type(locations) is str else [str(x) for x in locations] ) givenModules = ( [] if modules is None else [normpath(x.strip()) for x in itemize(modules, "\n")] if type(modules) is str else [normpath(str(x)) for x in modules] ) self.locations = mLocations + givenLocations self.modules = mModules + givenModules def getModule( self, org, repo, relative, checkout, backend=None, isBase=False, specs=None ): """Prepare to load a single module. Eventually, all TF data will be downloaded from local directories, bases on a list of location paths and module paths. This function computes the contribution of a single module to both the location paths and the module paths. Parameters ---------- org: string GitHub organization or GitLab group of the module repo: string: GitHub repository or GitLab project of the module relative: string Path within the repository of the module checkout: string A specifier to use a specific release or commit of a data repository. backend: string The backend if different from the backend of the main module isBase: boolean, optional False Whether this module is the main data of the corpus. specs: dict, optional False Additional informational attributes of the module, e.g. a DOI """ backend = self.backend if backend is None else backendRep(backend, "norm") bRep = backendRep(backend, "spec", default=self.backend) version = self.version silent = self.silent mLocations = self.mLocations provenance = self.provenance seen = self.seen app = self.app _browse = app._browse aContext = app.context branch = aContext.provenanceSpec["branch"] relative = prefixSlash(normpath(relative)) moduleRef = f"{bRep}{org}/{repo}{relative}" if moduleRef in self.seen: return True if org is None or repo is None: relativeBare = relative.removeprefix("/") repoLocation = relativeBare mLocations.append(relativeBare) (commit, local, release) = (None, None, None) else: (commit, release, local, localBase, localDir) = checkoutRepo( backend, _browse=_browse, org=org, repo=repo, folder=relative, version=version, checkout=checkout, withPaths=False, keep=False, silent=silent, ) if not localBase: return False repoLocation = f"{localBase}/{org}/{repo}" mLocations.append(f"{localBase}/{localDir}") seen.add(moduleRef) if isBase:
info = {} for item in ( ("doi", None), ("corpus", f"{org}/{repo}{relative}"), ): (key, default) = item info[key] = ( getattr(aContext, key) if isBase else specs[key] if specs and key in specs else default ) provenance.append( ( ("corpus", info["corpus"]), ("version
app.repoLocation = repoLocation
conditional_block
data.py
: def __init__( self, app, backend, moduleRefs, locations, modules, version, checkout, silent ): """Collects TF data according to specifications. The specifications are passed as arguments when the object is initialized. Parameters ---------- backend: string `github` or `gitlab` or a GitLab instance such as `gitlab.huc.knaw.nl`. app: obj The high-level API object moduleRefs: tuple Each member consists of a module ref, which is a tuple of information that defines a module. locations: string|tuple One or more directory paths. They will be combined with the `modules` argument and used as locations to search for TF data files. modules: string|tuple One or more directory path segments. They will be appended to the paths given by the `locations` argument to form search locations for TF data files. version: string The version of TF data that should be retrievend. Version is a directory level just below the search locations. checkout: string A specifier to use a specific release or commit of a data repository. silent: string, optional tf.core.timestamp.SILENT_D See `tf.core.timestamp.Timestamp` """ self.backend = backend self.app = app self.moduleRefs = ( [] if moduleRefs is None else moduleRefs.split(",") if type(moduleRefs) is str else list(moduleRefs) ) self.locationsArg = locations self.modulesArg = modules self.version = version self.checkout = checkout self.silent = silent def getMain(self): """Get the main data of the corpus. This is specified by the `org`, `repo` and `relative` settings under `provenanceSpec` in `config.yaml`. See Also -------- tf.advanced.settings: options allowed in `config.yaml` """ app = self.app checkout = self.checkout aContext = app.context org = aContext.org repo = aContext.repo relative = prefixSlash(aContext.relative) appPath = aContext.appPath appName = aContext.appName if appName.startswith("app:"): appParent = appPath.rsplit("/", 1)[0] relative = f"{appParent}{relative}" elif org is None or repo is None: appPathRep = f"{appPath}/" if appPath else "" relative = f"{appPathRep}{appName}" self.checkout = "local" if not self.getModule(org, repo, prefixSlash(relative), checkout, isBase=True): self.good = False def getStandard(self): """Get the data of the standard modules specified by the settings of the corpus. These are specified in the `moduleSpecs` setting under `provenanceSpecs` in `config.yaml`. They will be loaded *after* the extra modules specified in the **mod** parameter, and only in as far they have not been specifief in the **mod** parameter. In this way you can pass overriding checkout specifiers to the standard modules. See Also -------- tf.advanced.settings: options allowed in `config.yaml` """ app = self.app loadData = app.loadData if not loadData or loadData == "core": return aContext = app.context moduleSpecs = aContext.moduleSpecs seen = self.seen checkout = self.checkout backend = self.backend for m in moduleSpecs or []: org = m["org"] repo = m["repo"] relative = m["relative"] theCheckout = m.get("checkout", checkout) theBackend = m.get("backend", backend) bRep = backendRep(theBackend, "spec", default=backend) ref = f"{bRep}{org}/{repo}{relative}" if ref in seen: continue if not self.getModule( org, repo, relative, theCheckout, backend=theBackend, specs=m, ): self.good = False def getRefs(self): """Get data from additional modules. These are specified in the `moduleRefs` parameter of `AppData`. We store the set of special modules in order to skip them later when we are loading the standard modules. """ backend = self.backend refs = self.moduleRefs for ref in refs: refPure = ref.rsplit(":", 1)[0] if refPure in self.seen: continue parts = splitModRef(ref) if not parts: self.good = False continue parts[2] = prefixSlash(normpath(parts[2])) # the relative bit theBackend = ( None if parts[-1] is None or parts[-1] == backend else parts[-1] ) if not self.getModule(*parts[0:-1], backend=theBackend): self.good = False def getModules(self): """Get data from additional local directories. These are specified in the `locations` and `modules` parameters of `AppData`. """ self.provenance = [] provenance = self.provenance self.mLocations = [] mLocations = self.mLocations self.locations = None self.modules = None self.good = True self.seen = set() self.getMain() self.getRefs() self.getStandard() version = self.version good = self.good app = self.app if good: app.mLocations = mLocations app.provenance = provenance else: return mModules = [] if mLocations: mModules.append(version or "") locations = self.locationsArg modules = self.modulesArg givenLocations = ( [] if locations is None else [expandDir(app, x.strip()) for x in itemize(locations, "\n")] if type(locations) is str else [str(x) for x in locations] ) givenModules = ( [] if modules is None else [normpath(x.strip()) for x in itemize(modules, "\n")] if type(modules) is str else [normpath(str(x)) for x in modules] ) self.locations = mLocations + givenLocations self.modules = mModules + givenModules def getModule( self, org, repo, relative, checkout, backend=None, isBase=False, specs=None ): """Prepare to load a single module. Eventually, all TF data will be downloaded from local directories, bases on a list of location paths and module paths. This function computes the contribution of a single module to both the location paths and the module paths. Parameters ---------- org: string GitHub organization or GitLab group of the module repo: string: GitHub repository or GitLab project of the module relative: string Path within the repository of the module checkout: string A specifier to use a specific release or commit of a data repository. backend: string The backend if different from the backend of the main module isBase: boolean, optional False Whether this module is the main data of the corpus. specs: dict, optional False Additional informational attributes of the module, e.g. a DOI """ backend = self.backend if backend is None else backendRep(backend, "norm") bRep = backendRep(backend, "spec", default=self.backend) version = self.version silent = self.silent mLocations = self.mLocations provenance = self.provenance seen = self.seen app = self.app _browse = app._browse aContext = app.context branch = aContext.provenanceSpec["branch"] relative = prefixSlash(normpath(relative)) moduleRef = f"{bRep}{org}/{repo}{relative}" if moduleRef in self.seen: return True if org is None or repo is None: relativeBare = relative.removeprefix("/") repoLocation = relativeBare mLocations.append(relativeBare) (commit, local, release) = (None, None, None) else: (commit, release, local, localBase, localDir) = checkoutRepo( backend, _browse=_browse, org=org, repo=repo, folder=relative, version=version, checkout=checkout, withPaths=False, keep=False, silent=silent, ) if not localBase: return False repoLocation = f"{localBase}/{org}/{repo}" mLocations.append(f"{localBase}/{localDir}") seen.add(moduleRef) if isBase: app.repoLocation = repoLocation info = {} for item in ( ("doi", None), ("corpus", f"{org}/{repo}{relative}"), ): (key, default) = item info[key] = ( getattr(aContext, key) if isBase else specs[key] if specs and key in specs else default ) provenance.append( ( ("cor
AppData
identifier_name
data.py
One or more directory path segments. They will be appended to the paths given by the `locations` argument to form search locations for TF data files. version: string The version of TF data that should be retrievend. Version is a directory level just below the search locations. checkout: string A specifier to use a specific release or commit of a data repository. silent: string, optional tf.core.timestamp.SILENT_D See `tf.core.timestamp.Timestamp` """ self.backend = backend self.app = app self.moduleRefs = ( [] if moduleRefs is None else moduleRefs.split(",") if type(moduleRefs) is str else list(moduleRefs) ) self.locationsArg = locations self.modulesArg = modules self.version = version self.checkout = checkout self.silent = silent def getMain(self): """Get the main data of the corpus. This is specified by the `org`, `repo` and `relative` settings under `provenanceSpec` in `config.yaml`. See Also -------- tf.advanced.settings: options allowed in `config.yaml` """ app = self.app checkout = self.checkout aContext = app.context org = aContext.org repo = aContext.repo relative = prefixSlash(aContext.relative) appPath = aContext.appPath appName = aContext.appName if appName.startswith("app:"): appParent = appPath.rsplit("/", 1)[0] relative = f"{appParent}{relative}" elif org is None or repo is None: appPathRep = f"{appPath}/" if appPath else "" relative = f"{appPathRep}{appName}" self.checkout = "local" if not self.getModule(org, repo, prefixSlash(relative), checkout, isBase=True): self.good = False def getStandard(self): """Get the data of the standard modules specified by the settings of the corpus. These are specified in the `moduleSpecs` setting under `provenanceSpecs` in `config.yaml`. They will be loaded *after* the extra modules specified in the **mod** parameter, and only in as far they have not been specifief in the **mod** parameter. In this way you can pass overriding checkout specifiers to the standard modules. See Also -------- tf.advanced.settings: options allowed in `config.yaml` """ app = self.app loadData = app.loadData if not loadData or loadData == "core": return aContext = app.context moduleSpecs = aContext.moduleSpecs seen = self.seen checkout = self.checkout backend = self.backend for m in moduleSpecs or []: org = m["org"] repo = m["repo"] relative = m["relative"] theCheckout = m.get("checkout", checkout) theBackend = m.get("backend", backend) bRep = backendRep(theBackend, "spec", default=backend) ref = f"{bRep}{org}/{repo}{relative}" if ref in seen: continue if not self.getModule( org, repo, relative, theCheckout, backend=theBackend, specs=m, ): self.good = False def getRefs(self): """Get data from additional modules. These are specified in the `moduleRefs` parameter of `AppData`. We store the set of special modules in order to skip them later when we are loading the standard modules. """ backend = self.backend refs = self.moduleRefs for ref in refs: refPure = ref.rsplit(":", 1)[0] if refPure in self.seen: continue parts = splitModRef(ref) if not parts: self.good = False continue parts[2] = prefixSlash(normpath(parts[2])) # the relative bit theBackend = ( None if parts[-1] is None or parts[-1] == backend else parts[-1] ) if not self.getModule(*parts[0:-1], backend=theBackend): self.good = False def getModules(self): """Get data from additional local directories. These are specified in the `locations` and `modules` parameters of `AppData`. """ self.provenance = [] provenance = self.provenance self.mLocations = [] mLocations = self.mLocations self.locations = None self.modules = None self.good = True self.seen = set() self.getMain() self.getRefs() self.getStandard() version = self.version good = self.good app = self.app if good: app.mLocations = mLocations app.provenance = provenance else: return mModules = [] if mLocations: mModules.append(version or "") locations = self.locationsArg modules = self.modulesArg givenLocations = ( [] if locations is None else [expandDir(app, x.strip()) for x in itemize(locations, "\n")] if type(locations) is str else [str(x) for x in locations] ) givenModules = ( [] if modules is None else [normpath(x.strip()) for x in itemize(modules, "\n")] if type(modules) is str else [normpath(str(x)) for x in modules] ) self.locations = mLocations + givenLocations self.modules = mModules + givenModules def getModule( self, org, repo, relative, checkout, backend=None, isBase=False, specs=None ): """Prepare to load a single module. Eventually, all TF data will be downloaded from local directories, bases on a list of location paths and module paths. This function computes the contribution of a single module to both the location paths and the module paths. Parameters ---------- org: string GitHub organization or GitLab group of the module repo: string: GitHub repository or GitLab project of the module relative: string Path within the repository of the module checkout: string A specifier to use a specific release or commit of a data repository. backend: string The backend if different from the backend of the main module isBase: boolean, optional False Whether this module is the main data of the corpus. specs: dict, optional False Additional informational attributes of the module, e.g. a DOI """ backend = self.backend if backend is None else backendRep(backend, "norm") bRep = backendRep(backend, "spec", default=self.backend) version = self.version silent = self.silent mLocations = self.mLocations provenance = self.provenance seen = self.seen app = self.app _browse = app._browse aContext = app.context branch = aContext.provenanceSpec["branch"] relative = prefixSlash(normpath(relative)) moduleRef = f"{bRep}{org}/{repo}{relative}" if moduleRef in self.seen: return True if org is None or repo is None: relativeBare = relative.removeprefix("/") repoLocation = relativeBare mLocations.append(relativeBare) (commit, local, release) = (None, None, None) else: (commit, release, local, localBase, localDir) = checkoutRepo( backend, _browse=_browse, org=org, repo=repo, folder=relative, version=version, checkout=checkout, withPaths=False, keep=False, silent=silent, ) if not localBase: return False repoLocation = f"{localBase}/{org}/{repo}" mLocations.append(f"{localBase}/{localDir}") seen.add(moduleRef) if isBase: app.repoLocation = repoLocation info = {} for item in ( ("doi", None), ("corpus", f"{org}/{repo}{relative}"), ): (key, default) = item info[key] = ( getattr(aContext, key) if isBase else specs[key] if specs and key in specs else default ) provenance.append( ( ("corpus", info["cor
def __init__( self, app, backend, moduleRefs, locations, modules, version, checkout, silent ): """Collects TF data according to specifications. The specifications are passed as arguments when the object is initialized. Parameters ---------- backend: string `github` or `gitlab` or a GitLab instance such as `gitlab.huc.knaw.nl`. app: obj The high-level API object moduleRefs: tuple Each member consists of a module ref, which is a tuple of information that defines a module. locations: string|tuple One or more directory paths. They will be combined with the `modules` argument and used as locations to search for TF data files. modules: string|tuple
identifier_body
preprocess_03.py
print('Check: ' + str(length) + ' < ' + str(total_time*fs1//(2*q))) raise ValueError('Length FFT_side != length: ' + str(len(FFT_side)) + ' != ' + str(length)) FFT_log = [] # normalize FFT for value in FFT_side: value = np.log(value) FFT_log.append(value) max_val = getMax(FFT_log)[1] FFT_norm = [] for value in FFT_log: FFT_norm.append(value/max_val) FFT_side = np.array(FFT_norm) FFT_divided = FFT_side[range(length//divide)] #plot = True if plot == True: freqs = scipy.fftpack.fftfreq(sig3.size, 1/fs2) freqs_divided = np.array(freqs[range(len(FFT_divided))]) plt.plot(freqs_divided,FFT_divided) # plotting the complete fft spectrum plt.show() return FFT_divided def processMPCC(filename,subsample=2048): #assume 8000Hz #amplify high frequencies #Setup try: fs, signal = wavfile.read(filename) # File assumed to be in the same directory except ValueError or UnboundLocalError: print(filename + ' failed to process.') print('Failed Read') print() return 'failed' half = len(signal)//2 side = subsample//2 signal = signal[half-side:half+side] if side != len(signal)//2: print(filename + ' failed to process.') print('N too small, N: ' + str(len(signal)) + ', subsample: ' + str(subsample)) print() return 'failed' sig = signal[:,0] #get first channel #Pre-Emphasis pre_emphasis = 0.97 e_sig = sig[1:] - pre_emphasis * sig[0:-1] #emphasized signal sig_len = len(e_sig) #Framing fr_size = 0.025 # frame size (sec) fr_overlap = 0.01 # frame stride, frame overlap (sec) fr_len = int(round(fr_size * fs)) # frame length (sec/sec) fr_step = int(round(fr_overlap * fs)) # amt to step frame each time num_fr = int(np.ceil(np.abs(sig_len - fr_len) / fr_step)) #Number of Frames padding = num_fr * fr_step + fr_len # Amount of padding between frames z = [0 for _ in range(padding-sig_len)] z = np.array(z) pad_sig = np.append(e_sig, z) # Pad Signal so frames equal size #idx = np.tile(np.linspace(0, fr_len,fr_len), (num_fr, 1)) + np.tile(np.linspace(0, num_fr * fr_step, fr_step * num_fr), (fr_len, 1)).T #fr = pad_sig[idx] idx = np.tile(np.arange(0, fr_len), (num_fr, 1)) + np.transpose(np.tile(np.arange(0, num_fr * fr_step, fr_step), (fr_len, 1))) fr = pad_sig[idx.astype(np.int32)] #Window NFFT = 512 fr = fr * ( 0.54 - 0.46 * np.cos((2 * np.pi * NFFT) / (fr_len - 1)) ) # Hamming Window #Fourier-Transform and Power Spectrum #NFFT = NFFT mag_fr = np.absolute(np.fft.rfft(fr, NFFT)) # Magnitude of the FFT pow_fr = (1.0 / NFFT) * ((mag_fr) ** 2) # Power Spectrum #Filter Banks nfilt = 40 f_low = 0 f_high = (2595 * np.log10(1 + (fs / 2) / 700)) # Convert Hz to Mel mel_points = np.linspace(f_low, f_high, nfilt + 2) # Equally spaced in Mel scale hz_points = (700 * (10**(mel_points / 2595) - 1)) # Convert Mel to Hz b = np.floor((NFFT + 1) * hz_points / fs) #bin fbank = np.zeros((nfilt, int(np.floor(NFFT / 2 + 1)))) for i in range(1, nfilt + 1): f_m_minus = int(b[i - 1]) # left f_m = int(b[i]) # center f_m_plus = int(b[i + 1]) # right for j in range(f_m_minus, f_m): fbank[i - 1, j] = (j - b[i - 1]) / (b[i] - b[i - 1]) for j in range(f_m, f_m_plus): fbank[i - 1, j] = (b[i + 1] - j) / (b[i + 1] - b[i]) fb = np.dot(pow_fr, np.transpose(fbank)) # filter banks fb = np.where(fb == 0, np.finfo(float).eps, fb) # Numerical Stability fb = 20 * np.log10(fb) # convert to dB #Mel-frequency Cepstral Coefficients (MFCCs) num_ceps = 12 mfcc = dct(fb, type=2, axis=1, norm='ortho')[:, 1 : (num_ceps + 1)] # Keep 2-13 #Sinusoidal Filtering c_lift = 22 # dim of MFCC vector (n_fr, n_coeff) = mfcc.shape #number of frames number of coeff ncoeff_array = np.arange(n_coeff) lift = 1 + (c_lift / 2) * np.sin(np.pi * ncoeff_array / c_lift) mfcc = mfcc * lift #Mean Normalization epsilon = 1e-8 for i in range(len(fb)): fb[i] -= mean(fb) + epsilon for i in range(len(mfcc)): mfcc[i] -= mean(mfcc) + epsilon output = [] for i in range(len(mfcc)): for j in range(len(mfcc[0])): output.append(mfcc[i][j]) m = getMax(output)[1] for i,value in enumerate(output): output[i] = value/m return np.array(output) def mean(array_list): """Returns the mean of an array or list""" count = 0.0 for value in array_list: count += value return count/len(array_list) def downsample(sig,fs,q): """ sig (list,array): sound/data signal q (int): downsample factor """ N = len(sig)//q new_sig = [] for i in range(len(sig)//q): new_sig.append(sig[i*q]) new_sig = np.array(new_sig) return (fs//q,new_sig) class Preprocess:
# self.X is dictionary of dir:[input_nodes1,input_nodes2] # self.Y is dictionary of dir:[output_nodes1,output_nodes2] # self.Y corresponds to self.X self.X = [] # list of input vectors self.Y = [] # list of output vectors #if process == False: #self.loadData(data_file) #else: #process == True: #self.processData(data_file,directory,comment) def getXY(self): """Returns X (List of Input Vectors), and Y (List of Output Vectors) for preprocessed data ex) X = [[0,0],[0,1],[1,0],[1,1]] ex) Y = [[0],[1
def __init__(self): """data_file (string): contains the file to load or store data, ex)data.txt process (bool): if False, load data from data_file, if True, process data in directory & store in data_file directory (string): (optional) directory of data to be processed """ # Ex) self.output['cel']: np.array([1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]) self.output = {} # directory names are names of instruments #self.dirs = # ['cel', 'cla', 'flu', 'gac', 'gel', 'org', 'pia', 'sax', 'tru', 'vio', 'voi'] self.dirs = [] # list of names of subdirectories in directory # example: self.files['sax'] = # IRMAS-TrainingData\sax\006__[sax][nod][cla]1686__1.wav self.files = {} # dictionary of dir:[file,file,file] # self.data = {} # dictionary of dir:[input_nodes,input_nodes]
identifier_body
preprocess_03.py
print('Check: ' + str(length) + ' < ' + str(total_time*fs1//(2*q))) raise ValueError('Length FFT_side != length: ' + str(len(FFT_side)) + ' != ' + str(length)) FFT_log = [] # normalize FFT for value in FFT_side: value = np.log(value) FFT_log.append(value) max_val = getMax(FFT_log)[1] FFT_norm = [] for value in FFT_log: FFT_norm.append(value/max_val) FFT_side = np.array(FFT_norm) FFT_divided = FFT_side[range(length//divide)] #plot = True if plot == True: freqs = scipy.fftpack.fftfreq(sig3.size, 1/fs2) freqs_divided = np.array(freqs[range(len(FFT_divided))]) plt.plot(freqs_divided,FFT_divided) # plotting the complete fft spectrum plt.show() return FFT_divided def processMPCC(filename,subsample=2048): #assume 8000Hz #amplify high frequencies #Setup try: fs, signal = wavfile.read(filename) # File assumed to be in the same directory except ValueError or UnboundLocalError: print(filename + ' failed to process.') print('Failed Read') print() return 'failed' half = len(signal)//2 side = subsample//2 signal = signal[half-side:half+side] if side != len(signal)//2: print(filename + ' failed to process.') print('N too small, N: ' + str(len(signal)) + ', subsample: ' + str(subsample)) print() return 'failed' sig = signal[:,0] #get first channel #Pre-Emphasis pre_emphasis = 0.97 e_sig = sig[1:] - pre_emphasis * sig[0:-1] #emphasized signal sig_len = len(e_sig) #Framing fr_size = 0.025 # frame size (sec) fr_overlap = 0.01 # frame stride, frame overlap (sec) fr_len = int(round(fr_size * fs)) # frame length (sec/sec) fr_step = int(round(fr_overlap * fs)) # amt to step frame each time num_fr = int(np.ceil(np.abs(sig_len - fr_len) / fr_step)) #Number of Frames padding = num_fr * fr_step + fr_len # Amount of padding between frames z = [0 for _ in range(padding-sig_len)] z = np.array(z) pad_sig = np.append(e_sig, z) # Pad Signal so frames equal size #idx = np.tile(np.linspace(0, fr_len,fr_len), (num_fr, 1)) + np.tile(np.linspace(0, num_fr * fr_step, fr_step * num_fr), (fr_len, 1)).T #fr = pad_sig[idx] idx = np.tile(np.arange(0, fr_len), (num_fr, 1)) + np.transpose(np.tile(np.arange(0, num_fr * fr_step, fr_step), (fr_len, 1))) fr = pad_sig[idx.astype(np.int32)] #Window NFFT = 512 fr = fr * ( 0.54 - 0.46 * np.cos((2 * np.pi * NFFT) / (fr_len - 1)) ) # Hamming Window #Fourier-Transform and Power Spectrum #NFFT = NFFT mag_fr = np.absolute(np.fft.rfft(fr, NFFT)) # Magnitude of the FFT pow_fr = (1.0 / NFFT) * ((mag_fr) ** 2) # Power Spectrum #Filter Banks nfilt = 40 f_low = 0 f_high = (2595 * np.log10(1 + (fs / 2) / 700)) # Convert Hz to Mel mel_points = np.linspace(f_low, f_high, nfilt + 2) # Equally spaced in Mel scale hz_points = (700 * (10**(mel_points / 2595) - 1)) # Convert Mel to Hz b = np.floor((NFFT + 1) * hz_points / fs) #bin fbank = np.zeros((nfilt, int(np.floor(NFFT / 2 + 1)))) for i in range(1, nfilt + 1): f_m_minus = int(b[i - 1]) # left f_m = int(b[i]) # center f_m_plus = int(b[i + 1]) # right for j in range(f_m_minus, f_m): fbank[i - 1, j] = (j - b[i - 1]) / (b[i] - b[i - 1]) for j in range(f_m, f_m_plus): fbank[i - 1, j] = (b[i + 1] - j) / (b[i + 1] - b[i]) fb = np.dot(pow_fr, np.transpose(fbank)) # filter banks fb = np.where(fb == 0, np.finfo(float).eps, fb) # Numerical Stability fb = 20 * np.log10(fb) # convert to dB #Mel-frequency Cepstral Coefficients (MFCCs) num_ceps = 12 mfcc = dct(fb, type=2, axis=1, norm='ortho')[:, 1 : (num_ceps + 1)] # Keep 2-13 #Sinusoidal Filtering c_lift = 22 # dim of MFCC vector (n_fr, n_coeff) = mfcc.shape #number of frames number of coeff ncoeff_array = np.arange(n_coeff) lift = 1 + (c_lift / 2) * np.sin(np.pi * ncoeff_array / c_lift) mfcc = mfcc * lift #Mean Normalization epsilon = 1e-8 for i in range(len(fb)): fb[i] -= mean(fb) + epsilon for i in range(len(mfcc)): mfcc[i] -= mean(mfcc) + epsilon output = []
m = getMax(output)[1] for i,value in enumerate(output): output[i] = value/m return np.array(output) def mean(array_list): """Returns the mean of an array or list""" count = 0.0 for value in array_list: count += value return count/len(array_list) def downsample(sig,fs,q): """ sig (list,array): sound/data signal q (int): downsample factor """ N = len(sig)//q new_sig = [] for i in range(len(sig)//q): new_sig.append(sig[i*q]) new_sig = np.array(new_sig) return (fs//q,new_sig) class Preprocess: def __init__(self): """data_file (string): contains the file to load or store data, ex)data.txt process (bool): if False, load data from data_file, if True, process data in directory & store in data_file directory (string): (optional) directory of data to be processed """ # Ex) self.output['cel']: np.array([1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]) self.output = {} # directory names are names of instruments #self.dirs = # ['cel', 'cla', 'flu', 'gac', 'gel', 'org', 'pia', 'sax', 'tru', 'vio', 'voi'] self.dirs = [] # list of names of subdirectories in directory # example: self.files['sax'] = # IRMAS-TrainingData\sax\006__[sax][nod][cla]1686__1.wav self.files = {} # dictionary of dir:[file,file,file] # self.data = {} # dictionary of dir:[input_nodes,input_nodes] # self.X is dictionary of dir:[input_nodes1,input_nodes2] # self.Y is dictionary of dir:[output_nodes1,output_nodes2] # self.Y corresponds to self.X self.X = [] # list of input vectors self.Y = [] # list of output vectors #if process == False: #self.loadData(data_file) #else: #process == True: #self.processData(data_file,directory,comment) def getXY(self): """Returns X (List of Input Vectors), and Y (List of Output Vectors) for preprocessed data ex) X = [[0,0],[0,1],[1,0],[1,1]] ex) Y = [[0],[1
for i in range(len(mfcc)): for j in range(len(mfcc[0])): output.append(mfcc[i][j])
random_line_split
preprocess_03.py
]) plt.plot(freqs_divided,FFT_divided) # plotting the complete fft spectrum plt.show() return FFT_divided def processMPCC(filename,subsample=2048): #assume 8000Hz #amplify high frequencies #Setup try: fs, signal = wavfile.read(filename) # File assumed to be in the same directory except ValueError or UnboundLocalError: print(filename + ' failed to process.') print('Failed Read') print() return 'failed' half = len(signal)//2 side = subsample//2 signal = signal[half-side:half+side] if side != len(signal)//2: print(filename + ' failed to process.') print('N too small, N: ' + str(len(signal)) + ', subsample: ' + str(subsample)) print() return 'failed' sig = signal[:,0] #get first channel #Pre-Emphasis pre_emphasis = 0.97 e_sig = sig[1:] - pre_emphasis * sig[0:-1] #emphasized signal sig_len = len(e_sig) #Framing fr_size = 0.025 # frame size (sec) fr_overlap = 0.01 # frame stride, frame overlap (sec) fr_len = int(round(fr_size * fs)) # frame length (sec/sec) fr_step = int(round(fr_overlap * fs)) # amt to step frame each time num_fr = int(np.ceil(np.abs(sig_len - fr_len) / fr_step)) #Number of Frames padding = num_fr * fr_step + fr_len # Amount of padding between frames z = [0 for _ in range(padding-sig_len)] z = np.array(z) pad_sig = np.append(e_sig, z) # Pad Signal so frames equal size #idx = np.tile(np.linspace(0, fr_len,fr_len), (num_fr, 1)) + np.tile(np.linspace(0, num_fr * fr_step, fr_step * num_fr), (fr_len, 1)).T #fr = pad_sig[idx] idx = np.tile(np.arange(0, fr_len), (num_fr, 1)) + np.transpose(np.tile(np.arange(0, num_fr * fr_step, fr_step), (fr_len, 1))) fr = pad_sig[idx.astype(np.int32)] #Window NFFT = 512 fr = fr * ( 0.54 - 0.46 * np.cos((2 * np.pi * NFFT) / (fr_len - 1)) ) # Hamming Window #Fourier-Transform and Power Spectrum #NFFT = NFFT mag_fr = np.absolute(np.fft.rfft(fr, NFFT)) # Magnitude of the FFT pow_fr = (1.0 / NFFT) * ((mag_fr) ** 2) # Power Spectrum #Filter Banks nfilt = 40 f_low = 0 f_high = (2595 * np.log10(1 + (fs / 2) / 700)) # Convert Hz to Mel mel_points = np.linspace(f_low, f_high, nfilt + 2) # Equally spaced in Mel scale hz_points = (700 * (10**(mel_points / 2595) - 1)) # Convert Mel to Hz b = np.floor((NFFT + 1) * hz_points / fs) #bin fbank = np.zeros((nfilt, int(np.floor(NFFT / 2 + 1)))) for i in range(1, nfilt + 1): f_m_minus = int(b[i - 1]) # left f_m = int(b[i]) # center f_m_plus = int(b[i + 1]) # right for j in range(f_m_minus, f_m): fbank[i - 1, j] = (j - b[i - 1]) / (b[i] - b[i - 1]) for j in range(f_m, f_m_plus): fbank[i - 1, j] = (b[i + 1] - j) / (b[i + 1] - b[i]) fb = np.dot(pow_fr, np.transpose(fbank)) # filter banks fb = np.where(fb == 0, np.finfo(float).eps, fb) # Numerical Stability fb = 20 * np.log10(fb) # convert to dB #Mel-frequency Cepstral Coefficients (MFCCs) num_ceps = 12 mfcc = dct(fb, type=2, axis=1, norm='ortho')[:, 1 : (num_ceps + 1)] # Keep 2-13 #Sinusoidal Filtering c_lift = 22 # dim of MFCC vector (n_fr, n_coeff) = mfcc.shape #number of frames number of coeff ncoeff_array = np.arange(n_coeff) lift = 1 + (c_lift / 2) * np.sin(np.pi * ncoeff_array / c_lift) mfcc = mfcc * lift #Mean Normalization epsilon = 1e-8 for i in range(len(fb)): fb[i] -= mean(fb) + epsilon for i in range(len(mfcc)): mfcc[i] -= mean(mfcc) + epsilon output = [] for i in range(len(mfcc)): for j in range(len(mfcc[0])): output.append(mfcc[i][j]) m = getMax(output)[1] for i,value in enumerate(output): output[i] = value/m return np.array(output) def mean(array_list): """Returns the mean of an array or list""" count = 0.0 for value in array_list: count += value return count/len(array_list) def downsample(sig,fs,q): """ sig (list,array): sound/data signal q (int): downsample factor """ N = len(sig)//q new_sig = [] for i in range(len(sig)//q): new_sig.append(sig[i*q]) new_sig = np.array(new_sig) return (fs//q,new_sig) class Preprocess: def __init__(self): """data_file (string): contains the file to load or store data, ex)data.txt process (bool): if False, load data from data_file, if True, process data in directory & store in data_file directory (string): (optional) directory of data to be processed """ # Ex) self.output['cel']: np.array([1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]) self.output = {} # directory names are names of instruments #self.dirs = # ['cel', 'cla', 'flu', 'gac', 'gel', 'org', 'pia', 'sax', 'tru', 'vio', 'voi'] self.dirs = [] # list of names of subdirectories in directory # example: self.files['sax'] = # IRMAS-TrainingData\sax\006__[sax][nod][cla]1686__1.wav self.files = {} # dictionary of dir:[file,file,file] # self.data = {} # dictionary of dir:[input_nodes,input_nodes] # self.X is dictionary of dir:[input_nodes1,input_nodes2] # self.Y is dictionary of dir:[output_nodes1,output_nodes2] # self.Y corresponds to self.X self.X = [] # list of input vectors self.Y = [] # list of output vectors #if process == False: #self.loadData(data_file) #else: #process == True: #self.processData(data_file,directory,comment) def getXY(self): """Returns X (List of Input Vectors), and Y (List of Output Vectors) for preprocessed data ex) X = [[0,0],[0,1],[1,0],[1,1]] ex) Y = [[0],[1],[1],[0]] """ return (self.X,self.Y) def getInputLength(self): """Returns length of Input Layer""" return len(self.X[0]) def getOutputLength(self): """Returns length of Output Layer""" return len(self.Y[0]) def getFileList(self): """Returns a dictionary with key:value 'Output Name':[file list] ex) {'sax':['sax1.wav','sax2.wav','sax3.wav']} """ return self.files def getOutputVectors(self): """ Returns a dictionary with key:value 'OutputName':output vector Ex) output['cel']: np.array([1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]) """ return self.output def
getOutputNames
identifier_name
preprocess_03.py
): contains the file to load or store data, ex)data.txt process (bool): if False, load data from data_file, if True, process data in directory & store in data_file directory (string): (optional) directory of data to be processed """ # Ex) self.output['cel']: np.array([1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]) self.output = {} # directory names are names of instruments #self.dirs = # ['cel', 'cla', 'flu', 'gac', 'gel', 'org', 'pia', 'sax', 'tru', 'vio', 'voi'] self.dirs = [] # list of names of subdirectories in directory # example: self.files['sax'] = # IRMAS-TrainingData\sax\006__[sax][nod][cla]1686__1.wav self.files = {} # dictionary of dir:[file,file,file] # self.data = {} # dictionary of dir:[input_nodes,input_nodes] # self.X is dictionary of dir:[input_nodes1,input_nodes2] # self.Y is dictionary of dir:[output_nodes1,output_nodes2] # self.Y corresponds to self.X self.X = [] # list of input vectors self.Y = [] # list of output vectors #if process == False: #self.loadData(data_file) #else: #process == True: #self.processData(data_file,directory,comment) def getXY(self): """Returns X (List of Input Vectors), and Y (List of Output Vectors) for preprocessed data ex) X = [[0,0],[0,1],[1,0],[1,1]] ex) Y = [[0],[1],[1],[0]] """ return (self.X,self.Y) def getInputLength(self): """Returns length of Input Layer""" return len(self.X[0]) def getOutputLength(self): """Returns length of Output Layer""" return len(self.Y[0]) def getFileList(self): """Returns a dictionary with key:value 'Output Name':[file list] ex) {'sax':['sax1.wav','sax2.wav','sax3.wav']} """ return self.files def getOutputVectors(self): """ Returns a dictionary with key:value 'OutputName':output vector Ex) output['cel']: np.array([1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]) """ return self.output def getOutputNames(self): """Returns a list of the names of the output vectors ex) ['cel', 'cla', 'flu', 'gac', 'gel', 'org', 'pia', 'sax', 'tru', 'vio', 'voi'] """ return self.dirs def loadData(self,data_file): """Loads the data in data_file into Trainer""" #Load the data from the json with open(data_file) as json_file: data = json.load(json_file) # Clear all instance variables self.dirs = [] self.files = {} self.X = [] self.Y = [] self.output = {} # stored the data into the instance variables self.dirs = data['dirs'] #good self.files = data['files'] # good # self.output is a dict() with string:np.array output = data['output'] for e in output: self.output[e] = np.array(output[e]) # -> fine #self.X is a list of np.arrays X = data['X'] for x in X: self.X.append(np.array(x))# -> fine #self.Y is a list of np.arrays Y = data['Y'] for y in Y: self.Y.append(list(y))# -> fine #Test prints, uncomment to test if data looks correct #print('self.dirs = ' + str(self.dirs)) #print() #print('self.files = ' + str(self.files)) #print() #print('self.output = ' + str(self.output)) #print() #print('self.X = ' + str(self.X)) #print() #print('self.Y = ' + str(self.Y)) #print() print('Preprocessed data loaded from ' + str(data_file)) print(data['comment']) return def processData(self,data_file,directory,comment = '',way='mpcc',opt=[1024]): """Processes the data in directory and stores it in data_file directory (string): folder of data to be processed data_file (string): name of file for data to be stored ex) data.txt comment (string): optional message to be stored with data way = 'fft', opts is a list containing length (int): Number of datapoints of one-sided fft (must be even,preferably a power of 2) q (int): Downsampling Rate (must be even, preferably power of 2) fs_in (int): throw ValueError if fs of filename != fs_i divide (int): 1/divide*Nsamples is taken from FFT (preferably even) plot (bool): ( plots the one sided FFT if True, otherwise does not plot Note: length < total_time*fs/(q) Ex) length = 1024 < (0.25sec)*(44100Hz)/(4) = 2756 way = 'mpcc', opts is a list containing subsample (int) = Number of subsamples to take from audio file. """ self.dirs = [name for name in os.listdir(directory) if os.path.isdir(os.path.join(directory, name))] # directory names are names of instruments #self.dirs = # ['cel', 'cla', 'flu', 'gac', 'gel', 'org', 'pia', 'sax', 'tru', 'vio', 'voi'] self.dirs = [name for name in os.listdir(directory) if os.path.isdir(os.path.join(directory, name))] # example: self.files['sax'] = # IRMAS-TrainingData\sax\006__[sax][nod][cla]1686__1.wav self.files = {} for d in self.dirs: self.files[d] = [] sub_dir = os.path.join(directory, d) for filename in glob.glob(os.path.join(sub_dir, '*.wav')): self.files[d].append(filename) # Ex) self.output['cel']: np.array([1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]) i = 0 for name in self.dirs: temp = [] for j in range(len(self.dirs)): if i == j: temp.append(1) else: temp.append(0) self.output[name] = np.array(temp) i +=1 #self.X = [] # list of input vectors #self.Y = [] # list of output vectors t0 = time.time() for name in self.dirs: t1 = time.time() for file in self.files[name]: #input_vector = processFile(file,length=length1,q=q1,fs_in=fs_in1,divide=divide1,plot = False) if way == 'mpcc': input_vector = processMPCC(file,*opt) elif way == 'fft': input_vector = processFFT(file,*opt) else: raise ValueError('Invalid Way, valid types include: \'mpcc\' or \'fft\'') if input_vector != 'failed': self.X.append(input_vector) self.Y.append(self.output[name]) print('Time take to process '+str(name)+ ': ' + str((time.time()-t1)/60)[0:4] + ' min.') print('Total Processing Time: ' + str((time.time()-t0)/60)[0:4] + ' min.') # Now we can store all of the data in a json # Need to store self.X, self.Y, self.dirs,self.output,self.files,self.data # self.dirs is a list of strings -> fine # self.files is a dict() with string:string -> fine # self.output is a dict() with string:np.array output = {} for d in self.output: out_list = [] for value in self.output[d]: out_list.append(int(value)) output[d] = out_list # -> fine #self.X is a list of np.arrays X = [] for i in range(len(self.X)): x = [] for ele in self.X[i]: x.append(float(ele)) X.append(x) # -> fine #self.Y is a list of np.arrays Y = [] for i in range(len(self.Y)):
y = [] for ele in self.Y[i]: y.append(float(ele)) Y.append(y) # -> fine
conditional_block
lib.rs
VertexHandle}; /// The three basic elements in a polygon mesh. #[derive(Debug, Clone, Copy, PartialEq, Eq)] pub enum
{ Edge, Face, Vertex, } // =========================================================================== // ===== `Sealed` trait // =========================================================================== pub(crate) mod sealed { /// A trait that cannot be implemented outside of this crate. /// /// This is helpful for all "real" traits in this library that only /// abstract over a closed set of types. Thus, users shouldn't be able to /// implement those traits for their types. Adding `Sealed` as supertrait /// solves this problem. pub trait Sealed {} } // =========================================================================== // ===== Macros // =========================================================================== /// Derive macro for the [`Empty` trait][traits::Empty]. /// /// ``` /// use lox::Empty; // this imports the custom-derive and not the trait! /// /// #[derive(Empty)] /// struct MyStruct { /// a: Vec<u32>, // => `vec![]` /// b: Option<String>, // => `None` /// c: (), // => `()` /// } /// ``` /// /// This can only be derived for structs. All struct fields need to implement /// `Empty` in order for the derive to work. If your struct has generic /// parameters, they won't be bounded with `Empty` in the generated impl block. /// This is useful most of the time, because things like `Vec<T>` and /// `Option<T>` don't require `T: Empty` to implement `Empty`. But this means /// that you sometimes have to add a global `Empty` bound to your parameter or /// implement `Empty` manually. pub use lox_macros::Empty; /// Derive macro for [the `MemSink` trait][io::MemSink]. /// /// You can easily derive `MemSink` for your own types. To do that, you have to /// attach `#[derive(MemSink)]` to your struct definition (note: currently, the /// trait can only be derived for structs with named fields). You also have to /// annotate your fields with `#[lox(...)]` attributes to tell the derive macro /// what a field should be used for. Example: /// /// ``` /// use lox::{ /// MemSink, VertexHandle, /// cgmath::Point3, /// ds::HalfEdgeMesh, /// map::DenseMap, /// }; /// /// /// #[derive(MemSink)] /// struct MyMesh { /// #[lox(core_mesh)] /// mesh: HalfEdgeMesh, /// /// #[lox(vertex_position)] /// positions: DenseMap<VertexHandle, Point3<f32>>, /// } /// ``` /// /// There is one required field: the core mesh field. That field's type has to /// implement several mesh traits, in particular `MeshMut` and `TriMeshMut`. /// You have to annotate that mesh with `#[lox(core_mesh)]`. /// /// Additionally, you can have fields for each mesh property, like vertex /// position or face colors. The type of those fields has to implement /// `PropStoreMut` with a compatible element type. You have to annotate these /// property fields with the corresponding attribute. The available properties /// are: /// /// - `vertex_position` /// - `vertex_normal` /// - `vertex_color` /// - `face_normal` /// - `face_color` /// /// Furthermore, there are some configurations (like the cast mode) that can be /// configured via `lox(...)` attributes as well. See below for more /// information. /// /// /// ## Cast modes /// /// You can set a *cast mode* for each field. A `MemSink` has to be able to /// "handle" any primitive type as the source is allowed to call the property /// methods with any type. The sink can handle types either by casting or by /// returning an error. The field's cast mode determines which casts are /// allowed and which are not. Possible cast modes: /// /// - `cast = "none"` /// - `cast = "lossless"` /// - `cast = "rounding"` /// - `cast = "clamping"` /// - `cast = "lossy"` (*default*) /// /// The `none` mode does not allow casting at all. If the type provided by the /// source does not match the type in your struct, an error is returned. All /// other modes correspond to the cast modes in the [`cast` /// module][crate::cast]. /// /// Note that the cast modes are used by `derive(MemSource)` as well. /// /// You can specify the cast mode either per field or globally on the whole /// struct. The mode of the struct applies to all fields that don't have a /// field-specific mode. /// /// ``` /// # use lox::{ /// # MemSink, VertexHandle, /// # cgmath::{Point3, Vector3}, /// # ds::HalfEdgeMesh, /// # map::DenseMap, /// # }; /// # /// #[derive(MemSink)] /// #[lox(cast = "none")] /// struct MyMesh { /// #[lox(core_mesh)] /// mesh: HalfEdgeMesh, /// /// #[lox(vertex_position)] /// positions: DenseMap<VertexHandle, Point3<f32>>, /// /// #[lox(vertex_normal, cast = "lossy")] /// normals: DenseMap<VertexHandle, Vector3<f32>>, /// } /// ``` /// /// In this example, the vertex positions inherit the "struct global" cast mode /// (`none`), while the vertex normals override that mode to `lossy`. /// /// /// ### Exact traits required for each field /// /// Traits required for the `core_mesh` field: /// - TODO /// /// Traits required for property fields. For type `T` of the field: /// - `T` must implement [`PropStoreMut`][crate::map::PropStoreMut] (with /// fitting handle type). Additionally: /// - For `vertex_position`: `T::Target` must implement /// [`Pos3Like`][crate::prop::Pos3Like]. /// - For `*_normal`: `T::Target` must implement /// [`Vec3Like`][crate::prop::Vec3Like]. /// - For `*_color`: `T::Target` must implement /// [`ColorLike`][crate::prop::ColorLike] and `T::Target::Channel` must /// implement [`Primitive`]. #[cfg(feature= "io")] pub use lox_macros::MemSink; /// Derive macro for [the `MemSource` trait][io::MemSource]. /// /// You can easily derive `MemSource` for your own types. To do that, you have /// to attach `#[derive(MemSource)]` to your struct definition (note: /// currently, the trait can only be derived for structs with named fields). /// You also have to annotate your fields with `#[lox(...)]` attributes to tell /// the derive macro what a field should be used for. Example: /// /// ``` /// use lox::{ /// MemSource, VertexHandle, /// cgmath::Point3, /// ds::SharedVertexMesh, /// map::DenseMap, /// }; /// /// /// #[derive(MemSource)] /// struct MyMesh { /// #[lox(core_mesh)] /// mesh: SharedVertexMesh, /// /// #[lox(vertex_position)] /// positions: DenseMap<VertexHandle, Point3<f32>>, /// } /// ``` /// /// Deriving this trait works very similar to deriving [`MemSink`]. See its /// documentation for more information on the custom derive. /// /// /// ### Exact traits required for each field /// /// Traits required for the `core_mesh` field: /// - TODO /// /// Traits required for property fields. For type `T` of the field: /// - `T` must implement [`PropStore`][crate::map::PropStore] (with fitting /// handle type). Additionally: /// - For `vertex_position`: `T::Target` must implement /// [`Pos3Like`][crate::prop::Pos3Like] and `T::Target::Scalar` must /// implement [`Primitive`]. /// - For `*_normal`: `T::Target` must implement /// [`Vec3Like`][crate::prop::Vec3Like] and `T::Target::Scalar` must /// implement [`Primitive`]. /// - For `*_color`: `T::Target` must implement /// [`ColorLike`][crate::prop::ColorLike] and `T::Target::Channel` must /// implement [`Primitive`]. #[cfg(feature= "io")] pub use lox_macros::MemSource; /// Convenience macro to quickly create a small mesh. /// /// (This is just a dummy macro to add documentation to the actual proc-macro /// reexported from `lox-macros`. See [#58700][i58700] and [#58696][i58696] for /// more information.) /// /// **Note about unstable features**: this proc macro needs to be invoked in /// expression context, which is still unstable. So your crate needs to enable /// the `proc_macro_hygiene` feature for this to work. /// /// [i58700]: https://github.com/rust-lang/rust/issues/58700 /// [i58696]: https://github.com/rust-lang/rust/issues/586
MeshElement
identifier_name
lib.rs
VertexHandle}; /// The three basic elements in a polygon mesh. #[derive(Debug, Clone, Copy, PartialEq, Eq)] pub enum MeshElement { Edge, Face, Vertex, } // =========================================================================== // ===== `Sealed` trait // =========================================================================== pub(crate) mod sealed { /// A trait that cannot be implemented outside of this crate. /// /// This is helpful for all "real" traits in this library that only /// abstract over a closed set of types. Thus, users shouldn't be able to /// implement those traits for their types. Adding `Sealed` as supertrait /// solves this problem. pub trait Sealed {} } // =========================================================================== // ===== Macros // =========================================================================== /// Derive macro for the [`Empty` trait][traits::Empty]. /// /// ``` /// use lox::Empty; // this imports the custom-derive and not the trait! /// /// #[derive(Empty)] /// struct MyStruct { /// a: Vec<u32>, // => `vec![]` /// b: Option<String>, // => `None` /// c: (), // => `()` /// } /// ``` /// /// This can only be derived for structs. All struct fields need to implement /// `Empty` in order for the derive to work. If your struct has generic /// parameters, they won't be bounded with `Empty` in the generated impl block. /// This is useful most of the time, because things like `Vec<T>` and /// `Option<T>` don't require `T: Empty` to implement `Empty`. But this means /// that you sometimes have to add a global `Empty` bound to your parameter or /// implement `Empty` manually. pub use lox_macros::Empty; /// Derive macro for [the `MemSink` trait][io::MemSink]. /// /// You can easily derive `MemSink` for your own types. To do that, you have to /// attach `#[derive(MemSink)]` to your struct definition (note: currently, the /// trait can only be derived for structs with named fields). You also have to /// annotate your fields with `#[lox(...)]` attributes to tell the derive macro /// what a field should be used for. Example: /// /// ``` /// use lox::{ /// MemSink, VertexHandle, /// cgmath::Point3, /// ds::HalfEdgeMesh, /// map::DenseMap, /// }; /// /// /// #[derive(MemSink)] /// struct MyMesh { /// #[lox(core_mesh)] /// mesh: HalfEdgeMesh, /// /// #[lox(vertex_position)] /// positions: DenseMap<VertexHandle, Point3<f32>>, /// } /// ``` /// /// There is one required field: the core mesh field. That field's type has to /// implement several mesh traits, in particular `MeshMut` and `TriMeshMut`. /// You have to annotate that mesh with `#[lox(core_mesh)]`. /// /// Additionally, you can have fields for each mesh property, like vertex /// position or face colors. The type of those fields has to implement /// `PropStoreMut` with a compatible element type. You have to annotate these /// property fields with the corresponding attribute. The available properties /// are: /// /// - `vertex_position` /// - `vertex_normal` /// - `vertex_color` /// - `face_normal` /// - `face_color` /// /// Furthermore, there are some configurations (like the cast mode) that can be /// configured via `lox(...)` attributes as well. See below for more /// information. /// /// /// ## Cast modes /// /// You can set a *cast mode* for each field. A `MemSink` has to be able to /// "handle" any primitive type as the source is allowed to call the property /// methods with any type. The sink can handle types either by casting or by /// returning an error. The field's cast mode determines which casts are /// allowed and which are not. Possible cast modes: /// /// - `cast = "none"` /// - `cast = "lossless"` /// - `cast = "rounding"` /// - `cast = "clamping"` /// - `cast = "lossy"` (*default*) /// /// The `none` mode does not allow casting at all. If the type provided by the /// source does not match the type in your struct, an error is returned. All /// other modes correspond to the cast modes in the [`cast` /// module][crate::cast]. /// /// Note that the cast modes are used by `derive(MemSource)` as well. /// /// You can specify the cast mode either per field or globally on the whole /// struct. The mode of the struct applies to all fields that don't have a /// field-specific mode. /// /// ``` /// # use lox::{ /// # MemSink, VertexHandle, /// # cgmath::{Point3, Vector3}, /// # ds::HalfEdgeMesh, /// # map::DenseMap, /// # }; /// # /// #[derive(MemSink)] /// #[lox(cast = "none")] /// struct MyMesh { /// #[lox(core_mesh)] /// mesh: HalfEdgeMesh, /// /// #[lox(vertex_position)] /// positions: DenseMap<VertexHandle, Point3<f32>>, /// /// #[lox(vertex_normal, cast = "lossy")] /// normals: DenseMap<VertexHandle, Vector3<f32>>, /// } /// ``` /// /// In this example, the vertex positions inherit the "struct global" cast mode /// (`none`), while the vertex normals override that mode to `lossy`. /// /// /// ### Exact traits required for each field /// /// Traits required for the `core_mesh` field: /// - TODO /// /// Traits required for property fields. For type `T` of the field: /// - `T` must implement [`PropStoreMut`][crate::map::PropStoreMut] (with /// fitting handle type). Additionally: /// - For `vertex_position`: `T::Target` must implement /// [`Pos3Like`][crate::prop::Pos3Like]. /// - For `*_normal`: `T::Target` must implement /// [`Vec3Like`][crate::prop::Vec3Like]. /// - For `*_color`: `T::Target` must implement /// [`ColorLike`][crate::prop::ColorLike] and `T::Target::Channel` must /// implement [`Primitive`]. #[cfg(feature= "io")] pub use lox_macros::MemSink; /// Derive macro for [the `MemSource` trait][io::MemSource]. /// /// You can easily derive `MemSource` for your own types. To do that, you have /// to attach `#[derive(MemSource)]` to your struct definition (note: /// currently, the trait can only be derived for structs with named fields). /// You also have to annotate your fields with `#[lox(...)]` attributes to tell /// the derive macro what a field should be used for. Example: /// /// ``` /// use lox::{
/// }; /// /// /// #[derive(MemSource)] /// struct MyMesh { /// #[lox(core_mesh)] /// mesh: SharedVertexMesh, /// /// #[lox(vertex_position)] /// positions: DenseMap<VertexHandle, Point3<f32>>, /// } /// ``` /// /// Deriving this trait works very similar to deriving [`MemSink`]. See its /// documentation for more information on the custom derive. /// /// /// ### Exact traits required for each field /// /// Traits required for the `core_mesh` field: /// - TODO /// /// Traits required for property fields. For type `T` of the field: /// - `T` must implement [`PropStore`][crate::map::PropStore] (with fitting /// handle type). Additionally: /// - For `vertex_position`: `T::Target` must implement /// [`Pos3Like`][crate::prop::Pos3Like] and `T::Target::Scalar` must /// implement [`Primitive`]. /// - For `*_normal`: `T::Target` must implement /// [`Vec3Like`][crate::prop::Vec3Like] and `T::Target::Scalar` must /// implement [`Primitive`]. /// - For `*_color`: `T::Target` must implement /// [`ColorLike`][crate::prop::ColorLike] and `T::Target::Channel` must /// implement [`Primitive`]. #[cfg(feature= "io")] pub use lox_macros::MemSource; /// Convenience macro to quickly create a small mesh. /// /// (This is just a dummy macro to add documentation to the actual proc-macro /// reexported from `lox-macros`. See [#58700][i58700] and [#58696][i58696] for /// more information.) /// /// **Note about unstable features**: this proc macro needs to be invoked in /// expression context, which is still unstable. So your crate needs to enable /// the `proc_macro_hygiene` feature for this to work. /// /// [i58700]: https://github.com/rust-lang/rust/issues/58700 /// [i58696]: https://github.com/rust-lang/rust/issues/5869
/// MemSource, VertexHandle, /// cgmath::Point3, /// ds::SharedVertexMesh, /// map::DenseMap,
random_line_split
main.rs
.iter().enumerate() { raw_futures.push(get_listing(&client, &marker.id, index)); } let unpin_futures: Vec<_> = raw_futures.into_iter().map(Box::pin).collect(); let mut mut_futures = unpin_futures; let mut documents: HashMap<String, Html> = HashMap::new(); let mut size: usize = 0; let mut current: f32 = 0.0; let total: usize = mut_futures.len(); while !mut_futures.is_empty() { match future::select_all(mut_futures).await { (Ok((id, _idx, document, content_length)), _index, remaining) => { current += 1.0; let percentage = (((current / total as f32) * 100.0) / 2.0) as usize; io::stdout() .write( format!( "\rdownloading listings {}/{}: [{}>{}]", current, total, "=".repeat(percentage), " ".repeat(50 - percentage), ) .as_bytes(), ) .unwrap(); io::stdout().flush().unwrap(); size += content_length; documents.insert(id, document); mut_futures = remaining; } (Err(_e), _index, remaining) => { error!("document failed"); mut_futures = remaining; } } } println!("\n"); info!( "downloaded {:.2?}MB from {} listings\n\t\t\t\t└──{:?}{}", size as f32 / 1000000.0, total, documents.iter().map(|v| v.0).collect::<Vec<&String>>(), " ".repeat(50) ); Ok(documents) } fn get_desired_listings(listing_map: &HashMap<String, Html>) -> Vec<DesiredListing> { let selector = Selector::parse(".facts___list___items .facts___item").unwrap(); let mut desired_listings: Vec<DesiredListing> = vec![]; for (key, value) in listing_map { let mut dl = DesiredListing::new(); let div = value.select(&selector).collect::<Vec<_>>(); for node in div { let mut node_vec = node .text() .collect::<Vec<&str>>() .iter() .map(|&v| v.trim()) .collect::<Vec<&str>>(); node_vec.retain(|&v| v != ""); if node_vec[0] == "Days on URE" && (node_vec[1] == "Just Listed" || node_vec[1].to_string().parse::<usize>().unwrap() >= 20) { dl.interested = true; } if node_vec[0] == "Status" && node_vec[1] == "Active" { dl.active = true; } } if dl.is_desired() { dl.mls = String::from(key); desired_listings.push(dl); } } desired_listings } fn remove_duplicates(listings: &mut UreData) { let mut dup_idx: Vec<usize> = vec![]; let mut existing = get_checked_listings(); for (idx, listing) in listings.markers.iter().enumerate() { if existing.contains(&listing.id) { dup_idx.push(idx); } } if dup_idx.len() > 0 { for i in dup_idx.into_iter().rev() { listings.markers.remove(i); } } if listings.markers.len() > 0 { for listing in listings.markers.iter() { existing.push(listing.id.clone()); } write_checked_listings(&existing).unwrap(); } else { info!("no new listings"); } } fn build_listing_message(listings: &Vec<DesiredListing>) -> String { let mut message_str = String::from(""); for listing in listings { message_str.push_str(&format!( "https://www.utahrealestate.com/{}\n\n", listing.mls )); } message_str } async fn send_messages(client: &request::Client, message: &str) -> SureResult<()> { let credentials = get_twilio_credentials(); let mut raw_futures = vec![]; for number in credentials.alert_numbers.iter() { raw_futures.push(send_message(&client, &message, number)) } let unpin_futures: Vec<_> = raw_futures.into_iter().map(Box::pin).collect(); let mut mut_futures = unpin_futures; while !mut_futures.is_empty() { match future::select_all(mut_futures).await { (Ok(_res), _index, remaining) => mut_futures = remaining, (Err(_e), _index, remaining) => mut_futures = remaining, } } Ok(()) } async fn get_listing( client: &request::Client, id: &str, index: usize, ) -> SureResult<(String, usize, Html, usize)> { let url = format!("https://www.utahrealestate.com/{}", id); let res = client .get(&url) .header(USER_AGENT, SURE_USER_AGENT) .send() .await?; let body = res.text().await?; let document = Html::parse_document(&body); Ok((String::from(id), index, document, body.len())) } async fn send_message(client: &request::Client, message: &str, to: &str) -> SureResult<()> { let credentials = get_twilio_credentials(); let message_url = format!( "{}/Accounts/{}/Messages.json", TWILIO_BASE_URL, credentials.sid ); let mut headers = HeaderMap::new(); headers.insert( AUTHORIZATION, format!("Basic {}", credentials.basic_auth()) .parse() .unwrap(), ); let params = [ ("From", &credentials.number), ("Body", &message.to_string()), ("To", &to.to_string()), ]; let res = client .post(&message_url) .headers(headers) .form(&params) .send() .await?; if res.status() == 201 { info!("message sent"); } else { error!( "error sending message: {:?}\n\t└──{}\n\t└──{:?}", res.status(), res.text().await?, params ) } Ok(()) } /// /// Utility Functions /// fn get_checked_listings() -> Vec<String> { let mut checked_mls: Vec<String> = vec![]; if let Ok(lines) = read_lines(&get_sure_filepath("listings.txt")) { for line in lines { if let Ok(l) = line { checked_mls.push(String::from(l.trim())) } } } checked_mls } fn write_checked_listings(checked: &Vec<String>) -> SureResult<()> { let mut contents = String::from(""); let mut file = OpenOptions::new() .write(true) .create(true) .open(&get_sure_filepath("listings.txt"))?; file.set_len(0)?; file.seek(SeekFrom::Start(0))?; let mut file = LineWriter::new(file); let mut sorted = checked .iter() .map(|v| v.parse::<usize>().unwrap()) .collect::<Vec<usize>>(); sorted.sort(); for c in sorted { contents.push_str(&format!("{}\n", c)); } file.write_all(contents.as_bytes())?; Ok(()) } fn get_ure_search_params() -> String { let mut param_encoded = String::from(""); if let Ok(lines) = read_lines(&get_sure_filepath("queries.env")) { for line in lines { if let Ok(l) = line { param_encoded.push_str(&format!("{}&", l)); } } } String::from(param_encoded) } fn get_twilio_credentials() -> TwilioAuth { let mut auth = TwilioAuth::new(); if let Ok(lines) = read_lines(&get_sure_filepath("twilio.env")) { for line in lines { if let Ok(i) = line { let config_item: Vec<&str> = i.split('=').collect(); if config_item[0] == "AccountSID" { auth.sid = String::from(config_item[1]); } if config_item[0] == "AuthToken" { auth.auth_token = String::from(config_item[1]); } if config_item[0] == "TwilioNumber" { auth.number = String::from(config_item[1]); } if config_item[0] == "AlertNumbers" { let numbers: Vec<String> = config_item[1] .split(",") .into_iter() .map(String::from) .collect(); auth.alert_numbers = numbers; } } } } auth } fn read_lines(filename: &str) -> io::Result<io::Lines<io::BufReader<File>>> { let file = File::open(filename)?; Ok(io::BufReader::new(file).lines()) } fn get_sure_filepath(
filename: &str) -
identifier_name
main.rs
Ok(()) } async fn get_session_id(client: &request::Client) -> SureResult<String> { let re = Regex::new(r#"(PHPSESSID=[\w\S]+);"#).unwrap(); let res = client .get("https://www.utahrealestate.com/index/public.index") .header(USER_AGENT, SURE_USER_AGENT) .send() .await?; let sessid = res.headers().get("set-cookie").unwrap().to_str().unwrap(); let mut id = String::from(""); for cap in re.captures_iter(sessid) { id = String::from(&cap[1]); } if id == "" { panic!("unable to find session id"); } Ok(id) } fn get_listings<'a>( client: &'a request::Client, session_id: &'a str, retry_count: usize, ) -> BoxFuture<'a, SureResult<UreData>> { if retry_count > 3 { error!("exceeded retry count - URE must be down"); std::process::exit(0); } async move { let params = get_ure_search_params(); let mut headers = HeaderMap::new(); headers.insert(USER_AGENT, SURE_USER_AGENT.parse().unwrap()); headers.insert( CONTENT_TYPE, "application/x-www-form-urlencoded".parse().unwrap(), ); headers.insert("PHPSESSID", session_id.parse().unwrap()); let res = client .post("https://www.utahrealestate.com/search/chained.update/param_reset/county_code,o_county_code,city,o_city,zip,o_zip,geometry,o_geometry/count/false/criteria/false/pg/1/limit/50/dh/1190") .headers(headers) .body(params) .send() .await?; let res_text = res.text().await?; match serde_json::from_str(&res_text) { Ok(v) => Ok(v), Err(_) => { error!("failed to parse text, retrying"); Ok(get_listings(client, session_id, retry_count + 1).await?) } } }.boxed() } async fn scrape_listings( client: &request::Client, data: &UreData, ) -> SureResult<HashMap<String, Html>> { let mut raw_futures = vec![]; for (index, marker) in data.markers.iter().enumerate() { raw_futures.push(get_listing(&client, &marker.id, index)); } let unpin_futures: Vec<_> = raw_futures.into_iter().map(Box::pin).collect(); let mut mut_futures = unpin_futures; let mut documents: HashMap<String, Html> = HashMap::new(); let mut size: usize = 0; let mut current: f32 = 0.0; let total: usize = mut_futures.len(); while !mut_futures.is_empty() { match future::select_all(mut_futures).await {
io::stdout() .write( format!( "\rdownloading listings {}/{}: [{}>{}]", current, total, "=".repeat(percentage), " ".repeat(50 - percentage), ) .as_bytes(), ) .unwrap(); io::stdout().flush().unwrap(); size += content_length; documents.insert(id, document); mut_futures = remaining; } (Err(_e), _index, remaining) => { error!("document failed"); mut_futures = remaining; } } } println!("\n"); info!( "downloaded {:.2?}MB from {} listings\n\t\t\t\t└──{:?}{}", size as f32 / 1000000.0, total, documents.iter().map(|v| v.0).collect::<Vec<&String>>(), " ".repeat(50) ); Ok(documents) } fn get_desired_listings(listing_map: &HashMap<String, Html>) -> Vec<DesiredListing> { let selector = Selector::parse(".facts___list___items .facts___item").unwrap(); let mut desired_listings: Vec<DesiredListing> = vec![]; for (key, value) in listing_map { let mut dl = DesiredListing::new(); let div = value.select(&selector).collect::<Vec<_>>(); for node in div { let mut node_vec = node .text() .collect::<Vec<&str>>() .iter() .map(|&v| v.trim()) .collect::<Vec<&str>>(); node_vec.retain(|&v| v != ""); if node_vec[0] == "Days on URE" && (node_vec[1] == "Just Listed" || node_vec[1].to_string().parse::<usize>().unwrap() >= 20) { dl.interested = true; } if node_vec[0] == "Status" && node_vec[1] == "Active" { dl.active = true; } } if dl.is_desired() { dl.mls = String::from(key); desired_listings.push(dl); } } desired_listings } fn remove_duplicates(listings: &mut UreData) { let mut dup_idx: Vec<usize> = vec![]; let mut existing = get_checked_listings(); for (idx, listing) in listings.markers.iter().enumerate() { if existing.contains(&listing.id) { dup_idx.push(idx); } } if dup_idx.len() > 0 { for i in dup_idx.into_iter().rev() { listings.markers.remove(i); } } if listings.markers.len() > 0 { for listing in listings.markers.iter() { existing.push(listing.id.clone()); } write_checked_listings(&existing).unwrap(); } else { info!("no new listings"); } } fn build_listing_message(listings: &Vec<DesiredListing>) -> String { let mut message_str = String::from(""); for listing in listings { message_str.push_str(&format!( "https://www.utahrealestate.com/{}\n\n", listing.mls )); } message_str } async fn send_messages(client: &request::Client, message: &str) -> SureResult<()> { let credentials = get_twilio_credentials(); let mut raw_futures = vec![]; for number in credentials.alert_numbers.iter() { raw_futures.push(send_message(&client, &message, number)) } let unpin_futures: Vec<_> = raw_futures.into_iter().map(Box::pin).collect(); let mut mut_futures = unpin_futures; while !mut_futures.is_empty() { match future::select_all(mut_futures).await { (Ok(_res), _index, remaining) => mut_futures = remaining, (Err(_e), _index, remaining) => mut_futures = remaining, } } Ok(()) } async fn get_listing( client: &request::Client, id: &str, index: usize, ) -> SureResult<(String, usize, Html, usize)> { let url = format!("https://www.utahrealestate.com/{}", id); let res = client .get(&url) .header(USER_AGENT, SURE_USER_AGENT) .send() .await?; let body = res.text().await?; let document = Html::parse_document(&body); Ok((String::from(id), index, document, body.len())) } async fn send_message(client: &request::Client, message: &str, to: &str) -> SureResult<()> { let credentials = get_twilio_credentials(); let message_url = format!( "{}/Accounts/{}/Messages.json", TWILIO_BASE_URL, credentials.sid ); let mut headers = HeaderMap::new(); headers.insert( AUTHORIZATION, format!("Basic {}", credentials.basic_auth()) .parse() .unwrap(), ); let params = [ ("From", &credentials.number), ("Body", &message.to_string()), ("To", &to.to_string()), ]; let res = client .post(&message_url) .headers(headers) .form(&params) .send() .await?; if res.status() == 201 { info!("message sent"); } else { error!( "error sending message: {:?}\n\t└──{}\n\t└──{:?}", res.status(), res.text().await?, params ) } Ok(()) } /// /// Utility Functions /// fn get_checked_listings() -> Vec<String> { let mut checked_mls: Vec<String> = vec![]; if let Ok(lines) = read_lines(&get_sure_filepath("listings.txt")) { for line in lines { if let Ok(l) = line { checked_mls.push(String::from(l.trim())) } }
(Ok((id, _idx, document, content_length)), _index, remaining) => { current += 1.0; let percentage = (((current / total as f32) * 100.0) / 2.0) as usize;
random_line_split
main.rs
Ok(()) } async fn get_session_id(client: &request::Client) -> SureResult<String> { let re = Regex::new(r#"(PHPSESSID=[\w\S]+);"#).unwrap(); let res = client .get("https://www.utahrealestate.com/index/public.index") .header(USER_AGENT, SURE_USER_AGENT) .send() .await?; let sessid = res.headers().get("set-cookie").unwrap().to_str().unwrap(); let mut id = String::from(""); for cap in re.captures_iter(sessid) { id = String::from(&cap[1]); } if id == "" { panic!("unable to find session id"); } Ok(id) } fn get_listings<'a>( client: &'a request::Client, session_id: &'a str, retry_count: usize, ) -> BoxFuture<'a, SureResult<UreData>> { if retry_count > 3 { error!("exceeded retry count - URE must be down"); std::process::exit(0); } async move { let params = get_ure_search_params(); let mut headers = HeaderMap::new(); headers.insert(USER_AGENT, SURE_USER_AGENT.parse().unwrap()); headers.insert( CONTENT_TYPE, "application/x-www-form-urlencoded".parse().unwrap(), ); headers.insert("PHPSESSID", session_id.parse().unwrap()); let res = client .post("https://www.utahrealestate.com/search/chained.update/param_reset/county_code,o_county_code,city,o_city,zip,o_zip,geometry,o_geometry/count/false/criteria/false/pg/1/limit/50/dh/1190") .headers(headers) .body(params) .send() .await?; let res_text = res.text().await?; match serde_json::from_str(&res_text) { Ok(v) => Ok(v), Err(_) => { error!("failed to parse text, retrying"); Ok(get_listings(client, session_id, retry_count + 1).await?) } } }.boxed() } async fn scrape_listings( client: &request::Client, data: &UreData, ) -> SureResult<HashMap<String, Html>> { let mut raw_futures = vec![]; for (index, marker) in data.markers.iter().enumerate() { raw_futures.push(get_listing(&client, &marker.id, index)); } let unpin_futures: Vec<_> = raw_futures.into_iter().map(Box::pin).collect(); let mut mut_futures = unpin_futures; let mut documents: HashMap<String, Html> = HashMap::new(); let mut size: usize = 0; let mut current: f32 = 0.0; let total: usize = mut_futures.len(); while !mut_futures.is_empty() { match future::select_all(mut_futures).await { (Ok((id, _idx, document, content_length)), _index, remaining) => { current += 1.0; let percentage = (((current / total as f32) * 100.0) / 2.0) as usize; io::stdout() .write( format!( "\rdownloading listings {}/{}: [{}>{}]", current, total, "=".repeat(percentage), " ".repeat(50 - percentage), ) .as_bytes(), ) .unwrap(); io::stdout().flush().unwrap(); size += content_length; documents.insert(id, document); mut_futures = remaining; } (Err(_e), _index, remaining) => { error!("document failed"); mut_futures = remaining; } } } println!("\n"); info!( "downloaded {:.2?}MB from {} listings\n\t\t\t\t└──{:?}{}", size as f32 / 1000000.0, total, documents.iter().map(|v| v.0).collect::<Vec<&String>>(), " ".repeat(50) ); Ok(documents) } fn get_desired_listings(listing_map: &HashMap<String, Html>) -> Vec<DesiredListing> { let selector = Selector::parse(".facts___list___items .facts___item").unwrap(); let mut desired_listings: Vec<DesiredListing> = vec![]; for (key, value) in listing_map { let mut dl = DesiredListing::new(); let div = value.select(&selector).collect::<Vec<_>>(); for node in div { let mut node_vec = node .text() .collect::<Vec<&str>>() .iter() .map(|&v| v.trim()) .collect::<Vec<&str>>(); node_vec.retain(|&v| v != ""); if node_vec[0] == "Days on URE" && (node_vec[1] == "Just Listed" || node_vec[1].to_string().parse::<usize>().unwrap() >= 20) { dl.interested = true; } if node_vec[0] == "Status" && node_vec[1] == "Active" { dl.active = true; } } if dl.is_desired() { dl.mls = String::from(key); desired_listings.push(dl); } } desired_listings } fn remove_duplicates(listings: &mut UreData) { let mut dup_idx: Vec<usize> = vec![]; let mut existing = get_checked_listings(); for (idx, listing) in listings.markers.iter().enumerate() { if existing.contains(&listing.id) { dup_idx.push(idx); } } if dup_idx.len() > 0 { for i in dup_idx.into_iter().rev() { listings.markers.remove(i); } } if listings.markers.len() > 0 { for listing in listings.markers.iter() { existing.push(listing.id.clone()); } write_checked_listings(&existing).unwrap(); } else { info!("no new listings"); } } fn build_listing_message(listings: &Vec<DesiredListing>) -> String { let mut message_str = String::from(""); for listing in listings { message_str.push_str(&format!( "https://www.utahrealestate.com/{}\n\n", listing.mls )); } message_str } async fn send_messages(client: &request::Client, message: &str) -> SureResult<()> { let credentials = get_twilio_credentials(); let mut raw_futures = vec![]; for number in credentials.alert_numbers.iter() { raw_futures.push(send_message(&client, &message, number)) } let unpin_futures: Vec<_> = raw_futures.into_iter().map(Box::pin).collect(); let mut mut_futures = unpin_futures; while !mut_futures.is_empty() { match future::select_all(mut_futures).await { (Ok(_res), _index, remaining) => mut_futures = remaining, (Err(_e), _index, remaining) => mut_futures = remaining, } } Ok(()) } async fn get_listing( client: &request::Client, id: &str, index: usize, ) -> SureResult<(String, usize, Html, usize)> { let url = format!("https://www.utahrealestate.com/{}", id); let res = client .get(&url) .header(USER_AGENT, SURE_USER_AGENT) .send() .await?; let body = res.text().await?; let document = Html::parse_document(&body); Ok((String::from(id), index, document, body.len())) } async fn send_message(client: &request::Client, message: &str, to: &str) -> SureResult<()> { let credentials = get_twilio_credentials(); let message_url = format!( "{}/Accounts/{}/Messages.json", TWILIO_BASE_URL, credentials.sid ); let mut headers = HeaderMap::new(); headers.insert( AUTHORIZATION, format!("Basic {}", credentials.basic_auth()) .parse() .unwrap(), ); let params = [ ("From", &credentials.number), ("Body", &message.to_string()), ("To", &to.to_string()), ]; let res = client .post(&message_url) .headers(headers) .form(&params) .send() .await?; if res.status() == 201 { info!("message sent"); } else { er
// Utility Functions /// fn get_checked_listings() -> Vec<String> { let mut checked_mls: Vec<String> = vec![]; if let Ok(lines) = read_lines(&get_sure_filepath("listings.txt")) { for line in lines { if let Ok(l) = line { checked_mls.push(String::from(l.trim())) }
ror!( "error sending message: {:?}\n\t└──{}\n\t└──{:?}", res.status(), res.text().await?, params ) } Ok(()) } /// /
conditional_block
main.rs
/// easily remove all the neighbors of a vertex from a state very efficiently. neighbors: Vec<BitSet>, /// For each vertex 'i', the value of 'weight[i]' denotes the weight associated /// to vertex i in the problem instance. The goal of MISP is to select the nodes /// from the underlying graph such that the resulting set is an independent set /// where the sum of the weights of selected vertices is maximum. weight: Vec<isize>, } /// A constant to mean take the node in the independent set. const YES: isize = 1; /// A constant to mean leave the node out of the independent set. const NO: isize = 0; /// The Misp class implements the 'Problem' trait. This means Misp is the definition /// of the DP model. That DP model is pretty straightforward, still you might want /// to check the implementation of the branching heuristic (next_variable method) /// since it does interesting stuffs. impl Problem for Misp { type State = BitSet; fn nb_variables(&self) -> usize { self.nb_vars } fn initial_state(&self) -> Self::State { (0..self.nb_variables()).collect() } fn initial_value(&self) -> isize { 0 } fn transition(&self, state: &Self::State, decision: Decision) -> Self::State { let mut res = state.clone(); res.remove(decision.variable.id()); if decision.value == YES { // intersect with complement of the neighbors for fast set difference res.intersect_with(&self.neighbors[decision.variable.id()]); } res } fn transition_cost(&self, _: &Self::State, decision: Decision) -> isize { if decision.value == NO { 0 } else { self.weight[decision.variable.id()] } } fn for_each_in_domain(&self, variable: Variable, state: &Self::State, f: &mut dyn DecisionCallback) { if state.contains(variable.id()) { f.apply(Decision{variable, value: YES}); f.apply(Decision{variable, value: NO }); } else { f.apply(Decision{variable, value: NO }); } } /// This method is (apparently) a bit more hairy. What it does is it simply decides to branch on /// the variable that occurs in the least number of states present in the next layer. The intuition /// here is to limit the max width as much as possible when developing the layers since all /// nodes that are not impacted by the change on the selectd vertex are simply copied over to the /// next layer. fn next_variable(&self, _: usize, next_layer: &mut dyn Iterator<Item = &Self::State>) -> Option<Variable> { // The thread local stuff is possibly one of the most surprising bits of this code. It declares // a static variable called VAR_HEURISTIC storing the counts of each vertex in the next layer. // The fact that it is static means that it will not be re-created (re allocated) upon each // pass. The fact that it is declared within a thread local block, means that this static var // will be created with a potentially mutable access for each thread. thread_local! { static VAR_HEURISTIC: RefCell<Vec<usize>> = RefCell::new(vec![]); } VAR_HEURISTIC.with(|heu| { let mut heu = heu.borrow_mut(); let heu: &mut Vec<usize> = heu.as_mut(); // initialize heu.reserve_exact(self.nb_variables()); if heu.is_empty() { for _ in 0..self.nb_variables() { heu.push(0); } } else { heu.iter_mut().for_each(|i| *i = 0); } // count the occurrence of each var for s in next_layer { for sit in s.iter() { heu[sit] += 1; } } // take the one occurring the least often heu.iter().copied().enumerate() .filter(|(_, v)| *v > 0) .min_by_key(|(_, v)| *v) .map(|(x, _)| Variable(x)) }) } fn is_impacted_by(&self, var: Variable, state: &Self::State) -> bool
} /// In addition to a dynamic programming (DP) model of the problem you want to solve, /// the branch and bound with MDD algorithm (and thus ddo) requires that you provide /// an additional relaxation allowing to control the maximum amount of space used by /// the decision diagrams that are compiled. /// /// That relaxation requires two operations: one to merge several nodes into one /// merged node that acts as an over approximation of the other nodes. The second /// operation is used to possibly offset some weight that would otherwise be lost /// to the arcs entering the newly created merged node. /// /// The role of this very simple structure is simply to provide an implementation /// of that relaxation. /// /// # Note: /// In addition to the aforementioned two operations, the MispRelax structure implements /// an optional `fast_upper_bound` method. Which one provides a useful bound to /// prune some portions of the state-space as the decision diagrams are compiled. /// (aka rough upper bound pruning). pub struct MispRelax<'a>{pb: &'a Misp} impl Relaxation for MispRelax<'_> { type State = BitSet; fn merge(&self, states: &mut dyn Iterator<Item = &Self::State>) -> Self::State { let mut state = BitSet::with_capacity(self.pb.nb_variables()); for s in states { state.union_with(s); } state } fn relax( &self, _source: &Self::State, _dest: &Self::State, _new: &Self::State, _decision: Decision, cost: isize, ) -> isize { cost } fn fast_upper_bound(&self, state: &Self::State) -> isize { state.iter().map(|x| self.pb.weight[x]).sum() } } /// The last bit of information which we need to provide when implementing a ddo-based /// solver is a `StateRanking`. This is an heuristic which is used to select the most /// and least promising nodes as a means to only delete/merge the *least* promising nodes /// when compiling restricted and relaxed DDs. pub struct MispRanking; impl StateRanking for MispRanking { type State = BitSet; fn compare(&self, a: &Self::State, b: &Self::State) -> std::cmp::Ordering { a.len().cmp(&b.len()) .then_with(|| a.cmp(b)) } } // ######################################################################################### // # THE INFORMATION BEYOND THIS LINE IS NOT DIRECTLY RELATED TO THE IMPLEMENTATION OF # // # A SOLVER BASED ON DDO. INSTEAD, THAT PORTION OF THE CODE CONTAINS GENERIC FUNCTION # // # THAT ARE USED TO READ AN INSTANCE FROM FILE, PROCESS COMMAND LINE ARGUMENTS, AND # // # THE MAIN FUNCTION. THESE ARE THUS NOT REQUIRED 'PER-SE', BUT I BELIEVE IT IS USEFUL # // # TO SHOW HOW IT CAN BE DONE IN AN EXAMPLE. # // ######################################################################################### /// This structure uses `clap-derive` annotations and define the arguments that can /// be passed on to the executable solver. #[derive(Parser, Debug)] #[command(author, version, about, long_about = None)] struct Args { /// The path to the instance file fname: String, /// The number of concurrent threads #[clap(short, long, default_value = "8")] threads: usize, /// The maximum amount of time you would like this solver to run #[clap(short, long)] duration: Option<u64>, /// The maximum number of nodes per layer #[clap(short, long)] width: Option<usize>, } /// This enumeration simply groups the kind of errors that might occur when parsing a /// misp instance from file. There can be io errors (file unavailable ?), format error /// (e.g. the file is not an instance but contains the text of your next paper), /// or parse int errors (which are actually a variant of the format error since it tells /// you that the parser expected an integer number but got ... something else). #[derive(Debug, thiserror::Error)] enum Error { /// There was an io related error #[error("io error {0}")] Io(#[from] std::io::Error), /// The parser expected to read something that was an integer but got some garbage #[error("parse int {0}")] ParseInt(#[from] ParseIntError), /// The file was not properly formatted. #[error("ill formed instance")] Format, } /// This function is used to read a misp instance from file. It returns either a /// misp instance if everything went on well or an error describing the problem. fn read_instance<P: AsRef<Path>>(fname: P) -> Result<Misp, Error> { let f = File::open(fname)?; let f = BufReader::new(f); let comment = Regex::
{ state.contains(var.id()) }
identifier_body
main.rs
/// of the DP model. That DP model is pretty straightforward, still you might want /// to check the implementation of the branching heuristic (next_variable method) /// since it does interesting stuffs. impl Problem for Misp { type State = BitSet; fn nb_variables(&self) -> usize { self.nb_vars } fn initial_state(&self) -> Self::State { (0..self.nb_variables()).collect() } fn initial_value(&self) -> isize { 0 } fn transition(&self, state: &Self::State, decision: Decision) -> Self::State { let mut res = state.clone(); res.remove(decision.variable.id()); if decision.value == YES { // intersect with complement of the neighbors for fast set difference res.intersect_with(&self.neighbors[decision.variable.id()]); } res } fn transition_cost(&self, _: &Self::State, decision: Decision) -> isize { if decision.value == NO { 0 } else { self.weight[decision.variable.id()] } } fn for_each_in_domain(&self, variable: Variable, state: &Self::State, f: &mut dyn DecisionCallback) { if state.contains(variable.id()) { f.apply(Decision{variable, value: YES}); f.apply(Decision{variable, value: NO }); } else { f.apply(Decision{variable, value: NO }); } } /// This method is (apparently) a bit more hairy. What it does is it simply decides to branch on /// the variable that occurs in the least number of states present in the next layer. The intuition /// here is to limit the max width as much as possible when developing the layers since all /// nodes that are not impacted by the change on the selectd vertex are simply copied over to the /// next layer. fn next_variable(&self, _: usize, next_layer: &mut dyn Iterator<Item = &Self::State>) -> Option<Variable> { // The thread local stuff is possibly one of the most surprising bits of this code. It declares // a static variable called VAR_HEURISTIC storing the counts of each vertex in the next layer. // The fact that it is static means that it will not be re-created (re allocated) upon each // pass. The fact that it is declared within a thread local block, means that this static var // will be created with a potentially mutable access for each thread. thread_local! { static VAR_HEURISTIC: RefCell<Vec<usize>> = RefCell::new(vec![]); } VAR_HEURISTIC.with(|heu| { let mut heu = heu.borrow_mut(); let heu: &mut Vec<usize> = heu.as_mut(); // initialize heu.reserve_exact(self.nb_variables()); if heu.is_empty() { for _ in 0..self.nb_variables() { heu.push(0); } } else { heu.iter_mut().for_each(|i| *i = 0); } // count the occurrence of each var for s in next_layer { for sit in s.iter() { heu[sit] += 1; } } // take the one occurring the least often heu.iter().copied().enumerate() .filter(|(_, v)| *v > 0) .min_by_key(|(_, v)| *v) .map(|(x, _)| Variable(x)) }) } fn is_impacted_by(&self, var: Variable, state: &Self::State) -> bool { state.contains(var.id()) } } /// In addition to a dynamic programming (DP) model of the problem you want to solve, /// the branch and bound with MDD algorithm (and thus ddo) requires that you provide /// an additional relaxation allowing to control the maximum amount of space used by /// the decision diagrams that are compiled. /// /// That relaxation requires two operations: one to merge several nodes into one /// merged node that acts as an over approximation of the other nodes. The second /// operation is used to possibly offset some weight that would otherwise be lost /// to the arcs entering the newly created merged node. /// /// The role of this very simple structure is simply to provide an implementation /// of that relaxation. /// /// # Note: /// In addition to the aforementioned two operations, the MispRelax structure implements /// an optional `fast_upper_bound` method. Which one provides a useful bound to /// prune some portions of the state-space as the decision diagrams are compiled. /// (aka rough upper bound pruning). pub struct MispRelax<'a>{pb: &'a Misp} impl Relaxation for MispRelax<'_> { type State = BitSet; fn merge(&self, states: &mut dyn Iterator<Item = &Self::State>) -> Self::State { let mut state = BitSet::with_capacity(self.pb.nb_variables()); for s in states { state.union_with(s); } state } fn relax( &self, _source: &Self::State, _dest: &Self::State, _new: &Self::State, _decision: Decision, cost: isize, ) -> isize { cost } fn fast_upper_bound(&self, state: &Self::State) -> isize { state.iter().map(|x| self.pb.weight[x]).sum() } } /// The last bit of information which we need to provide when implementing a ddo-based /// solver is a `StateRanking`. This is an heuristic which is used to select the most /// and least promising nodes as a means to only delete/merge the *least* promising nodes /// when compiling restricted and relaxed DDs. pub struct MispRanking; impl StateRanking for MispRanking { type State = BitSet; fn compare(&self, a: &Self::State, b: &Self::State) -> std::cmp::Ordering { a.len().cmp(&b.len()) .then_with(|| a.cmp(b)) } } // ######################################################################################### // # THE INFORMATION BEYOND THIS LINE IS NOT DIRECTLY RELATED TO THE IMPLEMENTATION OF # // # A SOLVER BASED ON DDO. INSTEAD, THAT PORTION OF THE CODE CONTAINS GENERIC FUNCTION # // # THAT ARE USED TO READ AN INSTANCE FROM FILE, PROCESS COMMAND LINE ARGUMENTS, AND # // # THE MAIN FUNCTION. THESE ARE THUS NOT REQUIRED 'PER-SE', BUT I BELIEVE IT IS USEFUL # // # TO SHOW HOW IT CAN BE DONE IN AN EXAMPLE. # // ######################################################################################### /// This structure uses `clap-derive` annotations and define the arguments that can /// be passed on to the executable solver. #[derive(Parser, Debug)] #[command(author, version, about, long_about = None)] struct Args { /// The path to the instance file fname: String, /// The number of concurrent threads #[clap(short, long, default_value = "8")] threads: usize, /// The maximum amount of time you would like this solver to run #[clap(short, long)] duration: Option<u64>, /// The maximum number of nodes per layer #[clap(short, long)] width: Option<usize>, } /// This enumeration simply groups the kind of errors that might occur when parsing a /// misp instance from file. There can be io errors (file unavailable ?), format error /// (e.g. the file is not an instance but contains the text of your next paper), /// or parse int errors (which are actually a variant of the format error since it tells /// you that the parser expected an integer number but got ... something else). #[derive(Debug, thiserror::Error)] enum Error { /// There was an io related error #[error("io error {0}")] Io(#[from] std::io::Error), /// The parser expected to read something that was an integer but got some garbage #[error("parse int {0}")] ParseInt(#[from] ParseIntError), /// The file was not properly formatted. #[error("ill formed instance")] Format, } /// This function is used to read a misp instance from file. It returns either a /// misp instance if everything went on well or an error describing the problem. fn read_instance<P: AsRef<Path>>(fname: P) -> Result<Misp, Error> { let f = File::open(fname)?; let f = BufReader::new(f); let comment = Regex::new(r"^c\s.*$").unwrap(); let pb_decl = Regex::new(r"^p\s+edge\s+(?P<vars>\d+)\s+(?P<edges>\d+)$").unwrap(); let node_decl = Regex::new(r"^n\s+(?P<node>\d+)\s+(?P<weight>-?\d+)").unwrap(); let edge_decl = Regex::new(r"^e\s+(?P<src>\d+)\s+(?P<dst>\d+)").unwrap(); let mut g = Misp{nb_vars: 0, neighbors: vec![], weight: vec![]}; for line in f.lines() { let line = line?; let line = line.trim(); if line.is_empty()
{ continue; }
conditional_block
main.rs
/// easily remove all the neighbors of a vertex from a state very efficiently. neighbors: Vec<BitSet>, /// For each vertex 'i', the value of 'weight[i]' denotes the weight associated /// to vertex i in the problem instance. The goal of MISP is to select the nodes /// from the underlying graph such that the resulting set is an independent set /// where the sum of the weights of selected vertices is maximum. weight: Vec<isize>, } /// A constant to mean take the node in the independent set. const YES: isize = 1; /// A constant to mean leave the node out of the independent set. const NO: isize = 0; /// The Misp class implements the 'Problem' trait. This means Misp is the definition /// of the DP model. That DP model is pretty straightforward, still you might want /// to check the implementation of the branching heuristic (next_variable method) /// since it does interesting stuffs. impl Problem for Misp { type State = BitSet;
fn initial_state(&self) -> Self::State { (0..self.nb_variables()).collect() } fn initial_value(&self) -> isize { 0 } fn transition(&self, state: &Self::State, decision: Decision) -> Self::State { let mut res = state.clone(); res.remove(decision.variable.id()); if decision.value == YES { // intersect with complement of the neighbors for fast set difference res.intersect_with(&self.neighbors[decision.variable.id()]); } res } fn transition_cost(&self, _: &Self::State, decision: Decision) -> isize { if decision.value == NO { 0 } else { self.weight[decision.variable.id()] } } fn for_each_in_domain(&self, variable: Variable, state: &Self::State, f: &mut dyn DecisionCallback) { if state.contains(variable.id()) { f.apply(Decision{variable, value: YES}); f.apply(Decision{variable, value: NO }); } else { f.apply(Decision{variable, value: NO }); } } /// This method is (apparently) a bit more hairy. What it does is it simply decides to branch on /// the variable that occurs in the least number of states present in the next layer. The intuition /// here is to limit the max width as much as possible when developing the layers since all /// nodes that are not impacted by the change on the selectd vertex are simply copied over to the /// next layer. fn next_variable(&self, _: usize, next_layer: &mut dyn Iterator<Item = &Self::State>) -> Option<Variable> { // The thread local stuff is possibly one of the most surprising bits of this code. It declares // a static variable called VAR_HEURISTIC storing the counts of each vertex in the next layer. // The fact that it is static means that it will not be re-created (re allocated) upon each // pass. The fact that it is declared within a thread local block, means that this static var // will be created with a potentially mutable access for each thread. thread_local! { static VAR_HEURISTIC: RefCell<Vec<usize>> = RefCell::new(vec![]); } VAR_HEURISTIC.with(|heu| { let mut heu = heu.borrow_mut(); let heu: &mut Vec<usize> = heu.as_mut(); // initialize heu.reserve_exact(self.nb_variables()); if heu.is_empty() { for _ in 0..self.nb_variables() { heu.push(0); } } else { heu.iter_mut().for_each(|i| *i = 0); } // count the occurrence of each var for s in next_layer { for sit in s.iter() { heu[sit] += 1; } } // take the one occurring the least often heu.iter().copied().enumerate() .filter(|(_, v)| *v > 0) .min_by_key(|(_, v)| *v) .map(|(x, _)| Variable(x)) }) } fn is_impacted_by(&self, var: Variable, state: &Self::State) -> bool { state.contains(var.id()) } } /// In addition to a dynamic programming (DP) model of the problem you want to solve, /// the branch and bound with MDD algorithm (and thus ddo) requires that you provide /// an additional relaxation allowing to control the maximum amount of space used by /// the decision diagrams that are compiled. /// /// That relaxation requires two operations: one to merge several nodes into one /// merged node that acts as an over approximation of the other nodes. The second /// operation is used to possibly offset some weight that would otherwise be lost /// to the arcs entering the newly created merged node. /// /// The role of this very simple structure is simply to provide an implementation /// of that relaxation. /// /// # Note: /// In addition to the aforementioned two operations, the MispRelax structure implements /// an optional `fast_upper_bound` method. Which one provides a useful bound to /// prune some portions of the state-space as the decision diagrams are compiled. /// (aka rough upper bound pruning). pub struct MispRelax<'a>{pb: &'a Misp} impl Relaxation for MispRelax<'_> { type State = BitSet; fn merge(&self, states: &mut dyn Iterator<Item = &Self::State>) -> Self::State { let mut state = BitSet::with_capacity(self.pb.nb_variables()); for s in states { state.union_with(s); } state } fn relax( &self, _source: &Self::State, _dest: &Self::State, _new: &Self::State, _decision: Decision, cost: isize, ) -> isize { cost } fn fast_upper_bound(&self, state: &Self::State) -> isize { state.iter().map(|x| self.pb.weight[x]).sum() } } /// The last bit of information which we need to provide when implementing a ddo-based /// solver is a `StateRanking`. This is an heuristic which is used to select the most /// and least promising nodes as a means to only delete/merge the *least* promising nodes /// when compiling restricted and relaxed DDs. pub struct MispRanking; impl StateRanking for MispRanking { type State = BitSet; fn compare(&self, a: &Self::State, b: &Self::State) -> std::cmp::Ordering { a.len().cmp(&b.len()) .then_with(|| a.cmp(b)) } } // ######################################################################################### // # THE INFORMATION BEYOND THIS LINE IS NOT DIRECTLY RELATED TO THE IMPLEMENTATION OF # // # A SOLVER BASED ON DDO. INSTEAD, THAT PORTION OF THE CODE CONTAINS GENERIC FUNCTION # // # THAT ARE USED TO READ AN INSTANCE FROM FILE, PROCESS COMMAND LINE ARGUMENTS, AND # // # THE MAIN FUNCTION. THESE ARE THUS NOT REQUIRED 'PER-SE', BUT I BELIEVE IT IS USEFUL # // # TO SHOW HOW IT CAN BE DONE IN AN EXAMPLE. # // ######################################################################################### /// This structure uses `clap-derive` annotations and define the arguments that can /// be passed on to the executable solver. #[derive(Parser, Debug)] #[command(author, version, about, long_about = None)] struct Args { /// The path to the instance file fname: String, /// The number of concurrent threads #[clap(short, long, default_value = "8")] threads: usize, /// The maximum amount of time you would like this solver to run #[clap(short, long)] duration: Option<u64>, /// The maximum number of nodes per layer #[clap(short, long)] width: Option<usize>, } /// This enumeration simply groups the kind of errors that might occur when parsing a /// misp instance from file. There can be io errors (file unavailable ?), format error /// (e.g. the file is not an instance but contains the text of your next paper), /// or parse int errors (which are actually a variant of the format error since it tells /// you that the parser expected an integer number but got ... something else). #[derive(Debug, thiserror::Error)] enum Error { /// There was an io related error #[error("io error {0}")] Io(#[from] std::io::Error), /// The parser expected to read something that was an integer but got some garbage #[error("parse int {0}")] ParseInt(#[from] ParseIntError), /// The file was not properly formatted. #[error("ill formed instance")] Format, } /// This function is used to read a misp instance from file. It returns either a /// misp instance if everything went on well or an error describing the problem. fn read_instance<P: AsRef<Path>>(fname: P) -> Result<Misp, Error> { let f = File::open(fname)?; let f = BufReader::new(f); let comment = Regex::new
fn nb_variables(&self) -> usize { self.nb_vars }
random_line_split