file_name
large_stringlengths
4
140
prefix
large_stringlengths
0
12.1k
suffix
large_stringlengths
0
12k
middle
large_stringlengths
0
7.51k
fim_type
large_stringclasses
4 values
utils.py
} -- config file containing data paths and tokenizer information metadata {bool} -- whether the data contains metadata for augmented embeddings proportion {float} -- proportion for splitting up train and test. (default: {0.7}) max_len {int} -- maximum token length for a text. (default: {128}) partition {dict} -- maps lists of training and validation data IDs (default: {None}) labels {dict} -- (default: {None}) Returns: partition {dict} -- list of ids in train and valid datasets torch.utils.data.Dataset -- dataset """ # columns if meta: [0] unique ID, [1] text, [2] metadata, [3] label # columns if no meta: [0] unique ID, [1] text, [2] label if config["metadata"]: unique_id_col = 0 text_col = 1 metadata_col = 2 label_col = 3 else: unique_id_col = 0 text_col = 1 label_col = 3 dataset = pd.read_csv(config['train_file'], header=None, sep='\t') print(dataset) # below fix null values wrecking encode_plus # convert labels to integer and drop nas dataset.iloc[:, label_col] = pd.to_numeric(dataset.iloc[:, label_col], errors = 'coerce' ) dataset = dataset[~ dataset[text_col].isnull()] # recreate the first column with the reset index. dataset = dataset[(dataset.iloc[:, label_col] == 1) | (dataset.iloc[:, label_col] == 0)] \ .reset_index().reset_index().drop(columns = ['index', 0]).rename(columns = {'level_0': 0}) print(dataset) # create list of train/valid IDs if not provided if not partition and not labels: ids = list(dataset.iloc[:,unique_id_col]) total_len = len(ids) np.random.shuffle(ids) labels = {} # metadata = {} partition = {'train': ids[ :int(total_len * 0.7)],
labels[i] = dataset.iloc[i][label_col] # set parameters for DataLoader -- num_workers = cores params = {'batch_size': 32, 'shuffle': True, 'num_workers': 0 } tokenizer = AutoTokenizer.from_pretrained(vocab) dataset[text_col] = dataset[text_col].apply(lambda x: tokenizer.encode_plus(str(x), \ max_length=max_len, \ add_special_tokens=True, \ pad_to_max_length=True, \ truncation=True)) if config['metadata']: # glove for metadata preprocessing glove = torchtext.vocab.GloVe(name="6B", dim=50) dataset[metadata_col] = dataset[metadata_col].apply(lambda y: __pad__(str(y).split(" "), 30)) dataset[metadata_col] = dataset[metadata_col].apply(lambda z: __glove_embed__(z, glove)) train_data = dataset[dataset[unique_id_col].isin(partition['train'])] valid_data = dataset[dataset[unique_id_col].isin(partition['valid'])] # create train/valid generators training_set = AbstractDataset(data=train_data, labels=labels, metadata=config['metadata'], list_IDs=partition['train'], max_len = max_len) training_generator = DataLoader(training_set, **params) validation_set = AbstractDataset(data=valid_data, labels=labels, metadata=config['metadata'], list_IDs=partition['valid'],max_len = max_len) validation_generator = DataLoader(validation_set, **params) return partition, training_generator, validation_generator def __pad__(sequence, max_l): """ Padding function for 1D sequences """ if max_l - len(sequence) < 0: sequence = sequence[:max_l] else: sequence = np.pad(sequence, (0, max_l - (len(sequence))), 'constant', constant_values=(0)) return sequence def __glove_embed__(sequence, model): """ Embed words in a sequence using GLoVE model """ embedded = [] for word in sequence: embedded.append(model[word]) return embedded def load_embeddings(config, name, vocab, training_generator, validation_generator): """Load embeddings either from cache or from scratch Args: config (json) -- file configurations. name -- vocab -- training_generator -- validation_generator -- Returns: embedding_shape, train_embeddings, valid_embeddings """ # Pickle embeddings should be AGNOSTIC to the name. This is because each pickled embedding is specific to the dataset and transformer. # Applies down the road when/if we attempt active learning data_name = config['train_file'].split('/')[-1][:-4] # retrieve file name without the extension train_embed_pkl_f = os.path.join(config['cache'], data_name + '_' + config['embedding_type'] + '_training_embeddings.p') valid_embed_pkl_f = os.path.join(config['cache'], data_name + '_' + config['embedding_type'] + '_validation_embeddings.p') if os.path.exists(train_embed_pkl_f): with open( train_embed_pkl_f, 'rb') as cache: train_embeddings = pickle.load(cache) with open(valid_embed_pkl_f, 'rb') as cache: valid_embeddings = pickle.load(cache) else: # get embeddings from scratch tokenizer = AutoTokenizer.from_pretrained(vocab) embedding_model = AbstractBert(vocab) if torch.cuda.device_count() > 1: print("GPUs Available: ", torch.cuda.device_count()) embedding_model = torch.nn.DataParallel(embedding_model, device_ids=[0, 1, 2]) use_cuda = torch.cuda.is_available() device = torch.device("cuda:0" if use_cuda else "cpu") embedding_model.eval().to(device) logger.info(' Getting BERT/ROBERTA embeddings...') train_embeddings = _get_bert_embeddings(training_generator, embedding_model, config["metadata"]) valid_embeddings = _get_bert_embeddings(validation_generator, embedding_model, config["metadata"]) # save embeddings pickle.dump(train_embeddings, open(train_embed_pkl_f, 'wb')) pickle.dump(valid_embeddings, open(valid_embed_pkl_f, 'wb')) logger.info(' Saved full BERT/ROBERTA embeddings.') embedding_shape = train_embeddings['embeddings'][1].shape[0] return embedding_shape, train_embeddings, valid_embeddings def _get_bert_embeddings(data_generator, embedding_model: torch.nn.Module, metadata: False): """Get BERT embeddings from a dataloader generator. Arguments: data_generator {data.Dataset} -- dataloader generator (AbstractDataset). embedding_model {torch.nn.Module} -- embedding model. Returns: embeddings {dict} -- dictionary containing ids, augmented embeddings, and labels. """ use_cuda = torch.cuda.is_available() device = torch.device("cuda:0" if use_cuda else "cpu") with torch.set_grad_enabled(False): embeddings = {'ids': [], 'embeddings': [], 'labels': [] } # get BERT training embeddings if metadata: for local_ids, local_data, local_meta, local_labels in data_generator: local_data, local_meta, local_labels = local_data.to(device).long().squeeze(1), \ local_meta, \ local_labels.to(device).long() #print(local_data[0].shape) augmented_embeddings = embedding_model(local_data, local_meta) embeddings['ids'].extend(np.array(local_ids)) embeddings['embeddings'].extend(np.array(augmented_embeddings.detach().cpu())) embeddings['labels'].extend(np.array(local_labels.detach().cpu().tolist())) else: for local_ids, local_data, local_labels in data_generator: local_data, local_labels = local_data.to(device).long().squeeze(1), \ local_labels.to(device).long() #print(local_data[0].shape) augmented_embeddings = embedding_model(local_data) embeddings['ids'].extend(np.array(local_ids)) embeddings['embeddings'].extend(np.array(augmented_embeddings.detach().cpu())) embeddings['labels'].extend(np.array(local_labels.detach().cpu().tolist())) return embeddings def get_pca_embeddings(config, name, training_embedding: dict, validation_embedding: dict): """Reduced embeddings using PCA. Args: training_embedding (dict) -- dictionary containing training embeddings validation_embedding (dict) -- dictionary containing validation embeddings Returns: generator -- Torch Dataloader tuple -- shape of embedding """ data_name = config['train_file'].split('/')[-1][:-4] # retrieve file name without the extension train_pca_pkl_f = os.path.join(config['pca_cache'], data_name + '_' + config['embedding_type'] + '_training_embeddings.p') valid_pca_pkl_f = os.path.join(config['pca_cache'], data_name + '_' + config['embedding_type'] + '_validation_embeddings.p') if os.path.exists(train_pca_pkl_f): logger.info(" Loading PCA-embeddings from cache ") with open(train_pca_pkl_f , 'rb') as cache: train_embeddings = pickle.load
'valid': ids[int(total_len * 0.7): ] } for i in dataset.iloc[:, unique_id_col]:
random_line_split
utils.py
} -- config file containing data paths and tokenizer information metadata {bool} -- whether the data contains metadata for augmented embeddings proportion {float} -- proportion for splitting up train and test. (default: {0.7}) max_len {int} -- maximum token length for a text. (default: {128}) partition {dict} -- maps lists of training and validation data IDs (default: {None}) labels {dict} -- (default: {None}) Returns: partition {dict} -- list of ids in train and valid datasets torch.utils.data.Dataset -- dataset """ # columns if meta: [0] unique ID, [1] text, [2] metadata, [3] label # columns if no meta: [0] unique ID, [1] text, [2] label if config["metadata"]: unique_id_col = 0 text_col = 1 metadata_col = 2 label_col = 3 else: unique_id_col = 0 text_col = 1 label_col = 3 dataset = pd.read_csv(config['train_file'], header=None, sep='\t') print(dataset) # below fix null values wrecking encode_plus # convert labels to integer and drop nas dataset.iloc[:, label_col] = pd.to_numeric(dataset.iloc[:, label_col], errors = 'coerce' ) dataset = dataset[~ dataset[text_col].isnull()] # recreate the first column with the reset index. dataset = dataset[(dataset.iloc[:, label_col] == 1) | (dataset.iloc[:, label_col] == 0)] \ .reset_index().reset_index().drop(columns = ['index', 0]).rename(columns = {'level_0': 0}) print(dataset) # create list of train/valid IDs if not provided if not partition and not labels: ids = list(dataset.iloc[:,unique_id_col]) total_len = len(ids) np.random.shuffle(ids) labels = {} # metadata = {} partition = {'train': ids[ :int(total_len * 0.7)], 'valid': ids[int(total_len * 0.7): ] } for i in dataset.iloc[:, unique_id_col]: labels[i] = dataset.iloc[i][label_col] # set parameters for DataLoader -- num_workers = cores params = {'batch_size': 32, 'shuffle': True, 'num_workers': 0 } tokenizer = AutoTokenizer.from_pretrained(vocab) dataset[text_col] = dataset[text_col].apply(lambda x: tokenizer.encode_plus(str(x), \ max_length=max_len, \ add_special_tokens=True, \ pad_to_max_length=True, \ truncation=True)) if config['metadata']: # glove for metadata preprocessing glove = torchtext.vocab.GloVe(name="6B", dim=50) dataset[metadata_col] = dataset[metadata_col].apply(lambda y: __pad__(str(y).split(" "), 30)) dataset[metadata_col] = dataset[metadata_col].apply(lambda z: __glove_embed__(z, glove)) train_data = dataset[dataset[unique_id_col].isin(partition['train'])] valid_data = dataset[dataset[unique_id_col].isin(partition['valid'])] # create train/valid generators training_set = AbstractDataset(data=train_data, labels=labels, metadata=config['metadata'], list_IDs=partition['train'], max_len = max_len) training_generator = DataLoader(training_set, **params) validation_set = AbstractDataset(data=valid_data, labels=labels, metadata=config['metadata'], list_IDs=partition['valid'],max_len = max_len) validation_generator = DataLoader(validation_set, **params) return partition, training_generator, validation_generator def __pad__(sequence, max_l): """ Padding function for 1D sequences """ if max_l - len(sequence) < 0:
else: sequence = np.pad(sequence, (0, max_l - (len(sequence))), 'constant', constant_values=(0)) return sequence def __glove_embed__(sequence, model): """ Embed words in a sequence using GLoVE model """ embedded = [] for word in sequence: embedded.append(model[word]) return embedded def load_embeddings(config, name, vocab, training_generator, validation_generator): """Load embeddings either from cache or from scratch Args: config (json) -- file configurations. name -- vocab -- training_generator -- validation_generator -- Returns: embedding_shape, train_embeddings, valid_embeddings """ # Pickle embeddings should be AGNOSTIC to the name. This is because each pickled embedding is specific to the dataset and transformer. # Applies down the road when/if we attempt active learning data_name = config['train_file'].split('/')[-1][:-4] # retrieve file name without the extension train_embed_pkl_f = os.path.join(config['cache'], data_name + '_' + config['embedding_type'] + '_training_embeddings.p') valid_embed_pkl_f = os.path.join(config['cache'], data_name + '_' + config['embedding_type'] + '_validation_embeddings.p') if os.path.exists(train_embed_pkl_f): with open( train_embed_pkl_f, 'rb') as cache: train_embeddings = pickle.load(cache) with open(valid_embed_pkl_f, 'rb') as cache: valid_embeddings = pickle.load(cache) else: # get embeddings from scratch tokenizer = AutoTokenizer.from_pretrained(vocab) embedding_model = AbstractBert(vocab) if torch.cuda.device_count() > 1: print("GPUs Available: ", torch.cuda.device_count()) embedding_model = torch.nn.DataParallel(embedding_model, device_ids=[0, 1, 2]) use_cuda = torch.cuda.is_available() device = torch.device("cuda:0" if use_cuda else "cpu") embedding_model.eval().to(device) logger.info(' Getting BERT/ROBERTA embeddings...') train_embeddings = _get_bert_embeddings(training_generator, embedding_model, config["metadata"]) valid_embeddings = _get_bert_embeddings(validation_generator, embedding_model, config["metadata"]) # save embeddings pickle.dump(train_embeddings, open(train_embed_pkl_f, 'wb')) pickle.dump(valid_embeddings, open(valid_embed_pkl_f, 'wb')) logger.info(' Saved full BERT/ROBERTA embeddings.') embedding_shape = train_embeddings['embeddings'][1].shape[0] return embedding_shape, train_embeddings, valid_embeddings def _get_bert_embeddings(data_generator, embedding_model: torch.nn.Module, metadata: False): """Get BERT embeddings from a dataloader generator. Arguments: data_generator {data.Dataset} -- dataloader generator (AbstractDataset). embedding_model {torch.nn.Module} -- embedding model. Returns: embeddings {dict} -- dictionary containing ids, augmented embeddings, and labels. """ use_cuda = torch.cuda.is_available() device = torch.device("cuda:0" if use_cuda else "cpu") with torch.set_grad_enabled(False): embeddings = {'ids': [], 'embeddings': [], 'labels': [] } # get BERT training embeddings if metadata: for local_ids, local_data, local_meta, local_labels in data_generator: local_data, local_meta, local_labels = local_data.to(device).long().squeeze(1), \ local_meta, \ local_labels.to(device).long() #print(local_data[0].shape) augmented_embeddings = embedding_model(local_data, local_meta) embeddings['ids'].extend(np.array(local_ids)) embeddings['embeddings'].extend(np.array(augmented_embeddings.detach().cpu())) embeddings['labels'].extend(np.array(local_labels.detach().cpu().tolist())) else: for local_ids, local_data, local_labels in data_generator: local_data, local_labels = local_data.to(device).long().squeeze(1), \ local_labels.to(device).long() #print(local_data[0].shape) augmented_embeddings = embedding_model(local_data) embeddings['ids'].extend(np.array(local_ids)) embeddings['embeddings'].extend(np.array(augmented_embeddings.detach().cpu())) embeddings['labels'].extend(np.array(local_labels.detach().cpu().tolist())) return embeddings def get_pca_embeddings(config, name, training_embedding: dict, validation_embedding: dict): """Reduced embeddings using PCA. Args: training_embedding (dict) -- dictionary containing training embeddings validation_embedding (dict) -- dictionary containing validation embeddings Returns: generator -- Torch Dataloader tuple -- shape of embedding """ data_name = config['train_file'].split('/')[-1][:-4] # retrieve file name without the extension train_pca_pkl_f = os.path.join(config['pca_cache'], data_name + '_' + config['embedding_type'] + '_training_embeddings.p') valid_pca_pkl_f = os.path.join(config['pca_cache'], data_name + '_' + config['embedding_type'] + '_validation_embeddings.p') if os.path.exists(train_pca_pkl_f): logger.info(" Loading PCA-embeddings from cache ") with open(train_pca_pkl_f , 'rb') as cache: train_embeddings = pickle
sequence = sequence[:max_l]
conditional_block
utils.py
=None, sep='\t') print(dataset) # below fix null values wrecking encode_plus # convert labels to integer and drop nas dataset.iloc[:, label_col] = pd.to_numeric(dataset.iloc[:, label_col], errors = 'coerce' ) dataset = dataset[~ dataset[text_col].isnull()] # recreate the first column with the reset index. dataset = dataset[(dataset.iloc[:, label_col] == 1) | (dataset.iloc[:, label_col] == 0)] \ .reset_index().reset_index().drop(columns = ['index', 0]).rename(columns = {'level_0': 0}) print(dataset) # create list of train/valid IDs if not provided if not partition and not labels: ids = list(dataset.iloc[:,unique_id_col]) total_len = len(ids) np.random.shuffle(ids) labels = {} # metadata = {} partition = {'train': ids[ :int(total_len * 0.7)], 'valid': ids[int(total_len * 0.7): ] } for i in dataset.iloc[:, unique_id_col]: labels[i] = dataset.iloc[i][label_col] # set parameters for DataLoader -- num_workers = cores params = {'batch_size': 32, 'shuffle': True, 'num_workers': 0 } tokenizer = AutoTokenizer.from_pretrained(vocab) dataset[text_col] = dataset[text_col].apply(lambda x: tokenizer.encode_plus(str(x), \ max_length=max_len, \ add_special_tokens=True, \ pad_to_max_length=True, \ truncation=True)) if config['metadata']: # glove for metadata preprocessing glove = torchtext.vocab.GloVe(name="6B", dim=50) dataset[metadata_col] = dataset[metadata_col].apply(lambda y: __pad__(str(y).split(" "), 30)) dataset[metadata_col] = dataset[metadata_col].apply(lambda z: __glove_embed__(z, glove)) train_data = dataset[dataset[unique_id_col].isin(partition['train'])] valid_data = dataset[dataset[unique_id_col].isin(partition['valid'])] # create train/valid generators training_set = AbstractDataset(data=train_data, labels=labels, metadata=config['metadata'], list_IDs=partition['train'], max_len = max_len) training_generator = DataLoader(training_set, **params) validation_set = AbstractDataset(data=valid_data, labels=labels, metadata=config['metadata'], list_IDs=partition['valid'],max_len = max_len) validation_generator = DataLoader(validation_set, **params) return partition, training_generator, validation_generator def __pad__(sequence, max_l): """ Padding function for 1D sequences """ if max_l - len(sequence) < 0: sequence = sequence[:max_l] else: sequence = np.pad(sequence, (0, max_l - (len(sequence))), 'constant', constant_values=(0)) return sequence def __glove_embed__(sequence, model): """ Embed words in a sequence using GLoVE model """ embedded = [] for word in sequence: embedded.append(model[word]) return embedded def load_embeddings(config, name, vocab, training_generator, validation_generator): """Load embeddings either from cache or from scratch Args: config (json) -- file configurations. name -- vocab -- training_generator -- validation_generator -- Returns: embedding_shape, train_embeddings, valid_embeddings """ # Pickle embeddings should be AGNOSTIC to the name. This is because each pickled embedding is specific to the dataset and transformer. # Applies down the road when/if we attempt active learning data_name = config['train_file'].split('/')[-1][:-4] # retrieve file name without the extension train_embed_pkl_f = os.path.join(config['cache'], data_name + '_' + config['embedding_type'] + '_training_embeddings.p') valid_embed_pkl_f = os.path.join(config['cache'], data_name + '_' + config['embedding_type'] + '_validation_embeddings.p') if os.path.exists(train_embed_pkl_f): with open( train_embed_pkl_f, 'rb') as cache: train_embeddings = pickle.load(cache) with open(valid_embed_pkl_f, 'rb') as cache: valid_embeddings = pickle.load(cache) else: # get embeddings from scratch tokenizer = AutoTokenizer.from_pretrained(vocab) embedding_model = AbstractBert(vocab) if torch.cuda.device_count() > 1: print("GPUs Available: ", torch.cuda.device_count()) embedding_model = torch.nn.DataParallel(embedding_model, device_ids=[0, 1, 2]) use_cuda = torch.cuda.is_available() device = torch.device("cuda:0" if use_cuda else "cpu") embedding_model.eval().to(device) logger.info(' Getting BERT/ROBERTA embeddings...') train_embeddings = _get_bert_embeddings(training_generator, embedding_model, config["metadata"]) valid_embeddings = _get_bert_embeddings(validation_generator, embedding_model, config["metadata"]) # save embeddings pickle.dump(train_embeddings, open(train_embed_pkl_f, 'wb')) pickle.dump(valid_embeddings, open(valid_embed_pkl_f, 'wb')) logger.info(' Saved full BERT/ROBERTA embeddings.') embedding_shape = train_embeddings['embeddings'][1].shape[0] return embedding_shape, train_embeddings, valid_embeddings def _get_bert_embeddings(data_generator, embedding_model: torch.nn.Module, metadata: False): """Get BERT embeddings from a dataloader generator. Arguments: data_generator {data.Dataset} -- dataloader generator (AbstractDataset). embedding_model {torch.nn.Module} -- embedding model. Returns: embeddings {dict} -- dictionary containing ids, augmented embeddings, and labels. """ use_cuda = torch.cuda.is_available() device = torch.device("cuda:0" if use_cuda else "cpu") with torch.set_grad_enabled(False): embeddings = {'ids': [], 'embeddings': [], 'labels': [] } # get BERT training embeddings if metadata: for local_ids, local_data, local_meta, local_labels in data_generator: local_data, local_meta, local_labels = local_data.to(device).long().squeeze(1), \ local_meta, \ local_labels.to(device).long() #print(local_data[0].shape) augmented_embeddings = embedding_model(local_data, local_meta) embeddings['ids'].extend(np.array(local_ids)) embeddings['embeddings'].extend(np.array(augmented_embeddings.detach().cpu())) embeddings['labels'].extend(np.array(local_labels.detach().cpu().tolist())) else: for local_ids, local_data, local_labels in data_generator: local_data, local_labels = local_data.to(device).long().squeeze(1), \ local_labels.to(device).long() #print(local_data[0].shape) augmented_embeddings = embedding_model(local_data) embeddings['ids'].extend(np.array(local_ids)) embeddings['embeddings'].extend(np.array(augmented_embeddings.detach().cpu())) embeddings['labels'].extend(np.array(local_labels.detach().cpu().tolist())) return embeddings def get_pca_embeddings(config, name, training_embedding: dict, validation_embedding: dict): """Reduced embeddings using PCA. Args: training_embedding (dict) -- dictionary containing training embeddings validation_embedding (dict) -- dictionary containing validation embeddings Returns: generator -- Torch Dataloader tuple -- shape of embedding """ data_name = config['train_file'].split('/')[-1][:-4] # retrieve file name without the extension train_pca_pkl_f = os.path.join(config['pca_cache'], data_name + '_' + config['embedding_type'] + '_training_embeddings.p') valid_pca_pkl_f = os.path.join(config['pca_cache'], data_name + '_' + config['embedding_type'] + '_validation_embeddings.p') if os.path.exists(train_pca_pkl_f): logger.info(" Loading PCA-embeddings from cache ") with open(train_pca_pkl_f , 'rb') as cache: train_embeddings = pickle.load(cache) with open(valid_pca_pkl_f, 'rb') as cache: valid_embeddings = pickle.load(cache) else: logger.info(' Standardizing ') ss = StandardScaler() train_embed_ss = ss.fit_transform(training_embedding['embeddings']) valid_embed_ss = ss.transform(validation_embedding['embeddings']) # Dimension reduction: PCA or UMAP (?) logger.info(' Doing PCA...') pca_model = decomposition.PCA(n_components = 0.90) # this can be a parameter down the road, but for debugging it's fine train_reduc = pca_model.fit_transform(train_embed_ss) val_reduc = pca_model.transform(valid_embed_ss) training_embedding['embeddings'] = train_reduc validation_embedding['embeddings'] = val_reduc train_embeddings = training_embedding.copy() valid_embeddings = validation_embedding.copy() # save embeddings pickle.dump(train_embeddings, open(train_pca_pkl_f, 'wb')) pickle.dump(valid_embeddings, open(valid_pca_pkl_f, 'wb')) embedding_shape = len(train_embeddings['embeddings'][0]) return embedding_shape, train_embeddings, valid_embeddings def
metrics
identifier_name
utils.py
} -- config file containing data paths and tokenizer information metadata {bool} -- whether the data contains metadata for augmented embeddings proportion {float} -- proportion for splitting up train and test. (default: {0.7}) max_len {int} -- maximum token length for a text. (default: {128}) partition {dict} -- maps lists of training and validation data IDs (default: {None}) labels {dict} -- (default: {None}) Returns: partition {dict} -- list of ids in train and valid datasets torch.utils.data.Dataset -- dataset """ # columns if meta: [0] unique ID, [1] text, [2] metadata, [3] label # columns if no meta: [0] unique ID, [1] text, [2] label if config["metadata"]: unique_id_col = 0 text_col = 1 metadata_col = 2 label_col = 3 else: unique_id_col = 0 text_col = 1 label_col = 3 dataset = pd.read_csv(config['train_file'], header=None, sep='\t') print(dataset) # below fix null values wrecking encode_plus # convert labels to integer and drop nas dataset.iloc[:, label_col] = pd.to_numeric(dataset.iloc[:, label_col], errors = 'coerce' ) dataset = dataset[~ dataset[text_col].isnull()] # recreate the first column with the reset index. dataset = dataset[(dataset.iloc[:, label_col] == 1) | (dataset.iloc[:, label_col] == 0)] \ .reset_index().reset_index().drop(columns = ['index', 0]).rename(columns = {'level_0': 0}) print(dataset) # create list of train/valid IDs if not provided if not partition and not labels: ids = list(dataset.iloc[:,unique_id_col]) total_len = len(ids) np.random.shuffle(ids) labels = {} # metadata = {} partition = {'train': ids[ :int(total_len * 0.7)], 'valid': ids[int(total_len * 0.7): ] } for i in dataset.iloc[:, unique_id_col]: labels[i] = dataset.iloc[i][label_col] # set parameters for DataLoader -- num_workers = cores params = {'batch_size': 32, 'shuffle': True, 'num_workers': 0 } tokenizer = AutoTokenizer.from_pretrained(vocab) dataset[text_col] = dataset[text_col].apply(lambda x: tokenizer.encode_plus(str(x), \ max_length=max_len, \ add_special_tokens=True, \ pad_to_max_length=True, \ truncation=True)) if config['metadata']: # glove for metadata preprocessing glove = torchtext.vocab.GloVe(name="6B", dim=50) dataset[metadata_col] = dataset[metadata_col].apply(lambda y: __pad__(str(y).split(" "), 30)) dataset[metadata_col] = dataset[metadata_col].apply(lambda z: __glove_embed__(z, glove)) train_data = dataset[dataset[unique_id_col].isin(partition['train'])] valid_data = dataset[dataset[unique_id_col].isin(partition['valid'])] # create train/valid generators training_set = AbstractDataset(data=train_data, labels=labels, metadata=config['metadata'], list_IDs=partition['train'], max_len = max_len) training_generator = DataLoader(training_set, **params) validation_set = AbstractDataset(data=valid_data, labels=labels, metadata=config['metadata'], list_IDs=partition['valid'],max_len = max_len) validation_generator = DataLoader(validation_set, **params) return partition, training_generator, validation_generator def __pad__(sequence, max_l):
def __glove_embed__(sequence, model): """ Embed words in a sequence using GLoVE model """ embedded = [] for word in sequence: embedded.append(model[word]) return embedded def load_embeddings(config, name, vocab, training_generator, validation_generator): """Load embeddings either from cache or from scratch Args: config (json) -- file configurations. name -- vocab -- training_generator -- validation_generator -- Returns: embedding_shape, train_embeddings, valid_embeddings """ # Pickle embeddings should be AGNOSTIC to the name. This is because each pickled embedding is specific to the dataset and transformer. # Applies down the road when/if we attempt active learning data_name = config['train_file'].split('/')[-1][:-4] # retrieve file name without the extension train_embed_pkl_f = os.path.join(config['cache'], data_name + '_' + config['embedding_type'] + '_training_embeddings.p') valid_embed_pkl_f = os.path.join(config['cache'], data_name + '_' + config['embedding_type'] + '_validation_embeddings.p') if os.path.exists(train_embed_pkl_f): with open( train_embed_pkl_f, 'rb') as cache: train_embeddings = pickle.load(cache) with open(valid_embed_pkl_f, 'rb') as cache: valid_embeddings = pickle.load(cache) else: # get embeddings from scratch tokenizer = AutoTokenizer.from_pretrained(vocab) embedding_model = AbstractBert(vocab) if torch.cuda.device_count() > 1: print("GPUs Available: ", torch.cuda.device_count()) embedding_model = torch.nn.DataParallel(embedding_model, device_ids=[0, 1, 2]) use_cuda = torch.cuda.is_available() device = torch.device("cuda:0" if use_cuda else "cpu") embedding_model.eval().to(device) logger.info(' Getting BERT/ROBERTA embeddings...') train_embeddings = _get_bert_embeddings(training_generator, embedding_model, config["metadata"]) valid_embeddings = _get_bert_embeddings(validation_generator, embedding_model, config["metadata"]) # save embeddings pickle.dump(train_embeddings, open(train_embed_pkl_f, 'wb')) pickle.dump(valid_embeddings, open(valid_embed_pkl_f, 'wb')) logger.info(' Saved full BERT/ROBERTA embeddings.') embedding_shape = train_embeddings['embeddings'][1].shape[0] return embedding_shape, train_embeddings, valid_embeddings def _get_bert_embeddings(data_generator, embedding_model: torch.nn.Module, metadata: False): """Get BERT embeddings from a dataloader generator. Arguments: data_generator {data.Dataset} -- dataloader generator (AbstractDataset). embedding_model {torch.nn.Module} -- embedding model. Returns: embeddings {dict} -- dictionary containing ids, augmented embeddings, and labels. """ use_cuda = torch.cuda.is_available() device = torch.device("cuda:0" if use_cuda else "cpu") with torch.set_grad_enabled(False): embeddings = {'ids': [], 'embeddings': [], 'labels': [] } # get BERT training embeddings if metadata: for local_ids, local_data, local_meta, local_labels in data_generator: local_data, local_meta, local_labels = local_data.to(device).long().squeeze(1), \ local_meta, \ local_labels.to(device).long() #print(local_data[0].shape) augmented_embeddings = embedding_model(local_data, local_meta) embeddings['ids'].extend(np.array(local_ids)) embeddings['embeddings'].extend(np.array(augmented_embeddings.detach().cpu())) embeddings['labels'].extend(np.array(local_labels.detach().cpu().tolist())) else: for local_ids, local_data, local_labels in data_generator: local_data, local_labels = local_data.to(device).long().squeeze(1), \ local_labels.to(device).long() #print(local_data[0].shape) augmented_embeddings = embedding_model(local_data) embeddings['ids'].extend(np.array(local_ids)) embeddings['embeddings'].extend(np.array(augmented_embeddings.detach().cpu())) embeddings['labels'].extend(np.array(local_labels.detach().cpu().tolist())) return embeddings def get_pca_embeddings(config, name, training_embedding: dict, validation_embedding: dict): """Reduced embeddings using PCA. Args: training_embedding (dict) -- dictionary containing training embeddings validation_embedding (dict) -- dictionary containing validation embeddings Returns: generator -- Torch Dataloader tuple -- shape of embedding """ data_name = config['train_file'].split('/')[-1][:-4] # retrieve file name without the extension train_pca_pkl_f = os.path.join(config['pca_cache'], data_name + '_' + config['embedding_type'] + '_training_embeddings.p') valid_pca_pkl_f = os.path.join(config['pca_cache'], data_name + '_' + config['embedding_type'] + '_validation_embeddings.p') if os.path.exists(train_pca_pkl_f): logger.info(" Loading PCA-embeddings from cache ") with open(train_pca_pkl_f , 'rb') as cache: train_embeddings = pickle.load
""" Padding function for 1D sequences """ if max_l - len(sequence) < 0: sequence = sequence[:max_l] else: sequence = np.pad(sequence, (0, max_l - (len(sequence))), 'constant', constant_values=(0)) return sequence
identifier_body
ioapic.rs
32 { self.read_reg(0x01) } pub fn read_ioapicarb(&mut self) -> u32 { self.read_reg(0x02) } pub fn read_ioredtbl(&mut self, idx: u8) -> u64 { assert!(idx < 24); let lo = self.read_reg(0x10 + idx * 2); let hi = self.read_reg(0x10 + idx * 2 + 1); u64::from(lo) | (u64::from(hi) << 32) } pub fn write_ioredtbl(&mut self, idx: u8, value: u64) { assert!(idx < 24); let lo = value as u32; let hi = (value >> 32) as u32; self.write_reg(0x10 + idx * 2, lo); self.write_reg(0x10 + idx * 2 + 1, hi); } pub fn max_redirection_table_entries(&mut self) -> u8 { let ver = self.read_ioapicver(); ((ver & 0x00FF_0000) >> 16) as u8 } pub fn id(&mut self) -> u8 { let id_reg = self.read_ioapicid(); ((id_reg & 0x0F00_0000) >> 24) as u8 } } pub struct IoApic { regs: Mutex<IoApicRegs>, gsi_start: u32, count: u8, } impl IoApic { pub fn new(regs_base: *const u32, gsi_start: u32) -> Self { let mut regs = IoApicRegs { pointer: regs_base }; let count = regs.max_redirection_table_entries(); Self { regs: Mutex::new(regs), gsi_start, count, } } /// Map an interrupt vector to a physical local APIC ID of a processor (thus physical mode). pub fn map(&self, idx: u8, info: MapInfo) { self.regs.lock().write_ioredtbl(idx, info.as_raw()) } pub fn set_mask(&self, gsi: u32, mask: bool) { let idx = (gsi - self.gsi_start) as u8; let mut guard = self.regs.lock(); let mut reg = guard.read_ioredtbl(idx); reg &= !(1 << 16); reg |= u64::from(mask) << 16; guard.write_ioredtbl(idx, reg); } } #[repr(u8)] #[derive(Clone, Copy, Debug)] pub enum ApicTriggerMode { Edge = 0, Level = 1, } #[repr(u8)] #[derive(Clone, Copy, Debug)] pub enum ApicPolarity { ActiveHigh = 0, ActiveLow = 1, } #[repr(u8)] #[derive(Clone, Copy, Debug)] pub enum DestinationMode { Physical = 0, Logical = 1, } #[repr(u8)] #[derive(Clone, Copy, Debug)] pub enum DeliveryMode { Fixed = 0b000, LowestPriority = 0b001, Smi = 0b010, Nmi = 0b100, Init = 0b101, ExtInt = 0b111, } #[derive(Clone, Copy, Debug)] pub struct MapInfo { pub dest: u8, pub mask: bool, pub trigger_mode: ApicTriggerMode, pub polarity: ApicPolarity, pub dest_mode: DestinationMode, pub delivery_mode: DeliveryMode, pub vector: u8, } impl MapInfo { pub fn as_raw(&self) -> u64 { assert!(self.vector >= 0x20); assert!(self.vector <= 0xFE); // TODO: Check for reserved fields. (u64::from(self.dest) << 56) | (u64::from(self.mask) << 16) | ((self.trigger_mode as u64) << 15) | ((self.polarity as u64) << 13) | ((self.dest_mode as u64) << 11) | ((self.delivery_mode as u64) << 8) | u64::from(self.vector) }
} impl fmt::Debug for IoApic { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { struct RedirTable<'a>(&'a Mutex<IoApicRegs>); impl<'a> fmt::Debug for RedirTable<'a> { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { let mut guard = self.0.lock(); let count = guard.max_redirection_table_entries(); f.debug_list().entries((0..count).map(|i| guard.read_ioredtbl(i))).finish() } } f.debug_struct("IoApic") .field("redir_table", &RedirTable(&self.regs)) .field("gsi_start", &self.gsi_start) .field("count", &self.count) .finish() } } #[derive(Clone, Copy, Debug)] pub enum TriggerMode { ConformsToSpecs, Edge, Level, } #[derive(Clone, Copy, Debug)] pub enum Polarity { ConformsToSpecs, ActiveHigh, ActiveLow, } #[derive(Clone, Copy, Debug)] pub struct Override { bus_irq: u8, gsi: u32, trigger_mode: TriggerMode, polarity: Polarity, } // static mut because only the AP initializes the I/O Apic, and when that is done, it's solely // accessed immutably. static mut IOAPICS: Option<Vec<IoApic>> = None; // static mut for the same reason as above static mut SRC_OVERRIDES: Option<Vec<Override>> = None; pub fn ioapics() -> &'static [IoApic] { unsafe { IOAPICS.as_ref().map_or(&[], |vector| &vector[..]) } } pub fn src_overrides() -> &'static [Override] { unsafe { SRC_OVERRIDES.as_ref().map_or(&[], |vector| &vector[..]) } } #[cfg(feature = "acpi")] pub unsafe fn handle_ioapic(mapper: &mut KernelMapper, madt_ioapic: &'static MadtIoApic) { // map the I/O APIC registers let frame = Frame::containing_address(PhysicalAddress::new(madt_ioapic.address as usize)); let page = Page::containing_address(VirtualAddress::new(crate::IOAPIC_OFFSET)); assert!(mapper.translate(page.start_address()).is_none()); mapper .get_mut() .expect("expected KernelMapper not to be locked re-entrant while mapping I/O APIC memory") .map_phys(page.start_address(), frame.start_address(), PageFlags::new().write(true).custom_flag(EntryFlags::NO_CACHE.bits(), true)) .expect("failed to map I/O APIC") .flush(); let ioapic_registers = page.start_address().data() as *const u32; let ioapic = IoApic::new(ioapic_registers, madt_ioapic.gsi_base); assert_eq!(ioapic.regs.lock().id(), madt_ioapic.id, "mismatched ACPI MADT I/O APIC ID, and the ID reported by the I/O APIC"); IOAPICS.get_or_insert_with(Vec::new).push(ioapic); } #[cfg(feature = "acpi")] pub unsafe fn handle_src_override(src_override: &'static MadtIntSrcOverride) { let flags = src_override.flags; let polarity_raw = (flags & 0x0003) as u8; let trigger_mode_raw = ((flags & 0x000C) >> 2) as u8; let polarity = match polarity_raw { 0b00 => Polarity::ConformsToSpecs, 0b01 => Polarity::ActiveHigh, 0b10 => return, // reserved 0b11 => Polarity::ActiveLow, _ => unreachable!(), }; let trigger_mode = match trigger_mode_raw { 0b00 => TriggerMode::ConformsToSpecs, 0b01 => TriggerMode::Edge, 0b10 => return, // reserved 0b11 => TriggerMode::Level, _ => unreachable!(), }; let over = Override { bus_irq: src_override.irq_source, gsi: src_override.gsi_base, polarity, trigger_mode, }; SRC_OVERRIDES.get_or_insert_with(Vec::new).push(over); } pub unsafe fn init(active_table: &mut KernelMapper) { let bsp_apic_id = cpuid().unwrap().get_feature_info().unwrap().initial_local_apic_id(); // TODO
random_line_split
ioapic.rs
(self.mask) << 16) | ((self.trigger_mode as u64) << 15) | ((self.polarity as u64) << 13) | ((self.dest_mode as u64) << 11) | ((self.delivery_mode as u64) << 8) | u64::from(self.vector) } } impl fmt::Debug for IoApic { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { struct RedirTable<'a>(&'a Mutex<IoApicRegs>); impl<'a> fmt::Debug for RedirTable<'a> { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { let mut guard = self.0.lock(); let count = guard.max_redirection_table_entries(); f.debug_list().entries((0..count).map(|i| guard.read_ioredtbl(i))).finish() } } f.debug_struct("IoApic") .field("redir_table", &RedirTable(&self.regs)) .field("gsi_start", &self.gsi_start) .field("count", &self.count) .finish() } } #[derive(Clone, Copy, Debug)] pub enum TriggerMode { ConformsToSpecs, Edge, Level, } #[derive(Clone, Copy, Debug)] pub enum Polarity { ConformsToSpecs, ActiveHigh, ActiveLow, } #[derive(Clone, Copy, Debug)] pub struct Override { bus_irq: u8, gsi: u32, trigger_mode: TriggerMode, polarity: Polarity, } // static mut because only the AP initializes the I/O Apic, and when that is done, it's solely // accessed immutably. static mut IOAPICS: Option<Vec<IoApic>> = None; // static mut for the same reason as above static mut SRC_OVERRIDES: Option<Vec<Override>> = None; pub fn ioapics() -> &'static [IoApic] { unsafe { IOAPICS.as_ref().map_or(&[], |vector| &vector[..]) } } pub fn src_overrides() -> &'static [Override] { unsafe { SRC_OVERRIDES.as_ref().map_or(&[], |vector| &vector[..]) } } #[cfg(feature = "acpi")] pub unsafe fn handle_ioapic(mapper: &mut KernelMapper, madt_ioapic: &'static MadtIoApic) { // map the I/O APIC registers let frame = Frame::containing_address(PhysicalAddress::new(madt_ioapic.address as usize)); let page = Page::containing_address(VirtualAddress::new(crate::IOAPIC_OFFSET)); assert!(mapper.translate(page.start_address()).is_none()); mapper .get_mut() .expect("expected KernelMapper not to be locked re-entrant while mapping I/O APIC memory") .map_phys(page.start_address(), frame.start_address(), PageFlags::new().write(true).custom_flag(EntryFlags::NO_CACHE.bits(), true)) .expect("failed to map I/O APIC") .flush(); let ioapic_registers = page.start_address().data() as *const u32; let ioapic = IoApic::new(ioapic_registers, madt_ioapic.gsi_base); assert_eq!(ioapic.regs.lock().id(), madt_ioapic.id, "mismatched ACPI MADT I/O APIC ID, and the ID reported by the I/O APIC"); IOAPICS.get_or_insert_with(Vec::new).push(ioapic); } #[cfg(feature = "acpi")] pub unsafe fn handle_src_override(src_override: &'static MadtIntSrcOverride) { let flags = src_override.flags; let polarity_raw = (flags & 0x0003) as u8; let trigger_mode_raw = ((flags & 0x000C) >> 2) as u8; let polarity = match polarity_raw { 0b00 => Polarity::ConformsToSpecs, 0b01 => Polarity::ActiveHigh, 0b10 => return, // reserved 0b11 => Polarity::ActiveLow, _ => unreachable!(), }; let trigger_mode = match trigger_mode_raw { 0b00 => TriggerMode::ConformsToSpecs, 0b01 => TriggerMode::Edge, 0b10 => return, // reserved 0b11 => TriggerMode::Level, _ => unreachable!(), }; let over = Override { bus_irq: src_override.irq_source, gsi: src_override.gsi_base, polarity, trigger_mode, }; SRC_OVERRIDES.get_or_insert_with(Vec::new).push(over); } pub unsafe fn init(active_table: &mut KernelMapper) { let bsp_apic_id = cpuid().unwrap().get_feature_info().unwrap().initial_local_apic_id(); // TODO: remove unwraps // search the madt for all IOAPICs. #[cfg(feature = "acpi")] { let madt: &'static Madt = match madt::MADT.as_ref() { Some(m) => m, // TODO: Parse MP tables too. None => return, }; if madt.flags & madt::FLAG_PCAT != 0 { pic::disable(); } // find all I/O APICs (usually one). for entry in madt.iter() { match entry { MadtEntry::IoApic(ioapic) => handle_ioapic(active_table, ioapic), MadtEntry::IntSrcOverride(src_override) => handle_src_override(src_override), _ => (), } } } println!("I/O APICs: {:?}, overrides: {:?}", ioapics(), src_overrides()); // map the legacy PC-compatible IRQs (0-15) to 32-47, just like we did with 8259 PIC (if it // wouldn't have been disabled due to this I/O APIC) for legacy_irq in 0..=15 { let (gsi, trigger_mode, polarity) = match get_override(legacy_irq) { Some(over) => (over.gsi, over.trigger_mode, over.polarity), None => { if src_overrides().iter().any(|over| over.gsi == u32::from(legacy_irq) && over.bus_irq != legacy_irq) && !src_overrides().iter().any(|over| over.bus_irq == legacy_irq) { // there's an IRQ conflict, making this legacy IRQ inaccessible. continue; } (legacy_irq.into(), TriggerMode::ConformsToSpecs, Polarity::ConformsToSpecs) } }; let apic = match find_ioapic(gsi) { Some(ioapic) => ioapic, None => { println!("Unable to find a suitable APIC for legacy IRQ {} (GSI {}). It will not be mapped.", legacy_irq, gsi); continue; } }; let redir_tbl_index = (gsi - apic.gsi_start) as u8; let map_info = MapInfo { // only send to the BSP dest: bsp_apic_id, dest_mode: DestinationMode::Physical, delivery_mode: DeliveryMode::Fixed, mask: false, polarity: match polarity { Polarity::ActiveHigh => ApicPolarity::ActiveHigh, Polarity::ActiveLow => ApicPolarity::ActiveLow, Polarity::ConformsToSpecs => ApicPolarity::ActiveHigh, }, trigger_mode: match trigger_mode { TriggerMode::Edge => ApicTriggerMode::Edge, TriggerMode::Level => ApicTriggerMode::Level, TriggerMode::ConformsToSpecs => ApicTriggerMode::Edge, }, vector: 32 + legacy_irq, }; apic.map(redir_tbl_index, map_info); } println!("I/O APICs: {:?}, overrides: {:?}", ioapics(), src_overrides()); irq::set_irq_method(irq::IrqMethod::Apic); // tell the firmware that we're using APIC rather than the default 8259 PIC. // FIXME: With ACPI moved to userspace, we should instead allow userspace to check whether the // IOAPIC has been initialized, and then subsequently let some ACPI driver call the AML from // userspace. /*#[cfg(feature = "acpi")] { let method = { let namespace_guard = crate::acpi::ACPI_TABLE.namespace.read(); if let Some(value) = namespace_guard.as_ref().unwrap().get("\\_PIC") { value.get_as_method().ok() } else { None } }; if let Some(m) = method { m.execute("\\_PIC".into(), vec!(crate::acpi::aml::AmlValue::Integer(1))); } }*/ } fn get_override(irq: u8) -> Option<&'static Override>
{ src_overrides().iter().find(|over| over.bus_irq == irq) }
identifier_body
ioapic.rs
32 { self.read_reg(0x01) } pub fn read_ioapicarb(&mut self) -> u32 { self.read_reg(0x02) } pub fn read_ioredtbl(&mut self, idx: u8) -> u64 { assert!(idx < 24); let lo = self.read_reg(0x10 + idx * 2); let hi = self.read_reg(0x10 + idx * 2 + 1); u64::from(lo) | (u64::from(hi) << 32) } pub fn write_ioredtbl(&mut self, idx: u8, value: u64) { assert!(idx < 24); let lo = value as u32; let hi = (value >> 32) as u32; self.write_reg(0x10 + idx * 2, lo); self.write_reg(0x10 + idx * 2 + 1, hi); } pub fn max_redirection_table_entries(&mut self) -> u8 { let ver = self.read_ioapicver(); ((ver & 0x00FF_0000) >> 16) as u8 } pub fn id(&mut self) -> u8 { let id_reg = self.read_ioapicid(); ((id_reg & 0x0F00_0000) >> 24) as u8 } } pub struct IoApic { regs: Mutex<IoApicRegs>, gsi_start: u32, count: u8, } impl IoApic { pub fn new(regs_base: *const u32, gsi_start: u32) -> Self { let mut regs = IoApicRegs { pointer: regs_base }; let count = regs.max_redirection_table_entries(); Self { regs: Mutex::new(regs), gsi_start, count, } } /// Map an interrupt vector to a physical local APIC ID of a processor (thus physical mode). pub fn map(&self, idx: u8, info: MapInfo) { self.regs.lock().write_ioredtbl(idx, info.as_raw()) } pub fn set_mask(&self, gsi: u32, mask: bool) { let idx = (gsi - self.gsi_start) as u8; let mut guard = self.regs.lock(); let mut reg = guard.read_ioredtbl(idx); reg &= !(1 << 16); reg |= u64::from(mask) << 16; guard.write_ioredtbl(idx, reg); } } #[repr(u8)] #[derive(Clone, Copy, Debug)] pub enum ApicTriggerMode { Edge = 0, Level = 1, } #[repr(u8)] #[derive(Clone, Copy, Debug)] pub enum ApicPolarity { ActiveHigh = 0, ActiveLow = 1, } #[repr(u8)] #[derive(Clone, Copy, Debug)] pub enum DestinationMode { Physical = 0, Logical = 1, } #[repr(u8)] #[derive(Clone, Copy, Debug)] pub enum DeliveryMode { Fixed = 0b000, LowestPriority = 0b001, Smi = 0b010, Nmi = 0b100, Init = 0b101, ExtInt = 0b111, } #[derive(Clone, Copy, Debug)] pub struct MapInfo { pub dest: u8, pub mask: bool, pub trigger_mode: ApicTriggerMode, pub polarity: ApicPolarity, pub dest_mode: DestinationMode, pub delivery_mode: DeliveryMode, pub vector: u8, } impl MapInfo { pub fn as_raw(&self) -> u64 { assert!(self.vector >= 0x20); assert!(self.vector <= 0xFE); // TODO: Check for reserved fields. (u64::from(self.dest) << 56) | (u64::from(self.mask) << 16) | ((self.trigger_mode as u64) << 15) | ((self.polarity as u64) << 13) | ((self.dest_mode as u64) << 11) | ((self.delivery_mode as u64) << 8) | u64::from(self.vector) } } impl fmt::Debug for IoApic { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { struct
<'a>(&'a Mutex<IoApicRegs>); impl<'a> fmt::Debug for RedirTable<'a> { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { let mut guard = self.0.lock(); let count = guard.max_redirection_table_entries(); f.debug_list().entries((0..count).map(|i| guard.read_ioredtbl(i))).finish() } } f.debug_struct("IoApic") .field("redir_table", &RedirTable(&self.regs)) .field("gsi_start", &self.gsi_start) .field("count", &self.count) .finish() } } #[derive(Clone, Copy, Debug)] pub enum TriggerMode { ConformsToSpecs, Edge, Level, } #[derive(Clone, Copy, Debug)] pub enum Polarity { ConformsToSpecs, ActiveHigh, ActiveLow, } #[derive(Clone, Copy, Debug)] pub struct Override { bus_irq: u8, gsi: u32, trigger_mode: TriggerMode, polarity: Polarity, } // static mut because only the AP initializes the I/O Apic, and when that is done, it's solely // accessed immutably. static mut IOAPICS: Option<Vec<IoApic>> = None; // static mut for the same reason as above static mut SRC_OVERRIDES: Option<Vec<Override>> = None; pub fn ioapics() -> &'static [IoApic] { unsafe { IOAPICS.as_ref().map_or(&[], |vector| &vector[..]) } } pub fn src_overrides() -> &'static [Override] { unsafe { SRC_OVERRIDES.as_ref().map_or(&[], |vector| &vector[..]) } } #[cfg(feature = "acpi")] pub unsafe fn handle_ioapic(mapper: &mut KernelMapper, madt_ioapic: &'static MadtIoApic) { // map the I/O APIC registers let frame = Frame::containing_address(PhysicalAddress::new(madt_ioapic.address as usize)); let page = Page::containing_address(VirtualAddress::new(crate::IOAPIC_OFFSET)); assert!(mapper.translate(page.start_address()).is_none()); mapper .get_mut() .expect("expected KernelMapper not to be locked re-entrant while mapping I/O APIC memory") .map_phys(page.start_address(), frame.start_address(), PageFlags::new().write(true).custom_flag(EntryFlags::NO_CACHE.bits(), true)) .expect("failed to map I/O APIC") .flush(); let ioapic_registers = page.start_address().data() as *const u32; let ioapic = IoApic::new(ioapic_registers, madt_ioapic.gsi_base); assert_eq!(ioapic.regs.lock().id(), madt_ioapic.id, "mismatched ACPI MADT I/O APIC ID, and the ID reported by the I/O APIC"); IOAPICS.get_or_insert_with(Vec::new).push(ioapic); } #[cfg(feature = "acpi")] pub unsafe fn handle_src_override(src_override: &'static MadtIntSrcOverride) { let flags = src_override.flags; let polarity_raw = (flags & 0x0003) as u8; let trigger_mode_raw = ((flags & 0x000C) >> 2) as u8; let polarity = match polarity_raw { 0b00 => Polarity::ConformsToSpecs, 0b01 => Polarity::ActiveHigh, 0b10 => return, // reserved 0b11 => Polarity::ActiveLow, _ => unreachable!(), }; let trigger_mode = match trigger_mode_raw { 0b00 => TriggerMode::ConformsToSpecs, 0b01 => TriggerMode::Edge, 0b10 => return, // reserved 0b11 => TriggerMode::Level, _ => unreachable!(), }; let over = Override { bus_irq: src_override.irq_source, gsi: src_override.gsi_base, polarity, trigger_mode, }; SRC_OVERRIDES.get_or_insert_with(Vec::new).push(over); } pub unsafe fn init(active_table: &mut KernelMapper) { let bsp_apic_id = cpuid().unwrap().get_feature_info().unwrap().initial_local_apic_id(); //
RedirTable
identifier_name
handler.go
existing one. CreateServiceInstance(*servicecatalog.Instance) (*servicecatalog.Instance, error) // CreateServiceBinding takes in a (possibly incomplete) // ServiceBinding and will either create or update an // existing one. CreateServiceBinding(*servicecatalog.Binding) (*servicecatalog.Binding, error) DeleteServiceBinding(*servicecatalog.Binding) error // CreateServiceBroker takes in a (possibly incomplete) // ServiceBroker and will either create or update an // existing one. CreateServiceBroker(*servicecatalog.Broker) (*servicecatalog.Broker, error) } type handler struct { k8sClient kubernetes.Interface apiClient apiclient.APIClient injector injector.BindingInjector newClientFunc func(name, url, username, password string) brokerapi.BrokerClient } func createHandler(k8sClient kubernetes.Interface, client apiclient.APIClient, injector injector.BindingInjector, newClientFn brokerapi.CreateFunc) *handler { return &handler{ k8sClient: k8sClient, apiClient: client, injector: injector, newClientFunc: newClientFn, } } func (h *handler) updateServiceInstance(in *servicecatalog.Instance) error { // Currently there's no difference between create / update, // but for prepping for future, split these into two different // methods for now. return h.createServiceInstance(in) } func (h *handler) createServiceInstance(in *servicecatalog.Instance) error { broker, err := util.GetBrokerByServiceClassName(h.apiClient.Brokers(), h.apiClient.ServiceClasses(), in.Spec.ServiceClassName) if err != nil { return err } authUsername, authPassword, err := GetAuthCredentialsFromBroker(h.k8sClient, broker) if err != nil { return err } client := h.newClientFunc(broker.Name, broker.Spec.URL, authUsername, authPassword) parameters := make(map[string]interface{}) if in.Spec.Parameters != nil && len(in.Spec.Parameters.Raw) > 0 { err = yaml.Unmarshal([]byte(in.Spec.Parameters.Raw), &parameters) if err != nil { glog.Errorf("Failed to unmarshal Instance parameters\n%s\n %v", in.Spec.Parameters, err) return err } } createReq := &brokerapi.CreateServiceInstanceRequest{ ServiceID: in.Spec.OSBServiceID, PlanID: in.Spec.OSBPlanID, Parameters: parameters, } _, err = client.CreateServiceInstance(in.Spec.OSBGUID, createReq) return err } // GetAuthCredentialsFromBroker returns the auth credentials, if any, // contained in the secret referenced in the Broker's AuthSecret field, or // returns an error. If the AuthSecret field is nil, empty values are // returned. func GetAuthCredentialsFromBroker(client kubernetes.Interface, broker *servicecatalog.Broker) (username, password string, err error) { if broker.Spec.AuthSecret == nil { return "", "", nil } authSecret, err := client.Core().Secrets(broker.Spec.AuthSecret.Namespace).Get(broker.Spec.AuthSecret.Name) if err != nil { return "", "", err } usernameBytes, ok := authSecret.Data["username"] if !ok { return "", "", fmt.Errorf("auth secret didn't contain username") } passwordBytes, ok := authSecret.Data["password"] if !ok { return "", "", fmt.Errorf("auth secret didn't contain password") } return string(usernameBytes), string(passwordBytes), nil } /////////////////////////////////////////////////////////////////////////////// // All the methods implementing the Handler interface go here for clarity sake. /////////////////////////////////////////////////////////////////////////////// func (h *handler) CreateServiceInstance(in *servicecatalog.Instance) (*servicecatalog.Instance, error) { serviceID, planID, planName, err := util.GetServicePlanInfo( h.apiClient.ServiceClasses(), in.Spec.ServiceClassName, in.Spec.PlanName, ) if err != nil { glog.Errorf("Error fetching service ID: %v", err) return nil, err } in.Spec.OSBServiceID = serviceID in.Spec.OSBPlanID = planID in.Spec.PlanName = planName if in.Spec.OSBGUID == "" { in.Spec.OSBGUID = uuid.NewV4().String() } glog.Infof("Instantiating service %s using service/plan %s : %s", in.Name, serviceID, planID) err = h.createServiceInstance(in) in.Status = servicecatalog.InstanceStatus{} if err != nil { in.Status.Conditions = []servicecatalog.InstanceCondition{ { Type: servicecatalog.InstanceConditionProvisionFailed, Status: servicecatalog.ConditionTrue, Reason: err.Error(), }, } glog.Errorf("Failed to create service instance: %v", err) } else { in.Status.Conditions = []servicecatalog.InstanceCondition{ { Type: servicecatalog.InstanceConditionReady, Status: servicecatalog.ConditionTrue, }, } } glog.Infof("Updating Service %s with State\n%v", in.Name, in.Status.Conditions[0].Type) return h.apiClient.Instances(in.ObjectMeta.Namespace).Update(in) } // DeleteServiceBinding executes all the actions needed before a binding resource can be // safely deleted. These actions include but are not necessarily limited to deleting the // kubernetes resources associated with the binding and calling the unbind REST operation // on the backing OSB API func (h *handler) DeleteServiceBinding(sb *servicecatalog.Binding) error { // this logic to set and update the timestamp is TPR specific. to be moved to the API server dts := metav1.Now() sb.DeletionTimestamp = &dts if _, err := h.apiClient.Bindings(sb.Namespace).Update(sb); err != nil { return err } // uninject if err := h.injector.Uninject(sb); err != nil { // if 0 conditions, uninject and drop condition for uninject // TODO: add failure condition (https://github.com/kubernetes-incubator/service-catalog/issues/305) return err } // TODO: add success condition (https://github.com/kubernetes-incubator/service-catalog/issues/305) if _, err := h.apiClient.Bindings(sb.Namespace).Update(sb); err != nil { return err } // TODO: unbind && add conditions (https://github.com/kubernetes-incubator/service-catalog/issues/305) if err := h.unbind(sb); err != nil { // TODO: add failure condition (https://github.com/kubernetes-incubator/service-catalog/issues/305) return err } // TODO: add success condition (https://github.com/kubernetes-incubator/service-catalog/issues/305) if _, err := h.apiClient.Bindings(sb.Namespace).Update(sb); err != nil { return err } // This is where the binding is _actually_ deleted after all necessary actions have been taken if err := h.apiClient.Bindings(sb.Namespace).Delete(sb.Name); err != nil { // TODO: add deletion error condition (https://github.com/kubernetes-incubator/service-catalog/issues/305) return err } return nil } func (h *handler) CreateServiceBinding(in *servicecatalog.Binding) (*servicecatalog.Binding, error)
client := h.newClientFunc(broker.Name, broker.Spec.URL, authUsername, authPassword) // Assign UUID to binding. if in.Spec.OSBGUID == "" { in.Spec.OSBGUID = uuid.NewV4().String() } parameters := make(map[string]interface{}) if in.Spec.Parameters != nil && len(in.Spec.Parameters.Raw) > 0 { err = yaml.Unmarshal([]byte(in.Spec.Parameters.Raw), &parameters) if err != nil { glog.Errorf("Failed to unmarshal Binding parameters\n%s\n %v", in.Spec.Parameters, err) return nil, err } } createReq := &brokerapi.BindingRequest{ ServiceID: inst.Spec.OSBServiceID, PlanID: inst.Spec.OSBPlanID, Parameters: parameters, } sbr, err := client.CreateServiceBinding(inst.Spec.OSBGUID, in.Spec.OSBGUID, createReq) in.Status = servicecatalog.BindingStatus{} if err != nil { in.Status.Conditions = []servicecatalog.BindingCondition{ { Type: servicecatalog.BindingConditionFailed, Status: servicecatalog.ConditionTrue, Reason: err.Error(), }, } glog.Errorf("Failed to create service instance: %v", err) } else { // Now try injection err := h.injector.Inject(in, &sbr.Credentials) if err != nil { in.Status.Conditions = []servicecatalog.BindingCondition{ { Type: servicecatalog.BindingConditionFailed, Status: servicecatalog
{ glog.Infof("Creating Service Binding: %v", in) inst, err := instanceForBinding(h.apiClient, in) if err != nil { return nil, err } sc, err := serviceClassForInstance(h.apiClient, inst) if err != nil { return nil, err } broker, err := brokerForServiceClass(h.apiClient, sc) if err != nil { return nil, err } authUsername, authPassword, err := GetAuthCredentialsFromBroker(h.k8sClient, broker) if err != nil { return nil, err }
identifier_body
handler.go
existing one. CreateServiceInstance(*servicecatalog.Instance) (*servicecatalog.Instance, error) // CreateServiceBinding takes in a (possibly incomplete) // ServiceBinding and will either create or update an // existing one. CreateServiceBinding(*servicecatalog.Binding) (*servicecatalog.Binding, error) DeleteServiceBinding(*servicecatalog.Binding) error // CreateServiceBroker takes in a (possibly incomplete) // ServiceBroker and will either create or update an // existing one. CreateServiceBroker(*servicecatalog.Broker) (*servicecatalog.Broker, error) } type handler struct { k8sClient kubernetes.Interface apiClient apiclient.APIClient injector injector.BindingInjector newClientFunc func(name, url, username, password string) brokerapi.BrokerClient } func createHandler(k8sClient kubernetes.Interface, client apiclient.APIClient, injector injector.BindingInjector, newClientFn brokerapi.CreateFunc) *handler { return &handler{ k8sClient: k8sClient, apiClient: client, injector: injector, newClientFunc: newClientFn, } } func (h *handler) updateServiceInstance(in *servicecatalog.Instance) error { // Currently there's no difference between create / update, // but for prepping for future, split these into two different // methods for now. return h.createServiceInstance(in) } func (h *handler) createServiceInstance(in *servicecatalog.Instance) error { broker, err := util.GetBrokerByServiceClassName(h.apiClient.Brokers(), h.apiClient.ServiceClasses(), in.Spec.ServiceClassName) if err != nil { return err } authUsername, authPassword, err := GetAuthCredentialsFromBroker(h.k8sClient, broker) if err != nil { return err } client := h.newClientFunc(broker.Name, broker.Spec.URL, authUsername, authPassword) parameters := make(map[string]interface{}) if in.Spec.Parameters != nil && len(in.Spec.Parameters.Raw) > 0 { err = yaml.Unmarshal([]byte(in.Spec.Parameters.Raw), &parameters) if err != nil { glog.Errorf("Failed to unmarshal Instance parameters\n%s\n %v", in.Spec.Parameters, err) return err } } createReq := &brokerapi.CreateServiceInstanceRequest{ ServiceID: in.Spec.OSBServiceID, PlanID: in.Spec.OSBPlanID, Parameters: parameters, } _, err = client.CreateServiceInstance(in.Spec.OSBGUID, createReq) return err } // GetAuthCredentialsFromBroker returns the auth credentials, if any, // contained in the secret referenced in the Broker's AuthSecret field, or // returns an error. If the AuthSecret field is nil, empty values are // returned. func GetAuthCredentialsFromBroker(client kubernetes.Interface, broker *servicecatalog.Broker) (username, password string, err error) { if broker.Spec.AuthSecret == nil { return "", "", nil } authSecret, err := client.Core().Secrets(broker.Spec.AuthSecret.Namespace).Get(broker.Spec.AuthSecret.Name) if err != nil { return "", "", err } usernameBytes, ok := authSecret.Data["username"] if !ok { return "", "", fmt.Errorf("auth secret didn't contain username") } passwordBytes, ok := authSecret.Data["password"] if !ok { return "", "", fmt.Errorf("auth secret didn't contain password") } return string(usernameBytes), string(passwordBytes), nil } /////////////////////////////////////////////////////////////////////////////// // All the methods implementing the Handler interface go here for clarity sake. /////////////////////////////////////////////////////////////////////////////// func (h *handler) CreateServiceInstance(in *servicecatalog.Instance) (*servicecatalog.Instance, error) { serviceID, planID, planName, err := util.GetServicePlanInfo( h.apiClient.ServiceClasses(), in.Spec.ServiceClassName, in.Spec.PlanName, ) if err != nil { glog.Errorf("Error fetching service ID: %v", err) return nil, err } in.Spec.OSBServiceID = serviceID in.Spec.OSBPlanID = planID in.Spec.PlanName = planName if in.Spec.OSBGUID == "" { in.Spec.OSBGUID = uuid.NewV4().String() } glog.Infof("Instantiating service %s using service/plan %s : %s", in.Name, serviceID, planID) err = h.createServiceInstance(in) in.Status = servicecatalog.InstanceStatus{} if err != nil { in.Status.Conditions = []servicecatalog.InstanceCondition{ { Type: servicecatalog.InstanceConditionProvisionFailed, Status: servicecatalog.ConditionTrue, Reason: err.Error(), }, } glog.Errorf("Failed to create service instance: %v", err) } else { in.Status.Conditions = []servicecatalog.InstanceCondition{ { Type: servicecatalog.InstanceConditionReady, Status: servicecatalog.ConditionTrue, }, } } glog.Infof("Updating Service %s with State\n%v", in.Name, in.Status.Conditions[0].Type) return h.apiClient.Instances(in.ObjectMeta.Namespace).Update(in) } // DeleteServiceBinding executes all the actions needed before a binding resource can be // safely deleted. These actions include but are not necessarily limited to deleting the // kubernetes resources associated with the binding and calling the unbind REST operation // on the backing OSB API func (h *handler) DeleteServiceBinding(sb *servicecatalog.Binding) error { // this logic to set and update the timestamp is TPR specific. to be moved to the API server dts := metav1.Now() sb.DeletionTimestamp = &dts if _, err := h.apiClient.Bindings(sb.Namespace).Update(sb); err != nil { return err } // uninject if err := h.injector.Uninject(sb); err != nil { // if 0 conditions, uninject and drop condition for uninject // TODO: add failure condition (https://github.com/kubernetes-incubator/service-catalog/issues/305) return err } // TODO: add success condition (https://github.com/kubernetes-incubator/service-catalog/issues/305) if _, err := h.apiClient.Bindings(sb.Namespace).Update(sb); err != nil { return err } // TODO: unbind && add conditions (https://github.com/kubernetes-incubator/service-catalog/issues/305) if err := h.unbind(sb); err != nil { // TODO: add failure condition (https://github.com/kubernetes-incubator/service-catalog/issues/305) return err } // TODO: add success condition (https://github.com/kubernetes-incubator/service-catalog/issues/305) if _, err := h.apiClient.Bindings(sb.Namespace).Update(sb); err != nil { return err } // This is where the binding is _actually_ deleted after all necessary actions have been taken if err := h.apiClient.Bindings(sb.Namespace).Delete(sb.Name); err != nil { // TODO: add deletion error condition (https://github.com/kubernetes-incubator/service-catalog/issues/305) return err } return nil } func (h *handler) CreateServiceBinding(in *servicecatalog.Binding) (*servicecatalog.Binding, error) { glog.Infof("Creating Service Binding: %v", in) inst, err := instanceForBinding(h.apiClient, in) if err != nil { return nil, err } sc, err := serviceClassForInstance(h.apiClient, inst) if err != nil { return nil, err } broker, err := brokerForServiceClass(h.apiClient, sc) if err != nil { return nil, err } authUsername, authPassword, err := GetAuthCredentialsFromBroker(h.k8sClient, broker) if err != nil { return nil, err } client := h.newClientFunc(broker.Name, broker.Spec.URL, authUsername, authPassword) // Assign UUID to binding. if in.Spec.OSBGUID == ""
parameters := make(map[string]interface{}) if in.Spec.Parameters != nil && len(in.Spec.Parameters.Raw) > 0 { err = yaml.Unmarshal([]byte(in.Spec.Parameters.Raw), &parameters) if err != nil { glog.Errorf("Failed to unmarshal Binding parameters\n%s\n %v", in.Spec.Parameters, err) return nil, err } } createReq := &brokerapi.BindingRequest{ ServiceID: inst.Spec.OSBServiceID, PlanID: inst.Spec.OSBPlanID, Parameters: parameters, } sbr, err := client.CreateServiceBinding(inst.Spec.OSBGUID, in.Spec.OSBGUID, createReq) in.Status = servicecatalog.BindingStatus{} if err != nil { in.Status.Conditions = []servicecatalog.BindingCondition{ { Type: servicecatalog.BindingConditionFailed, Status: servicecatalog.ConditionTrue, Reason: err.Error(), }, } glog.Errorf("Failed to create service instance: %v", err) } else { // Now try injection err := h.injector.Inject(in, &sbr.Credentials) if err != nil { in.Status.Conditions = []servicecatalog.BindingCondition{ { Type: servicecatalog.BindingConditionFailed, Status: servicecatalog
{ in.Spec.OSBGUID = uuid.NewV4().String() }
conditional_block
handler.go
catalogURLFormatString = "%s/v2/catalog" serviceInstanceFormatString = "%s/v2/service_instances/%s" bindingFormatString = "%s/v2/service_instances/%s/service_bindings/%s" defaultNamespace = "default" ) // Handler defines an interface used as a facade for data access operations. // The controller uses the functions of this interface as callbacks for various // events. type Handler interface { // CreateServiceInstance takes in a (possibly incomplete) // ServiceInstance and will either create or update an // existing one. CreateServiceInstance(*servicecatalog.Instance) (*servicecatalog.Instance, error) // CreateServiceBinding takes in a (possibly incomplete) // ServiceBinding and will either create or update an // existing one. CreateServiceBinding(*servicecatalog.Binding) (*servicecatalog.Binding, error) DeleteServiceBinding(*servicecatalog.Binding) error // CreateServiceBroker takes in a (possibly incomplete) // ServiceBroker and will either create or update an // existing one. CreateServiceBroker(*servicecatalog.Broker) (*servicecatalog.Broker, error) } type handler struct { k8sClient kubernetes.Interface apiClient apiclient.APIClient injector injector.BindingInjector newClientFunc func(name, url, username, password string) brokerapi.BrokerClient } func createHandler(k8sClient kubernetes.Interface, client apiclient.APIClient, injector injector.BindingInjector, newClientFn brokerapi.CreateFunc) *handler { return &handler{ k8sClient: k8sClient, apiClient: client, injector: injector, newClientFunc: newClientFn, } } func (h *handler) updateServiceInstance(in *servicecatalog.Instance) error { // Currently there's no difference between create / update, // but for prepping for future, split these into two different // methods for now. return h.createServiceInstance(in) } func (h *handler) createServiceInstance(in *servicecatalog.Instance) error { broker, err := util.GetBrokerByServiceClassName(h.apiClient.Brokers(), h.apiClient.ServiceClasses(), in.Spec.ServiceClassName) if err != nil { return err } authUsername, authPassword, err := GetAuthCredentialsFromBroker(h.k8sClient, broker) if err != nil { return err } client := h.newClientFunc(broker.Name, broker.Spec.URL, authUsername, authPassword) parameters := make(map[string]interface{}) if in.Spec.Parameters != nil && len(in.Spec.Parameters.Raw) > 0 { err = yaml.Unmarshal([]byte(in.Spec.Parameters.Raw), &parameters) if err != nil { glog.Errorf("Failed to unmarshal Instance parameters\n%s\n %v", in.Spec.Parameters, err) return err } } createReq := &brokerapi.CreateServiceInstanceRequest{ ServiceID: in.Spec.OSBServiceID, PlanID: in.Spec.OSBPlanID, Parameters: parameters, } _, err = client.CreateServiceInstance(in.Spec.OSBGUID, createReq) return err } // GetAuthCredentialsFromBroker returns the auth credentials, if any, // contained in the secret referenced in the Broker's AuthSecret field, or // returns an error. If the AuthSecret field is nil, empty values are // returned. func GetAuthCredentialsFromBroker(client kubernetes.Interface, broker *servicecatalog.Broker) (username, password string, err error) { if broker.Spec.AuthSecret == nil { return "", "", nil } authSecret, err := client.Core().Secrets(broker.Spec.AuthSecret.Namespace).Get(broker.Spec.AuthSecret.Name) if err != nil { return "", "", err } usernameBytes, ok := authSecret.Data["username"] if !ok { return "", "", fmt.Errorf("auth secret didn't contain username") } passwordBytes, ok := authSecret.Data["password"] if !ok { return "", "", fmt.Errorf("auth secret didn't contain password") } return string(usernameBytes), string(passwordBytes), nil } /////////////////////////////////////////////////////////////////////////////// // All the methods implementing the Handler interface go here for clarity sake. /////////////////////////////////////////////////////////////////////////////// func (h *handler) CreateServiceInstance(in *servicecatalog.Instance) (*servicecatalog.Instance, error) { serviceID, planID, planName, err := util.GetServicePlanInfo( h.apiClient.ServiceClasses(), in.Spec.ServiceClassName, in.Spec.PlanName, ) if err != nil { glog.Errorf("Error fetching service ID: %v", err) return nil, err } in.Spec.OSBServiceID = serviceID in.Spec.OSBPlanID = planID in.Spec.PlanName = planName if in.Spec.OSBGUID == "" { in.Spec.OSBGUID = uuid.NewV4().String() } glog.Infof("Instantiating service %s using service/plan %s : %s", in.Name, serviceID, planID) err = h.createServiceInstance(in) in.Status = servicecatalog.InstanceStatus{} if err != nil { in.Status.Conditions = []servicecatalog.InstanceCondition{ { Type: servicecatalog.InstanceConditionProvisionFailed, Status: servicecatalog.ConditionTrue, Reason: err.Error(), }, } glog.Errorf("Failed to create service instance: %v", err) } else { in.Status.Conditions = []servicecatalog.InstanceCondition{ { Type: servicecatalog.InstanceConditionReady, Status: servicecatalog.ConditionTrue, }, } } glog.Infof("Updating Service %s with State\n%v", in.Name, in.Status.Conditions[0].Type) return h.apiClient.Instances(in.ObjectMeta.Namespace).Update(in) } // DeleteServiceBinding executes all the actions needed before a binding resource can be // safely deleted. These actions include but are not necessarily limited to deleting the // kubernetes resources associated with the binding and calling the unbind REST operation // on the backing OSB API func (h *handler) DeleteServiceBinding(sb *servicecatalog.Binding) error { // this logic to set and update the timestamp is TPR specific. to be moved to the API server dts := metav1.Now() sb.DeletionTimestamp = &dts if _, err := h.apiClient.Bindings(sb.Namespace).Update(sb); err != nil { return err } // uninject if err := h.injector.Uninject(sb); err != nil { // if 0 conditions, uninject and drop condition for uninject // TODO: add failure condition (https://github.com/kubernetes-incubator/service-catalog/issues/305) return err } // TODO: add success condition (https://github.com/kubernetes-incubator/service-catalog/issues/305) if _, err := h.apiClient.Bindings(sb.Namespace).Update(sb); err != nil { return err } // TODO: unbind && add conditions (https://github.com/kubernetes-incubator/service-catalog/issues/305) if err := h.unbind(sb); err != nil { // TODO: add failure condition (https://github.com/kubernetes-incubator/service-catalog/issues/305) return err } // TODO: add success condition (https://github.com/kubernetes-incubator/service-catalog/issues/305) if _, err := h.apiClient.Bindings(sb.Namespace).Update(sb); err != nil { return err } // This is where the binding is _actually_ deleted after all necessary actions have been taken if err := h.apiClient.Bindings(sb.Namespace).Delete(sb.Name); err != nil { // TODO: add deletion error condition (https://github.com/kubernetes-incubator/service-catalog/issues/305) return err } return nil } func (h *handler) CreateServiceBinding(in *servicecatalog.Binding) (*servicecatalog.Binding, error) { glog.Infof("Creating Service Binding: %v", in) inst, err := instanceForBinding(h.apiClient, in) if err != nil { return nil, err } sc, err := serviceClassForInstance(h.apiClient, inst) if err != nil { return nil, err } broker, err := brokerForServiceClass(h.apiClient, sc) if err != nil { return nil, err } authUsername, authPassword, err := GetAuthCredentialsFromBroker(h.k8sClient, broker) if err != nil { return nil, err } client := h.newClientFunc(broker.Name, broker.Spec.URL, authUsername, authPassword) // Assign UUID to binding. if in.Spec.OSBGUID == "" { in.Spec.OSBGUID = uuid.NewV4().String() } parameters := make(map[string]interface{}) if in.Spec.Parameters != nil && len(in.Spec.Parameters.Raw) > 0 { err = yaml.Unmarshal([]byte(in.Spec.Parameters.Raw), &parameters) if err != nil { glog.Errorf("Failed to unmarshal Binding parameters\n%s\n %v", in.Spec.Parameters, err) return nil, err } } createReq := &brokerapi.BindingRequest{ ServiceID: inst.Spec.OSBServiceID, PlanID: inst.Spec.OSBPlanID, Parameters: parameters, } sbr, err := client.CreateServiceBinding(inst.Spec.OSBGUID, in.Spec.OSBGUID, createReq) in.Status = servicecatalog.BindingStatus{} if err != nil { in.Status.Conditions = []servicecatalog.Binding
random_line_split
handler.go
kubernetes.Interface apiClient apiclient.APIClient injector injector.BindingInjector newClientFunc func(name, url, username, password string) brokerapi.BrokerClient } func createHandler(k8sClient kubernetes.Interface, client apiclient.APIClient, injector injector.BindingInjector, newClientFn brokerapi.CreateFunc) *handler { return &handler{ k8sClient: k8sClient, apiClient: client, injector: injector, newClientFunc: newClientFn, } } func (h *handler) updateServiceInstance(in *servicecatalog.Instance) error { // Currently there's no difference between create / update, // but for prepping for future, split these into two different // methods for now. return h.createServiceInstance(in) } func (h *handler) createServiceInstance(in *servicecatalog.Instance) error { broker, err := util.GetBrokerByServiceClassName(h.apiClient.Brokers(), h.apiClient.ServiceClasses(), in.Spec.ServiceClassName) if err != nil { return err } authUsername, authPassword, err := GetAuthCredentialsFromBroker(h.k8sClient, broker) if err != nil { return err } client := h.newClientFunc(broker.Name, broker.Spec.URL, authUsername, authPassword) parameters := make(map[string]interface{}) if in.Spec.Parameters != nil && len(in.Spec.Parameters.Raw) > 0 { err = yaml.Unmarshal([]byte(in.Spec.Parameters.Raw), &parameters) if err != nil { glog.Errorf("Failed to unmarshal Instance parameters\n%s\n %v", in.Spec.Parameters, err) return err } } createReq := &brokerapi.CreateServiceInstanceRequest{ ServiceID: in.Spec.OSBServiceID, PlanID: in.Spec.OSBPlanID, Parameters: parameters, } _, err = client.CreateServiceInstance(in.Spec.OSBGUID, createReq) return err } // GetAuthCredentialsFromBroker returns the auth credentials, if any, // contained in the secret referenced in the Broker's AuthSecret field, or // returns an error. If the AuthSecret field is nil, empty values are // returned. func GetAuthCredentialsFromBroker(client kubernetes.Interface, broker *servicecatalog.Broker) (username, password string, err error) { if broker.Spec.AuthSecret == nil { return "", "", nil } authSecret, err := client.Core().Secrets(broker.Spec.AuthSecret.Namespace).Get(broker.Spec.AuthSecret.Name) if err != nil { return "", "", err } usernameBytes, ok := authSecret.Data["username"] if !ok { return "", "", fmt.Errorf("auth secret didn't contain username") } passwordBytes, ok := authSecret.Data["password"] if !ok { return "", "", fmt.Errorf("auth secret didn't contain password") } return string(usernameBytes), string(passwordBytes), nil } /////////////////////////////////////////////////////////////////////////////// // All the methods implementing the Handler interface go here for clarity sake. /////////////////////////////////////////////////////////////////////////////// func (h *handler) CreateServiceInstance(in *servicecatalog.Instance) (*servicecatalog.Instance, error) { serviceID, planID, planName, err := util.GetServicePlanInfo( h.apiClient.ServiceClasses(), in.Spec.ServiceClassName, in.Spec.PlanName, ) if err != nil { glog.Errorf("Error fetching service ID: %v", err) return nil, err } in.Spec.OSBServiceID = serviceID in.Spec.OSBPlanID = planID in.Spec.PlanName = planName if in.Spec.OSBGUID == "" { in.Spec.OSBGUID = uuid.NewV4().String() } glog.Infof("Instantiating service %s using service/plan %s : %s", in.Name, serviceID, planID) err = h.createServiceInstance(in) in.Status = servicecatalog.InstanceStatus{} if err != nil { in.Status.Conditions = []servicecatalog.InstanceCondition{ { Type: servicecatalog.InstanceConditionProvisionFailed, Status: servicecatalog.ConditionTrue, Reason: err.Error(), }, } glog.Errorf("Failed to create service instance: %v", err) } else { in.Status.Conditions = []servicecatalog.InstanceCondition{ { Type: servicecatalog.InstanceConditionReady, Status: servicecatalog.ConditionTrue, }, } } glog.Infof("Updating Service %s with State\n%v", in.Name, in.Status.Conditions[0].Type) return h.apiClient.Instances(in.ObjectMeta.Namespace).Update(in) } // DeleteServiceBinding executes all the actions needed before a binding resource can be // safely deleted. These actions include but are not necessarily limited to deleting the // kubernetes resources associated with the binding and calling the unbind REST operation // on the backing OSB API func (h *handler) DeleteServiceBinding(sb *servicecatalog.Binding) error { // this logic to set and update the timestamp is TPR specific. to be moved to the API server dts := metav1.Now() sb.DeletionTimestamp = &dts if _, err := h.apiClient.Bindings(sb.Namespace).Update(sb); err != nil { return err } // uninject if err := h.injector.Uninject(sb); err != nil { // if 0 conditions, uninject and drop condition for uninject // TODO: add failure condition (https://github.com/kubernetes-incubator/service-catalog/issues/305) return err } // TODO: add success condition (https://github.com/kubernetes-incubator/service-catalog/issues/305) if _, err := h.apiClient.Bindings(sb.Namespace).Update(sb); err != nil { return err } // TODO: unbind && add conditions (https://github.com/kubernetes-incubator/service-catalog/issues/305) if err := h.unbind(sb); err != nil { // TODO: add failure condition (https://github.com/kubernetes-incubator/service-catalog/issues/305) return err } // TODO: add success condition (https://github.com/kubernetes-incubator/service-catalog/issues/305) if _, err := h.apiClient.Bindings(sb.Namespace).Update(sb); err != nil { return err } // This is where the binding is _actually_ deleted after all necessary actions have been taken if err := h.apiClient.Bindings(sb.Namespace).Delete(sb.Name); err != nil { // TODO: add deletion error condition (https://github.com/kubernetes-incubator/service-catalog/issues/305) return err } return nil } func (h *handler) CreateServiceBinding(in *servicecatalog.Binding) (*servicecatalog.Binding, error) { glog.Infof("Creating Service Binding: %v", in) inst, err := instanceForBinding(h.apiClient, in) if err != nil { return nil, err } sc, err := serviceClassForInstance(h.apiClient, inst) if err != nil { return nil, err } broker, err := brokerForServiceClass(h.apiClient, sc) if err != nil { return nil, err } authUsername, authPassword, err := GetAuthCredentialsFromBroker(h.k8sClient, broker) if err != nil { return nil, err } client := h.newClientFunc(broker.Name, broker.Spec.URL, authUsername, authPassword) // Assign UUID to binding. if in.Spec.OSBGUID == "" { in.Spec.OSBGUID = uuid.NewV4().String() } parameters := make(map[string]interface{}) if in.Spec.Parameters != nil && len(in.Spec.Parameters.Raw) > 0 { err = yaml.Unmarshal([]byte(in.Spec.Parameters.Raw), &parameters) if err != nil { glog.Errorf("Failed to unmarshal Binding parameters\n%s\n %v", in.Spec.Parameters, err) return nil, err } } createReq := &brokerapi.BindingRequest{ ServiceID: inst.Spec.OSBServiceID, PlanID: inst.Spec.OSBPlanID, Parameters: parameters, } sbr, err := client.CreateServiceBinding(inst.Spec.OSBGUID, in.Spec.OSBGUID, createReq) in.Status = servicecatalog.BindingStatus{} if err != nil { in.Status.Conditions = []servicecatalog.BindingCondition{ { Type: servicecatalog.BindingConditionFailed, Status: servicecatalog.ConditionTrue, Reason: err.Error(), }, } glog.Errorf("Failed to create service instance: %v", err) } else { // Now try injection err := h.injector.Inject(in, &sbr.Credentials) if err != nil { in.Status.Conditions = []servicecatalog.BindingCondition{ { Type: servicecatalog.BindingConditionFailed, Status: servicecatalog.ConditionTrue, Reason: err.Error(), }, } glog.Errorf("Failed to create service instance: %v", err) } else { in.Status.Conditions = []servicecatalog.BindingCondition{ { Type: servicecatalog.BindingConditionReady, Status: servicecatalog.ConditionTrue, }, } } } glog.Infof("Updating Service Binding %s with State\n%v", in.Name, in.Status.Conditions[0].Type) return h.apiClient.Bindings(in.ObjectMeta.Namespace).Update(in) } func (h *handler)
CreateServiceBroker
identifier_name
Quirk.ts
} from "../Copy2Clipboard"; export abstract class Quirk { static inputField: HTMLTextAreaElement; static textFields: HTMLFieldSetElement; private readonly name: string; private shortName: string; private id: string; private readonly colorClass: string; input: string; private row: HTMLDivElement; private textArea: HTMLTextAreaElement; public activeCheckbox: HTMLInputElement; optionalCheckboxes: Array<OptionalCheckbox>; protected constructor(name: string, colorClass: string = "") { this.name = name; let spaceIndex = this.name.indexOf(" "); this.shortName = spaceIndex > 0 ? this.name.substr(0, spaceIndex) : name; this.id = this.shortName.toLocaleLowerCase(); this.optionalCheckboxes = new Array<OptionalCheckbox>(); this.colorClass = colorClass.length < 1 ? this.id : colorClass; } public render(category: Category): void { Quirk.textFields.insertAdjacentHTML('beforeend', renderHTML(this.name, this.id, this.colorClass)); this.row = <HTMLTableRowElement>document.getElementById(this.id + "-row") this.textArea = this.row.getElementsByTagName("textarea")[0]; this.textArea.onclick = selectAllAndCopy; // Create toggle checkbox. this.activeCheckbox = document.createElement("input"); this.activeCheckbox.classList.add("filled-in"); this.activeCheckbox.classList.add("checkbox-" + this.getColorClass()); this.activeCheckbox.type = "checkbox"; this.activeCheckbox.checked = true; this.activeCheckbox.onchange = () => this.updateVisibility(category); let td: HTMLTableCellElement = document.createElement("td"); td.insertAdjacentElement('beforeend', this.activeCheckbox); // Checkbox requires a span element adjacent to it for Materialize's theme to work. let span = document.createElement("span"); span.insertAdjacentText('beforeend', this.name); td.insertAdjacentElement('beforeend', span); let tr: HTMLTableRowElement = document.createElement("tr"); tr.classList.add("waves-effect"); tr.classList.add("waves-" + this.getColorClass()); tr.onclick = () => this.activeCheckbox.click(); tr.insertAdjacentElement('beforeend', td); let toggleCheckboxSet = category.getMainCheckboxSetElement(); toggleCheckboxSet.insertAdjacentElement('beforeend', tr); for (let i = 0; i < this.optionalCheckboxes.length; i++) { this.optionalCheckboxes[i].render(category, this.getID(), this); } } public getID(): string { return this.id; } public
(bruh: string): void { this.shortName = bruh; this.id = bruh.toLocaleLowerCase(); } public getShortName(): string { return this.shortName; } public getColorClass(): string { return this.colorClass; } public getTextAreaElement(): HTMLTextAreaElement { return this.textArea; } updateVisibility(category: Category): void { this.row.hidden = !this.activeCheckbox.checked; let optionals = <HTMLCollectionOf<HTMLElement>>document.getElementsByClassName(this.id + "-optional"); for (let i = 0; i < optionals.length; i++) { optionals[i].hidden = !this.activeCheckbox.checked; } let visible = !this.row.hidden // Save setting to cookies. setCookieBool(this.id, visible, 31); let optionalCheckboxSet: HTMLDivElement = <HTMLDivElement>document.getElementById(category.tabName.toLocaleLowerCase() + "-optional-checkboxes"); if (visible) { this.update(Quirk.inputField.value); if (optionalCheckboxSet.hidden && optionals.length > 0) { optionalCheckboxSet.hidden = false; } } else { // Check if any other optional checkboxes are visible. for (let i = 0; i < category.optionalCheckboxes.length; i++) { if (!category.optionalCheckboxes[i].hidden) { return; } } // Hide the table. optionalCheckboxSet.hidden = true; } } update(str: string): void { if (!this.activeCheckbox.checked || str.length < 1) { return; } this.input = str; this.quirkify(); this.updateTextField(); } updateTextField(): void { this.textArea.value = this.input; // Auto resize. Quirk.autoSize(this.textArea); } // Dynamically increase the height of a textarea. static autoSize(element: HTMLTextAreaElement): void { let minHeight: number = parseInt(window.getComputedStyle(element).getPropertyValue("min-height")); element.style.height = "auto"; // Lets the element shrink size. element.style.height = `${Math.max(minHeight, element.scrollHeight)}px`; } addCheckbox(label: string, title: string, defaultValue: boolean = false): OptionalCheckbox { let checkbox: OptionalCheckbox = new OptionalCheckbox(label, title, defaultValue) this.optionalCheckboxes.push(checkbox); return checkbox; } abstract quirkify(): void; protected lowerCase(pattern: string = ""): void { if (pattern.length < 1) { this.input = this.input.toLocaleLowerCase(); } else { let reg: RegExp = new RegExp(pattern, "gi"); this.input = this.input.replace(reg, function(match) { return match.toLocaleLowerCase(); }); } } protected upperCase(pattern: string = ""): void { if (pattern.length < 1) { this.input = this.input.toLocaleUpperCase(); } else { let reg: RegExp = new RegExp(pattern, "gi"); this.input = this.input.replace(reg, function(match) { return match.toLocaleUpperCase(); }); } } protected prefix(str: string): void { this.input = str + this.input; } protected suffix(str: string): void { this.input += str; } protected replaceString(pattern: string, replace: string): void { let reg: RegExp = new RegExp(pattern, "g"); this.input = this.input.replace(reg, replace); } protected replaceCaseInsensitive(pattern: string, replace: string): void { let reg: RegExp = new RegExp(pattern, "gi"); this.input = this.input.replace(reg, replace); } protected replaceMatchCase(pattern: string, replace: string): void { let reg: RegExp = new RegExp(pattern, "gi"); this.input = this.input.replace(reg, function(match) { return Quirk.matchCase(replace, match); }); } protected replaceWord(pattern: string, replace: string) { this.replaceString("\\b" + pattern + "\\b", replace); } protected replaceWordMatchCase(pattern: string, replace: string) { this.replaceMatchCase("\\b" + pattern + "\\b", replace); } // Function taken from https://stackoverflow.com/a/17265031/6446221. private static matchCase(text: string, pattern: string): string { // If the entire text is uppercase then uppercase the whole pattern regardless of lengths. if (pattern.toUpperCase() === pattern) { return text.toUpperCase(); } let result = ''; for (let i = 0; i < text.length; i++) { let c = text.charAt(i); let p = pattern.charCodeAt(i); if (p >= 65 && p < 65 + 26) { result += c.toUpperCase(); } else { result += c.toLowerCase(); } } return result; } randomReplace(pattern: string, replace: string, prob: number): void { let reg: RegExp = new RegExp(pattern, "g"); this.input = this.input.replace(reg, function(match) { if (Math.random() <= prob) { return replace; } return match; }); } // Troll-specific stuff below. // $1 - capture group for eyes. // $2 - capture group for mouth. replaceEmotes(replace: string): void { let eyes = "[:;]"; let mouth = "[\\)\\(Dd]"; this.upperCase(`(${eyes})(${mouth})`); let reg: RegExp = new RegExp(`(${eyes})(${mouth})`, "gi"); this.input = this.input.replace(reg, replace); } applyCatPuns(): void { this.replaceMatchCase("mother", "meowther"); this.replaceMatchCase("for", "fur"); this.replaceMatchCase("pause", "paws"); this.replaceMatchCase("cause", "claws"); this.replaceMatchCase("now", "meow"); this.replaceMatchCase("(per|pre)", "pur"); } applyFishPuns(): void { this.replaceMatchCase("kill", "krill"); this.replaceMatchCase("well", "whale"); this.replaceMatchCase("fine", "fin"); this.replaceMatchCase("see", "sea"); this.replaceMatchCase("should", "shoald"); this.replaceMatchCase("kid", "squid"); this.replaceMatchCase("sure", "shore"); this.replaceMatchCase("crap", "carp"); this.replaceMatchCase("(what are|what do)", "water"); } applyTiaraEmotes(): void { this.replaceEmotes("38$2"); } censorSwears(extreme: boolean = false): void { this.replaceWordMatchCase("fuck", "f*ck");
setShortName
identifier_name
Quirk.ts
from "../Copy2Clipboard"; export abstract class Quirk { static inputField: HTMLTextAreaElement; static textFields: HTMLFieldSetElement; private readonly name: string; private shortName: string; private id: string; private readonly colorClass: string; input: string; private row: HTMLDivElement; private textArea: HTMLTextAreaElement; public activeCheckbox: HTMLInputElement; optionalCheckboxes: Array<OptionalCheckbox>; protected constructor(name: string, colorClass: string = "") { this.name = name; let spaceIndex = this.name.indexOf(" "); this.shortName = spaceIndex > 0 ? this.name.substr(0, spaceIndex) : name; this.id = this.shortName.toLocaleLowerCase(); this.optionalCheckboxes = new Array<OptionalCheckbox>(); this.colorClass = colorClass.length < 1 ? this.id : colorClass; } public render(category: Category): void { Quirk.textFields.insertAdjacentHTML('beforeend', renderHTML(this.name, this.id, this.colorClass)); this.row = <HTMLTableRowElement>document.getElementById(this.id + "-row") this.textArea = this.row.getElementsByTagName("textarea")[0]; this.textArea.onclick = selectAllAndCopy; // Create toggle checkbox. this.activeCheckbox = document.createElement("input"); this.activeCheckbox.classList.add("filled-in"); this.activeCheckbox.classList.add("checkbox-" + this.getColorClass()); this.activeCheckbox.type = "checkbox"; this.activeCheckbox.checked = true; this.activeCheckbox.onchange = () => this.updateVisibility(category); let td: HTMLTableCellElement = document.createElement("td"); td.insertAdjacentElement('beforeend', this.activeCheckbox); // Checkbox requires a span element adjacent to it for Materialize's theme to work. let span = document.createElement("span"); span.insertAdjacentText('beforeend', this.name); td.insertAdjacentElement('beforeend', span); let tr: HTMLTableRowElement = document.createElement("tr"); tr.classList.add("waves-effect"); tr.classList.add("waves-" + this.getColorClass()); tr.onclick = () => this.activeCheckbox.click(); tr.insertAdjacentElement('beforeend', td); let toggleCheckboxSet = category.getMainCheckboxSetElement(); toggleCheckboxSet.insertAdjacentElement('beforeend', tr); for (let i = 0; i < this.optionalCheckboxes.length; i++) { this.optionalCheckboxes[i].render(category, this.getID(), this); } } public getID(): string { return this.id; } public setShortName(bruh: string): void { this.shortName = bruh; this.id = bruh.toLocaleLowerCase(); } public getShortName(): string { return this.shortName; } public getColorClass(): string { return this.colorClass; } public getTextAreaElement(): HTMLTextAreaElement { return this.textArea; } updateVisibility(category: Category): void { this.row.hidden = !this.activeCheckbox.checked; let optionals = <HTMLCollectionOf<HTMLElement>>document.getElementsByClassName(this.id + "-optional"); for (let i = 0; i < optionals.length; i++) { optionals[i].hidden = !this.activeCheckbox.checked; } let visible = !this.row.hidden // Save setting to cookies. setCookieBool(this.id, visible, 31); let optionalCheckboxSet: HTMLDivElement = <HTMLDivElement>document.getElementById(category.tabName.toLocaleLowerCase() + "-optional-checkboxes"); if (visible) { this.update(Quirk.inputField.value); if (optionalCheckboxSet.hidden && optionals.length > 0) { optionalCheckboxSet.hidden = false; } } else { // Check if any other optional checkboxes are visible. for (let i = 0; i < category.optionalCheckboxes.length; i++) { if (!category.optionalCheckboxes[i].hidden) { return; } } // Hide the table. optionalCheckboxSet.hidden = true; } } update(str: string): void { if (!this.activeCheckbox.checked || str.length < 1) { return; } this.input = str; this.quirkify(); this.updateTextField(); } updateTextField(): void
// Dynamically increase the height of a textarea. static autoSize(element: HTMLTextAreaElement): void { let minHeight: number = parseInt(window.getComputedStyle(element).getPropertyValue("min-height")); element.style.height = "auto"; // Lets the element shrink size. element.style.height = `${Math.max(minHeight, element.scrollHeight)}px`; } addCheckbox(label: string, title: string, defaultValue: boolean = false): OptionalCheckbox { let checkbox: OptionalCheckbox = new OptionalCheckbox(label, title, defaultValue) this.optionalCheckboxes.push(checkbox); return checkbox; } abstract quirkify(): void; protected lowerCase(pattern: string = ""): void { if (pattern.length < 1) { this.input = this.input.toLocaleLowerCase(); } else { let reg: RegExp = new RegExp(pattern, "gi"); this.input = this.input.replace(reg, function(match) { return match.toLocaleLowerCase(); }); } } protected upperCase(pattern: string = ""): void { if (pattern.length < 1) { this.input = this.input.toLocaleUpperCase(); } else { let reg: RegExp = new RegExp(pattern, "gi"); this.input = this.input.replace(reg, function(match) { return match.toLocaleUpperCase(); }); } } protected prefix(str: string): void { this.input = str + this.input; } protected suffix(str: string): void { this.input += str; } protected replaceString(pattern: string, replace: string): void { let reg: RegExp = new RegExp(pattern, "g"); this.input = this.input.replace(reg, replace); } protected replaceCaseInsensitive(pattern: string, replace: string): void { let reg: RegExp = new RegExp(pattern, "gi"); this.input = this.input.replace(reg, replace); } protected replaceMatchCase(pattern: string, replace: string): void { let reg: RegExp = new RegExp(pattern, "gi"); this.input = this.input.replace(reg, function(match) { return Quirk.matchCase(replace, match); }); } protected replaceWord(pattern: string, replace: string) { this.replaceString("\\b" + pattern + "\\b", replace); } protected replaceWordMatchCase(pattern: string, replace: string) { this.replaceMatchCase("\\b" + pattern + "\\b", replace); } // Function taken from https://stackoverflow.com/a/17265031/6446221. private static matchCase(text: string, pattern: string): string { // If the entire text is uppercase then uppercase the whole pattern regardless of lengths. if (pattern.toUpperCase() === pattern) { return text.toUpperCase(); } let result = ''; for (let i = 0; i < text.length; i++) { let c = text.charAt(i); let p = pattern.charCodeAt(i); if (p >= 65 && p < 65 + 26) { result += c.toUpperCase(); } else { result += c.toLowerCase(); } } return result; } randomReplace(pattern: string, replace: string, prob: number): void { let reg: RegExp = new RegExp(pattern, "g"); this.input = this.input.replace(reg, function(match) { if (Math.random() <= prob) { return replace; } return match; }); } // Troll-specific stuff below. // $1 - capture group for eyes. // $2 - capture group for mouth. replaceEmotes(replace: string): void { let eyes = "[:;]"; let mouth = "[\\)\\(Dd]"; this.upperCase(`(${eyes})(${mouth})`); let reg: RegExp = new RegExp(`(${eyes})(${mouth})`, "gi"); this.input = this.input.replace(reg, replace); } applyCatPuns(): void { this.replaceMatchCase("mother", "meowther"); this.replaceMatchCase("for", "fur"); this.replaceMatchCase("pause", "paws"); this.replaceMatchCase("cause", "claws"); this.replaceMatchCase("now", "meow"); this.replaceMatchCase("(per|pre)", "pur"); } applyFishPuns(): void { this.replaceMatchCase("kill", "krill"); this.replaceMatchCase("well", "whale"); this.replaceMatchCase("fine", "fin"); this.replaceMatchCase("see", "sea"); this.replaceMatchCase("should", "shoald"); this.replaceMatchCase("kid", "squid"); this.replaceMatchCase("sure", "shore"); this.replaceMatchCase("crap", "carp"); this.replaceMatchCase("(what are|what do)", "water"); } applyTiaraEmotes(): void { this.replaceEmotes("38$2"); } censorSwears(extreme: boolean = false): void { this.replaceWordMatchCase("fuck", "f*ck");
{ this.textArea.value = this.input; // Auto resize. Quirk.autoSize(this.textArea); }
identifier_body
Quirk.ts
from "../Copy2Clipboard"; export abstract class Quirk { static inputField: HTMLTextAreaElement; static textFields: HTMLFieldSetElement; private readonly name: string; private shortName: string; private id: string; private readonly colorClass: string; input: string; private row: HTMLDivElement; private textArea: HTMLTextAreaElement; public activeCheckbox: HTMLInputElement; optionalCheckboxes: Array<OptionalCheckbox>; protected constructor(name: string, colorClass: string = "") { this.name = name; let spaceIndex = this.name.indexOf(" "); this.shortName = spaceIndex > 0 ? this.name.substr(0, spaceIndex) : name; this.id = this.shortName.toLocaleLowerCase(); this.optionalCheckboxes = new Array<OptionalCheckbox>(); this.colorClass = colorClass.length < 1 ? this.id : colorClass; } public render(category: Category): void { Quirk.textFields.insertAdjacentHTML('beforeend', renderHTML(this.name, this.id, this.colorClass)); this.row = <HTMLTableRowElement>document.getElementById(this.id + "-row") this.textArea = this.row.getElementsByTagName("textarea")[0]; this.textArea.onclick = selectAllAndCopy; // Create toggle checkbox. this.activeCheckbox = document.createElement("input"); this.activeCheckbox.classList.add("filled-in"); this.activeCheckbox.classList.add("checkbox-" + this.getColorClass()); this.activeCheckbox.type = "checkbox"; this.activeCheckbox.checked = true; this.activeCheckbox.onchange = () => this.updateVisibility(category); let td: HTMLTableCellElement = document.createElement("td"); td.insertAdjacentElement('beforeend', this.activeCheckbox); // Checkbox requires a span element adjacent to it for Materialize's theme to work. let span = document.createElement("span"); span.insertAdjacentText('beforeend', this.name); td.insertAdjacentElement('beforeend', span); let tr: HTMLTableRowElement = document.createElement("tr"); tr.classList.add("waves-effect"); tr.classList.add("waves-" + this.getColorClass()); tr.onclick = () => this.activeCheckbox.click(); tr.insertAdjacentElement('beforeend', td); let toggleCheckboxSet = category.getMainCheckboxSetElement(); toggleCheckboxSet.insertAdjacentElement('beforeend', tr); for (let i = 0; i < this.optionalCheckboxes.length; i++) { this.optionalCheckboxes[i].render(category, this.getID(), this); } } public getID(): string { return this.id; } public setShortName(bruh: string): void { this.shortName = bruh; this.id = bruh.toLocaleLowerCase(); } public getShortName(): string { return this.shortName; } public getColorClass(): string { return this.colorClass; } public getTextAreaElement(): HTMLTextAreaElement { return this.textArea; } updateVisibility(category: Category): void { this.row.hidden = !this.activeCheckbox.checked; let optionals = <HTMLCollectionOf<HTMLElement>>document.getElementsByClassName(this.id + "-optional"); for (let i = 0; i < optionals.length; i++) { optionals[i].hidden = !this.activeCheckbox.checked; } let visible = !this.row.hidden // Save setting to cookies. setCookieBool(this.id, visible, 31); let optionalCheckboxSet: HTMLDivElement = <HTMLDivElement>document.getElementById(category.tabName.toLocaleLowerCase() + "-optional-checkboxes"); if (visible) { this.update(Quirk.inputField.value); if (optionalCheckboxSet.hidden && optionals.length > 0) { optionalCheckboxSet.hidden = false; } } else { // Check if any other optional checkboxes are visible. for (let i = 0; i < category.optionalCheckboxes.length; i++) { if (!category.optionalCheckboxes[i].hidden) { return; } } // Hide the table. optionalCheckboxSet.hidden = true; } } update(str: string): void { if (!this.activeCheckbox.checked || str.length < 1) { return; } this.input = str; this.quirkify(); this.updateTextField(); } updateTextField(): void { this.textArea.value = this.input; // Auto resize. Quirk.autoSize(this.textArea); } // Dynamically increase the height of a textarea. static autoSize(element: HTMLTextAreaElement): void { let minHeight: number = parseInt(window.getComputedStyle(element).getPropertyValue("min-height")); element.style.height = "auto"; // Lets the element shrink size. element.style.height = `${Math.max(minHeight, element.scrollHeight)}px`; } addCheckbox(label: string, title: string, defaultValue: boolean = false): OptionalCheckbox { let checkbox: OptionalCheckbox = new OptionalCheckbox(label, title, defaultValue) this.optionalCheckboxes.push(checkbox); return checkbox; } abstract quirkify(): void; protected lowerCase(pattern: string = ""): void { if (pattern.length < 1) { this.input = this.input.toLocaleLowerCase(); } else { let reg: RegExp = new RegExp(pattern, "gi"); this.input = this.input.replace(reg, function(match) { return match.toLocaleLowerCase(); }); } } protected upperCase(pattern: string = ""): void { if (pattern.length < 1) { this.input = this.input.toLocaleUpperCase(); } else { let reg: RegExp = new RegExp(pattern, "gi"); this.input = this.input.replace(reg, function(match) { return match.toLocaleUpperCase(); }); } } protected prefix(str: string): void { this.input = str + this.input; } protected suffix(str: string): void { this.input += str; } protected replaceString(pattern: string, replace: string): void { let reg: RegExp = new RegExp(pattern, "g"); this.input = this.input.replace(reg, replace); } protected replaceCaseInsensitive(pattern: string, replace: string): void { let reg: RegExp = new RegExp(pattern, "gi"); this.input = this.input.replace(reg, replace); } protected replaceMatchCase(pattern: string, replace: string): void { let reg: RegExp = new RegExp(pattern, "gi"); this.input = this.input.replace(reg, function(match) { return Quirk.matchCase(replace, match); }); } protected replaceWord(pattern: string, replace: string) { this.replaceString("\\b" + pattern + "\\b", replace); } protected replaceWordMatchCase(pattern: string, replace: string) { this.replaceMatchCase("\\b" + pattern + "\\b", replace); } // Function taken from https://stackoverflow.com/a/17265031/6446221. private static matchCase(text: string, pattern: string): string { // If the entire text is uppercase then uppercase the whole pattern regardless of lengths. if (pattern.toUpperCase() === pattern) { return text.toUpperCase(); } let result = ''; for (let i = 0; i < text.length; i++) { let c = text.charAt(i); let p = pattern.charCodeAt(i); if (p >= 65 && p < 65 + 26) { result += c.toUpperCase(); } else { result += c.toLowerCase(); } } return result; } randomReplace(pattern: string, replace: string, prob: number): void { let reg: RegExp = new RegExp(pattern, "g"); this.input = this.input.replace(reg, function(match) { if (Math.random() <= prob) { return replace; } return match; }); } // Troll-specific stuff below. // $1 - capture group for eyes. // $2 - capture group for mouth. replaceEmotes(replace: string): void { let eyes = "[:;]"; let mouth = "[\\)\\(Dd]"; this.upperCase(`(${eyes})(${mouth})`); let reg: RegExp = new RegExp(`(${eyes})(${mouth})`, "gi"); this.input = this.input.replace(reg, replace); } applyCatPuns(): void { this.replaceMatchCase("mother", "meowther"); this.replaceMatchCase("for", "fur"); this.replaceMatchCase("pause", "paws");
this.replaceMatchCase("now", "meow"); this.replaceMatchCase("(per|pre)", "pur"); } applyFishPuns(): void { this.replaceMatchCase("kill", "krill"); this.replaceMatchCase("well", "whale"); this.replaceMatchCase("fine", "fin"); this.replaceMatchCase("see", "sea"); this.replaceMatchCase("should", "shoald"); this.replaceMatchCase("kid", "squid"); this.replaceMatchCase("sure", "shore"); this.replaceMatchCase("crap", "carp"); this.replaceMatchCase("(what are|what do)", "water"); } applyTiaraEmotes(): void { this.replaceEmotes("38$2"); } censorSwears(extreme: boolean = false): void { this.replaceWordMatchCase("fuck", "f*ck"); this
this.replaceMatchCase("cause", "claws");
random_line_split
Quirk.ts
from "../Copy2Clipboard"; export abstract class Quirk { static inputField: HTMLTextAreaElement; static textFields: HTMLFieldSetElement; private readonly name: string; private shortName: string; private id: string; private readonly colorClass: string; input: string; private row: HTMLDivElement; private textArea: HTMLTextAreaElement; public activeCheckbox: HTMLInputElement; optionalCheckboxes: Array<OptionalCheckbox>; protected constructor(name: string, colorClass: string = "") { this.name = name; let spaceIndex = this.name.indexOf(" "); this.shortName = spaceIndex > 0 ? this.name.substr(0, spaceIndex) : name; this.id = this.shortName.toLocaleLowerCase(); this.optionalCheckboxes = new Array<OptionalCheckbox>(); this.colorClass = colorClass.length < 1 ? this.id : colorClass; } public render(category: Category): void { Quirk.textFields.insertAdjacentHTML('beforeend', renderHTML(this.name, this.id, this.colorClass)); this.row = <HTMLTableRowElement>document.getElementById(this.id + "-row") this.textArea = this.row.getElementsByTagName("textarea")[0]; this.textArea.onclick = selectAllAndCopy; // Create toggle checkbox. this.activeCheckbox = document.createElement("input"); this.activeCheckbox.classList.add("filled-in"); this.activeCheckbox.classList.add("checkbox-" + this.getColorClass()); this.activeCheckbox.type = "checkbox"; this.activeCheckbox.checked = true; this.activeCheckbox.onchange = () => this.updateVisibility(category); let td: HTMLTableCellElement = document.createElement("td"); td.insertAdjacentElement('beforeend', this.activeCheckbox); // Checkbox requires a span element adjacent to it for Materialize's theme to work. let span = document.createElement("span"); span.insertAdjacentText('beforeend', this.name); td.insertAdjacentElement('beforeend', span); let tr: HTMLTableRowElement = document.createElement("tr"); tr.classList.add("waves-effect"); tr.classList.add("waves-" + this.getColorClass()); tr.onclick = () => this.activeCheckbox.click(); tr.insertAdjacentElement('beforeend', td); let toggleCheckboxSet = category.getMainCheckboxSetElement(); toggleCheckboxSet.insertAdjacentElement('beforeend', tr); for (let i = 0; i < this.optionalCheckboxes.length; i++) { this.optionalCheckboxes[i].render(category, this.getID(), this); } } public getID(): string { return this.id; } public setShortName(bruh: string): void { this.shortName = bruh; this.id = bruh.toLocaleLowerCase(); } public getShortName(): string { return this.shortName; } public getColorClass(): string { return this.colorClass; } public getTextAreaElement(): HTMLTextAreaElement { return this.textArea; } updateVisibility(category: Category): void { this.row.hidden = !this.activeCheckbox.checked; let optionals = <HTMLCollectionOf<HTMLElement>>document.getElementsByClassName(this.id + "-optional"); for (let i = 0; i < optionals.length; i++)
let visible = !this.row.hidden // Save setting to cookies. setCookieBool(this.id, visible, 31); let optionalCheckboxSet: HTMLDivElement = <HTMLDivElement>document.getElementById(category.tabName.toLocaleLowerCase() + "-optional-checkboxes"); if (visible) { this.update(Quirk.inputField.value); if (optionalCheckboxSet.hidden && optionals.length > 0) { optionalCheckboxSet.hidden = false; } } else { // Check if any other optional checkboxes are visible. for (let i = 0; i < category.optionalCheckboxes.length; i++) { if (!category.optionalCheckboxes[i].hidden) { return; } } // Hide the table. optionalCheckboxSet.hidden = true; } } update(str: string): void { if (!this.activeCheckbox.checked || str.length < 1) { return; } this.input = str; this.quirkify(); this.updateTextField(); } updateTextField(): void { this.textArea.value = this.input; // Auto resize. Quirk.autoSize(this.textArea); } // Dynamically increase the height of a textarea. static autoSize(element: HTMLTextAreaElement): void { let minHeight: number = parseInt(window.getComputedStyle(element).getPropertyValue("min-height")); element.style.height = "auto"; // Lets the element shrink size. element.style.height = `${Math.max(minHeight, element.scrollHeight)}px`; } addCheckbox(label: string, title: string, defaultValue: boolean = false): OptionalCheckbox { let checkbox: OptionalCheckbox = new OptionalCheckbox(label, title, defaultValue) this.optionalCheckboxes.push(checkbox); return checkbox; } abstract quirkify(): void; protected lowerCase(pattern: string = ""): void { if (pattern.length < 1) { this.input = this.input.toLocaleLowerCase(); } else { let reg: RegExp = new RegExp(pattern, "gi"); this.input = this.input.replace(reg, function(match) { return match.toLocaleLowerCase(); }); } } protected upperCase(pattern: string = ""): void { if (pattern.length < 1) { this.input = this.input.toLocaleUpperCase(); } else { let reg: RegExp = new RegExp(pattern, "gi"); this.input = this.input.replace(reg, function(match) { return match.toLocaleUpperCase(); }); } } protected prefix(str: string): void { this.input = str + this.input; } protected suffix(str: string): void { this.input += str; } protected replaceString(pattern: string, replace: string): void { let reg: RegExp = new RegExp(pattern, "g"); this.input = this.input.replace(reg, replace); } protected replaceCaseInsensitive(pattern: string, replace: string): void { let reg: RegExp = new RegExp(pattern, "gi"); this.input = this.input.replace(reg, replace); } protected replaceMatchCase(pattern: string, replace: string): void { let reg: RegExp = new RegExp(pattern, "gi"); this.input = this.input.replace(reg, function(match) { return Quirk.matchCase(replace, match); }); } protected replaceWord(pattern: string, replace: string) { this.replaceString("\\b" + pattern + "\\b", replace); } protected replaceWordMatchCase(pattern: string, replace: string) { this.replaceMatchCase("\\b" + pattern + "\\b", replace); } // Function taken from https://stackoverflow.com/a/17265031/6446221. private static matchCase(text: string, pattern: string): string { // If the entire text is uppercase then uppercase the whole pattern regardless of lengths. if (pattern.toUpperCase() === pattern) { return text.toUpperCase(); } let result = ''; for (let i = 0; i < text.length; i++) { let c = text.charAt(i); let p = pattern.charCodeAt(i); if (p >= 65 && p < 65 + 26) { result += c.toUpperCase(); } else { result += c.toLowerCase(); } } return result; } randomReplace(pattern: string, replace: string, prob: number): void { let reg: RegExp = new RegExp(pattern, "g"); this.input = this.input.replace(reg, function(match) { if (Math.random() <= prob) { return replace; } return match; }); } // Troll-specific stuff below. // $1 - capture group for eyes. // $2 - capture group for mouth. replaceEmotes(replace: string): void { let eyes = "[:;]"; let mouth = "[\\)\\(Dd]"; this.upperCase(`(${eyes})(${mouth})`); let reg: RegExp = new RegExp(`(${eyes})(${mouth})`, "gi"); this.input = this.input.replace(reg, replace); } applyCatPuns(): void { this.replaceMatchCase("mother", "meowther"); this.replaceMatchCase("for", "fur"); this.replaceMatchCase("pause", "paws"); this.replaceMatchCase("cause", "claws"); this.replaceMatchCase("now", "meow"); this.replaceMatchCase("(per|pre)", "pur"); } applyFishPuns(): void { this.replaceMatchCase("kill", "krill"); this.replaceMatchCase("well", "whale"); this.replaceMatchCase("fine", "fin"); this.replaceMatchCase("see", "sea"); this.replaceMatchCase("should", "shoald"); this.replaceMatchCase("kid", "squid"); this.replaceMatchCase("sure", "shore"); this.replaceMatchCase("crap", "carp"); this.replaceMatchCase("(what are|what do)", "water"); } applyTiaraEmotes(): void { this.replaceEmotes("38$2"); } censorSwears(extreme: boolean = false): void { this.replaceWordMatchCase("fuck", "f*ck");
{ optionals[i].hidden = !this.activeCheckbox.checked; }
conditional_block
main.py
"cpc": "$CPC$", "pcc": "$PCC$", "drfc": "$D-RFC$", "wpc": "$WPC$" }) df_eval["Critic"] = df_eval["Critic"].replace( to_replace={ "concat": "MLP", "separable": "Separable", "innerprod": "Inner product", "bilinear": "Bilinear" }) return df_eval def apply_default_style(ax): ax.set_xlim([0, 20001]) ax.get_xaxis().set_major_formatter( FuncFormatter(lambda x, p: format(int(x/1000), ','))) ax.set_xlabel("Training steps (in thousands)") plt.tick_params(top=False, right=False, bottom=False, left=False) handles, labels = ax.get_legend_handles_labels() plt.legend(loc="lower right", handles=handles[1:], labels=labels[1:]) def get_testing_loss(x_array, session, loss, data_ph, dims, batch_size=512): total_loss = 0 for i in range(0, x_array.shape[0], batch_size): x_slice = x_array[i:i+batch_size, :dims] total_loss += x_slice.shape[0] * session.run(loss, feed_dict={data_ph: x_slice}) return total_loss / x_array.shape[0] def get_classification_accuracy(session, codes, data_ph, dims): x_train_mapped = map_data(x_train, session, codes, data_ph, dims) x_test_mapped = map_data(x_test, session, codes, data_ph, dims) accuracy = logistic_fit(x_train_mapped, y_train, x_test_mapped, y_test) return accuracy def map_data(x_array, session, codes, data_ph, dims, batch_size=512): x_mapped = [] for i in range(0, x_array.shape[0], batch_size): x_mapped.append( session.run(codes, feed_dict={data_ph: x_array[i:i+batch_size, :dims]})) return np.concatenate(x_mapped, axis=0) def reduce_logmeanexp_nodiag(x, axis=None): batch_size = x.shape[0] logsumexp = tf.reduce_logsumexp(input_tensor=x - tf.linalg.tensor_diag(np.inf * tf.ones(batch_size)), axis=axis) if axis: num_elem = batch_size - 1. else: num_elem = batch_size * (batch_size - 1.) return logsumexp - tf.math.log(num_elem) def tuba_lower_bound(scores, log_baseline=None): if log_baseline is not None: scores -= log_baseline[:, None] batch_size = tf.cast(scores.shape[0], tf.float32) # First term is an expectation over samples from the joint, # which are the diagonal elmements of the scores matrix. joint_term = tf.reduce_mean(input_tensor=tf.linalg.diag_part(scores)) # Second term is an expectation over samples from the marginal, # which are the off-diagonal elements of the scores matrix. marg_term = tf.exp(reduce_logmeanexp_nodiag(scores)) return 1. + joint_term - marg_term def nwj_lower_bound(scores): # equivalent to: tuba_lower_bound(scores, log_baseline=1.) return tuba_lower_bound(scores - 1.) @tf.function def js_fgan_lower_bound(f): """Lower bound on Jensen-Shannon divergence from Nowozin et al. (2016).""" f_diag = tf.linalg.tensor_diag_part(f) first_term = tf.reduce_mean(-tf.nn.softplus(-f_diag)) n = tf.cast(f.shape[0], tf.float32) second_term = (tf.reduce_sum(tf.nn.softplus(f)) - tf.reduce_sum(tf.nn.softplus(f_diag))) / (n * (n - 1.)) return first_term - second_term @tf.function def infonce_lower_bound(scores): """InfoNCE lower bound from van den Oord et al. (2018).""" nll = tf.reduce_mean(input_tensor=tf.linalg.diag_part(scores) - tf.reduce_logsumexp(input_tensor=scores, axis=1)) mi = tf.math.log(tf.cast(scores.shape[0], tf.float32)) + nll return mi @tf.function def our_lower_bound(scores): # scores: 128, 128 """Our lower bound""" batch_size = tf.cast(scores.shape[0], tf.float32) joint_term = tf.reduce_mean(input_tensor=tf.linalg.diag_part(scores)) # expectation scores_sq = scores**2 marg_num = batch_size * (batch_size - 1.) marg_term = tf.reduce_sum(input_tensor=scores_sq) - tf.reduce_sum(input_tensor=tf.linalg.diag_part(scores_sq)) marg_term = marg_term / marg_num # tf.print(joint_term - 0.5*marg_term) return joint_term - 0.5*marg_term # nll = tf.reduce_mean(tf.linalg.diag_part(scores) - 0.5 * tf.math.reduce_euclidean_norm(scores, axis=1)) # tf.print(nll) # mi = tf.math.log(tf.cast(scores.shape[0].value, tf.float32)) + nll # return mi @tf.function def skew_js_fgan_lower_bound(f): """skewed js lower bound (true cross entropy)""" n = tf.cast(f.shape[0], tf.float32) alpha = 1/n f_diag = tf.linalg.tensor_diag_part(f) first_term = tf.reduce_mean(-tf.nn.softplus(-f_diag)) second_term = (tf.reduce_sum(tf.nn.softplus(f)) - tf.reduce_sum(tf.nn.softplus(f_diag))) / (n * (n - 1.)) return alpha*first_term - (1-alpha)*second_term @tf.function def label_smooth_pcc(f): """ pcc with label smoothing trick""" n = f.shape[0] labels = tf.eye(n) labels = tf.reshape(labels,[-1,1]) labels = (1.0 - args.smoothing) * labels + args.smoothing / 2 pre_prob = tf.reshape(tf.sigmoid(f), [-1,1]) bce = tf.keras.losses.BinaryCrossentropy() return -bce(labels, pre_prob) @tf.function def predict_smooth_pcc(f): """ pcc with predictor smoothing trick""" n = f.shape[0] labels = tf.eye(n) labels = tf.reshape(labels,[-1,1]) # labels = (1.0 - args.smoothing) * labels + args.smoothing / 2 pre_prob = tf.reshape(tf.sigmoid(f), [-1,1]) pre_prob = (1.0 - args.smoothing) * pre_prob + args.smoothing / 2 bce = tf.keras.losses.BinaryCrossentropy() return -bce(labels, pre_prob) @tf.function def adap_label_smooth_pcc(f): n = f.shape[0] labels = tf.eye(n) labels = tf.reshape(labels,[-1,1]) # labels = (1.0 - args.smoothing) * labels + args.smoothing / 2 pre_prob = tf.reshape(tf.sigmoid(f), [-1,1]) alpha = acti_func(pre_prob, a, b, c) new_labels = (1.0 - alpha) * labels + alpha / 2 bce = tf.keras.losses.BinaryCrossentropy() return -bce(new_labels, pre_prob) @tf.function def adap_pred_smooth_pcc(f): n = f.shape[0] labels = tf.eye(n) labels = tf.reshape(labels,[-1,1]) # labels = (1.0 - args.smoothing) * labels + args.smoothing / 2 pre_prob = tf.reshape(tf.sigmoid(f), [-1,1]) # print('pre_prob:',pre_prob) alpha = acti_func(pre_prob, a, b, c) pre_prob = (1.0 - alpha) * pre_prob + alpha / 2 bce = tf.keras.losses.BinaryCrossentropy() return -bce(labels, pre_prob) # @title Define the linear evaluation protocol { display-mode: "form" } def logistic_fit(x_train, y_train, x_test, y_test): logistic_regressor = sk_linear.LogisticRegression( solver='saga', multi_class='multinomial', tol=.1, C=10.) from sklearn.preprocessing import MinMaxScaler scaler = Min
"""Convert results class to a data frame.""" label = "{}, {}, {}".format(nets, critic, loss) rows = list( zip( itertools.repeat(exp_name), itertools.repeat(nets), itertools.repeat(critic), itertools.repeat(loss), itertools.repeat(seed), result.iterations, [-loss for loss in result.testing_losses], # Loss -> bound. result.classification_accuracies, itertools.repeat(label))) df_eval = pd.DataFrame( rows, columns=("exp_name", "nets", "Critic", "Estimator", "run", "iteration", "bound_value", "accuracy", "label")) df_eval["Estimator"] = df_eval["Estimator"].replace( to_replace={
identifier_body
main.py
def skew_js_fgan_lower_bound(f): """skewed js lower bound (true cross entropy)""" n = tf.cast(f.shape[0], tf.float32) alpha = 1/n f_diag = tf.linalg.tensor_diag_part(f) first_term = tf.reduce_mean(-tf.nn.softplus(-f_diag)) second_term = (tf.reduce_sum(tf.nn.softplus(f)) - tf.reduce_sum(tf.nn.softplus(f_diag))) / (n * (n - 1.)) return alpha*first_term - (1-alpha)*second_term @tf.function def label_smooth_pcc(f): """ pcc with label smoothing trick""" n = f.shape[0] labels = tf.eye(n) labels = tf.reshape(labels,[-1,1]) labels = (1.0 - args.smoothing) * labels + args.smoothing / 2 pre_prob = tf.reshape(tf.sigmoid(f), [-1,1]) bce = tf.keras.losses.BinaryCrossentropy() return -bce(labels, pre_prob) @tf.function def predict_smooth_pcc(f): """ pcc with predictor smoothing trick""" n = f.shape[0] labels = tf.eye(n) labels = tf.reshape(labels,[-1,1]) # labels = (1.0 - args.smoothing) * labels + args.smoothing / 2 pre_prob = tf.reshape(tf.sigmoid(f), [-1,1]) pre_prob = (1.0 - args.smoothing) * pre_prob + args.smoothing / 2 bce = tf.keras.losses.BinaryCrossentropy() return -bce(labels, pre_prob) @tf.function def adap_label_smooth_pcc(f): n = f.shape[0] labels = tf.eye(n) labels = tf.reshape(labels,[-1,1]) # labels = (1.0 - args.smoothing) * labels + args.smoothing / 2 pre_prob = tf.reshape(tf.sigmoid(f), [-1,1]) alpha = acti_func(pre_prob, a, b, c) new_labels = (1.0 - alpha) * labels + alpha / 2 bce = tf.keras.losses.BinaryCrossentropy() return -bce(new_labels, pre_prob) @tf.function def adap_pred_smooth_pcc(f): n = f.shape[0] labels = tf.eye(n) labels = tf.reshape(labels,[-1,1]) # labels = (1.0 - args.smoothing) * labels + args.smoothing / 2 pre_prob = tf.reshape(tf.sigmoid(f), [-1,1]) # print('pre_prob:',pre_prob) alpha = acti_func(pre_prob, a, b, c) pre_prob = (1.0 - alpha) * pre_prob + alpha / 2 bce = tf.keras.losses.BinaryCrossentropy() return -bce(labels, pre_prob) # @title Define the linear evaluation protocol { display-mode: "form" } def logistic_fit(x_train, y_train, x_test, y_test): logistic_regressor = sk_linear.LogisticRegression( solver='saga', multi_class='multinomial', tol=.1, C=10.) from sklearn.preprocessing import MinMaxScaler scaler = MinMaxScaler() x_train = scaler.fit_transform(x_train) x_test = scaler.transform(x_test) logistic_regressor.fit(x_train, y_train.ravel()) return logistic_regressor.score(x_test, y_test.ravel()) # @title Define and load the dataset, check baseline in pixel space { display-mode: "form" } tf.compat.v1.reset_default_graph() def map_fn(example): image = example[FEATURE_INPUT] image = tf.cast(image, tf.float32) / 255.0 image = tf.reshape(image, [-1]) # Flatten. label = example[FEATURE_LABEL] return {FEATURE_INPUT: image, FEATURE_LABEL: label} def load_data(split): return (tfds.load(TFDS_NAME, data_dir='/public/wangxu/data/', split=split) .cache() .map(map_func=map_fn) .shuffle(1000)) def tfds_to_np(dataset): features = list(tfds.as_numpy(dataset)) images = np.stack([f[FEATURE_INPUT].ravel() for f in features]) labels = np.stack([f[FEATURE_LABEL].ravel() for f in features]) return images, labels dataset_train = load_data("train") dataset_test = load_data("test") x_train, y_train = tfds_to_np(dataset_train) x_test, y_test = tfds_to_np(dataset_test) tf.compat.v1.reset_default_graph() x_train_noisy = x_train + 0.05 * np.random.randn(*x_train.shape) x_test_noisy = x_test + 0.05 * np.random.randn(*x_test.shape) print("Fit on half the pixels: {}. It should be around 0.835.".format( logistic_fit(x_train_noisy[:, :DIMS//2], y_train, x_test_noisy[:, :DIMS//2], y_test))) def processed_train_data(dims, batch_size): dataset = load_data("train") dataset_batched = dataset.repeat().batch(batch_size, drop_remainder=True) get_next = tf.compat.v1.data.make_one_shot_iterator(dataset_batched).get_next() features = get_next[FEATURE_INPUT] labels = get_next[FEATURE_LABEL] # Martin: where the problem occurs x_1, x_2 = tf.split(features, [dims, DIMS-dims], axis=-1) return x_1, x_2, labels class MLP(tf.keras.Model): def __init__(self, layer_dimensions, shortcuts, dense_kwargs={}): super(MLP, self).__init__() self._layers = [tfkl.Dense(dimensions, **dense_kwargs) for dimensions in layer_dimensions[:-1]] dense_kwargs_copy = copy.deepcopy(dense_kwargs) dense_kwargs_copy["activation"] = None self._layers.append(tfkl.Dense(layer_dimensions[-1], **dense_kwargs_copy)) self._shortcuts = shortcuts @property def layers(self): return self._layers def __call__(self, inputs): x = inputs for layer in self.layers: x = layer(x) + x if self._shortcuts else layer(x) return x # LayerNorm implementation copied from # https://stackoverflow.com/questions/39095252/fail-to-implement-layer-normalization-with-keras class LayerNorm(tfkl.Layer): """ Layer Normalization in the style of https://arxiv.org/abs/1607.06450 """ def __init__(self, scale_initializer='ones', bias_initializer='zeros', axes=[1,2,3], epsilon=1e-6, **kwargs): super(LayerNorm, self).__init__(**kwargs) self.epsilon = epsilon self.scale_initializer = tf.keras.initializers.get(scale_initializer) self.bias_initializer = tf.keras.initializers.get(bias_initializer) self.axes = axes def build(self, input_shape): self.scale = self.add_weight(shape=(input_shape[-1],), initializer=self.scale_initializer, trainable=True, name='{}_scale'.format(self.name)) self.bias = self.add_weight(shape=(input_shape[-1],), initializer=self.bias_initializer, trainable=True, name='{}_bias'.format(self.name)) self.built = True def call(self, x, mask=None): mean = tf.keras.backend.mean(x, axis=self.axes, keepdims=True) std = tf.keras.backend.std(x, axis=self.axes, keepdims=True) norm = (x - mean) * (1/(std + self.epsilon)) return norm * self.scale + self.bias def compute_output_shape(self, input_shape): return input_shape class ConvNet(tf.keras.Sequential): def __init__(self, channels=64, kernel_size=5, input_dim=DIMS//2, output_dim=100, activation=tf.nn.relu): # Note: This works only for the specific data set considered here. super(ConvNet, self).__init__([ tfkl.Reshape((14, 28, 1), input_shape=(input_dim,)), tfkl.Conv2D(channels, kernel_size, strides=2, padding="same", activation=activation), tfkl.Conv2D(2*channels, kernel_size, strides=2, padding="same", activation=activation), LayerNorm(), tfkl.GlobalAveragePooling2D(), tfkl.Dense(output_dim), ]) from tensorflow_probability.python.internal import tensorshape_util import tensorflow.compat.v1 as tf1 from tensorflow_probability.python.bijectors import affine_scalar from tensorflow_probability.python.bijectors import bijector as bijector_lib # Modified from tensorflow_probability/python/bijectors/real_nvp.py class RealNVP(bijector_lib.Bijector): def __init__(self, num_masked, shift_and_log_scale_fn=None, bijector_fn=None, is_constant_jacobian=False, validate_args=False, name=None): name = name or 'real_nvp' if num_masked < 0:
raise ValueError('num_masked must be a non-negative integer.')
conditional_block
main.py
n - 1.)) return first_term - second_term @tf.function def infonce_lower_bound(scores): """InfoNCE lower bound from van den Oord et al. (2018).""" nll = tf.reduce_mean(input_tensor=tf.linalg.diag_part(scores) - tf.reduce_logsumexp(input_tensor=scores, axis=1)) mi = tf.math.log(tf.cast(scores.shape[0], tf.float32)) + nll return mi @tf.function def our_lower_bound(scores): # scores: 128, 128 """Our lower bound""" batch_size = tf.cast(scores.shape[0], tf.float32) joint_term = tf.reduce_mean(input_tensor=tf.linalg.diag_part(scores)) # expectation scores_sq = scores**2 marg_num = batch_size * (batch_size - 1.) marg_term = tf.reduce_sum(input_tensor=scores_sq) - tf.reduce_sum(input_tensor=tf.linalg.diag_part(scores_sq)) marg_term = marg_term / marg_num # tf.print(joint_term - 0.5*marg_term) return joint_term - 0.5*marg_term # nll = tf.reduce_mean(tf.linalg.diag_part(scores) - 0.5 * tf.math.reduce_euclidean_norm(scores, axis=1)) # tf.print(nll) # mi = tf.math.log(tf.cast(scores.shape[0].value, tf.float32)) + nll # return mi @tf.function def skew_js_fgan_lower_bound(f): """skewed js lower bound (true cross entropy)""" n = tf.cast(f.shape[0], tf.float32) alpha = 1/n f_diag = tf.linalg.tensor_diag_part(f) first_term = tf.reduce_mean(-tf.nn.softplus(-f_diag)) second_term = (tf.reduce_sum(tf.nn.softplus(f)) - tf.reduce_sum(tf.nn.softplus(f_diag))) / (n * (n - 1.)) return alpha*first_term - (1-alpha)*second_term @tf.function def label_smooth_pcc(f): """ pcc with label smoothing trick""" n = f.shape[0] labels = tf.eye(n) labels = tf.reshape(labels,[-1,1]) labels = (1.0 - args.smoothing) * labels + args.smoothing / 2 pre_prob = tf.reshape(tf.sigmoid(f), [-1,1]) bce = tf.keras.losses.BinaryCrossentropy() return -bce(labels, pre_prob) @tf.function def predict_smooth_pcc(f): """ pcc with predictor smoothing trick""" n = f.shape[0] labels = tf.eye(n) labels = tf.reshape(labels,[-1,1]) # labels = (1.0 - args.smoothing) * labels + args.smoothing / 2 pre_prob = tf.reshape(tf.sigmoid(f), [-1,1]) pre_prob = (1.0 - args.smoothing) * pre_prob + args.smoothing / 2 bce = tf.keras.losses.BinaryCrossentropy() return -bce(labels, pre_prob) @tf.function def adap_label_smooth_pcc(f): n = f.shape[0] labels = tf.eye(n) labels = tf.reshape(labels,[-1,1]) # labels = (1.0 - args.smoothing) * labels + args.smoothing / 2 pre_prob = tf.reshape(tf.sigmoid(f), [-1,1]) alpha = acti_func(pre_prob, a, b, c) new_labels = (1.0 - alpha) * labels + alpha / 2 bce = tf.keras.losses.BinaryCrossentropy() return -bce(new_labels, pre_prob) @tf.function def adap_pred_smooth_pcc(f): n = f.shape[0] labels = tf.eye(n) labels = tf.reshape(labels,[-1,1]) # labels = (1.0 - args.smoothing) * labels + args.smoothing / 2 pre_prob = tf.reshape(tf.sigmoid(f), [-1,1]) # print('pre_prob:',pre_prob) alpha = acti_func(pre_prob, a, b, c) pre_prob = (1.0 - alpha) * pre_prob + alpha / 2 bce = tf.keras.losses.BinaryCrossentropy() return -bce(labels, pre_prob) # @title Define the linear evaluation protocol { display-mode: "form" } def logistic_fit(x_train, y_train, x_test, y_test): logistic_regressor = sk_linear.LogisticRegression( solver='saga', multi_class='multinomial', tol=.1, C=10.) from sklearn.preprocessing import MinMaxScaler scaler = MinMaxScaler() x_train = scaler.fit_transform(x_train) x_test = scaler.transform(x_test) logistic_regressor.fit(x_train, y_train.ravel()) return logistic_regressor.score(x_test, y_test.ravel()) # @title Define and load the dataset, check baseline in pixel space { display-mode: "form" } tf.compat.v1.reset_default_graph() def map_fn(example): image = example[FEATURE_INPUT] image = tf.cast(image, tf.float32) / 255.0 image = tf.reshape(image, [-1]) # Flatten. label = example[FEATURE_LABEL] return {FEATURE_INPUT: image, FEATURE_LABEL: label} def load_data(split): return (tfds.load(TFDS_NAME, data_dir='/public/wangxu/data/', split=split) .cache() .map(map_func=map_fn) .shuffle(1000)) def tfds_to_np(dataset): features = list(tfds.as_numpy(dataset)) images = np.stack([f[FEATURE_INPUT].ravel() for f in features]) labels = np.stack([f[FEATURE_LABEL].ravel() for f in features]) return images, labels dataset_train = load_data("train") dataset_test = load_data("test") x_train, y_train = tfds_to_np(dataset_train) x_test, y_test = tfds_to_np(dataset_test) tf.compat.v1.reset_default_graph() x_train_noisy = x_train + 0.05 * np.random.randn(*x_train.shape) x_test_noisy = x_test + 0.05 * np.random.randn(*x_test.shape) print("Fit on half the pixels: {}. It should be around 0.835.".format( logistic_fit(x_train_noisy[:, :DIMS//2], y_train, x_test_noisy[:, :DIMS//2], y_test))) def processed_train_data(dims, batch_size): dataset = load_data("train") dataset_batched = dataset.repeat().batch(batch_size, drop_remainder=True) get_next = tf.compat.v1.data.make_one_shot_iterator(dataset_batched).get_next() features = get_next[FEATURE_INPUT] labels = get_next[FEATURE_LABEL] # Martin: where the problem occurs x_1, x_2 = tf.split(features, [dims, DIMS-dims], axis=-1) return x_1, x_2, labels class MLP(tf.keras.Model): def __init__(self, layer_dimensions, shortcuts, dense_kwargs={}): super(MLP, self).__init__() self._layers = [tfkl.Dense(dimensions, **dense_kwargs) for dimensions in layer_dimensions[:-1]] dense_kwargs_copy = copy.deepcopy(dense_kwargs) dense_kwargs_copy["activation"] = None self._layers.append(tfkl.Dense(layer_dimensions[-1], **dense_kwargs_copy)) self._shortcuts = shortcuts @property def layers(self): return self._layers def __call__(self, inputs): x = inputs for layer in self.layers: x = layer(x) + x if self._shortcuts else layer(x) return x # LayerNorm implementation copied from # https://stackoverflow.com/questions/39095252/fail-to-implement-layer-normalization-with-keras class LayerNorm(tfkl.Layer): """ Layer Normalization in the style of https://arxiv.org/abs/1607.06450 """ def __init__(self, scale_initializer='ones', bias_initializer='zeros', axes=[1,2,3], epsilon=1e-6, **kwargs): super(LayerNorm, self).__init__(**kwargs) self.epsilon = epsilon self.scale_initializer = tf.keras.initializers.get(scale_initializer) self.bias_initializer = tf.keras.initializers.get(bias_initializer) self.axes = axes def build(self, input_shape): self.scale = self.add_weight(shape=(input_shape[-1],), initializer=self.scale_initializer, trainable=True, name='{}_scale'.format(self.name)) self.bias = self.add_weight(shape=(input_shape[-1],), initializer=self.bias_initializer, trainable=True, name='{}_bias'.format(self.name)) self.built = True def call(self, x, mask=None): mean = tf.keras.backend.mean(x, axis=self.axes, keepdims=True) std = tf.keras.backend.std(x, axis=self.axes, keepdims=True) norm = (x - mean) * (1/(std + self.epsilon)) return norm * self.scale + self.bias def
compute_output_shape
identifier_name
main.py
:] return self._bijector_fn(y0, self._input_depth - self._num_masked, **condition_kwargs).inverse_log_det_jacobian( y1, event_ndims=1) def real_nvp_default_template(hidden_layers, shift_only=False, activation=tf.nn.relu, name=None, *args, # pylint: disable=keyword-arg-before-vararg **kwargs): with tf.compat.v1.name_scope(name or 'real_nvp_default_template'): def _fn(x, output_units, **condition_kwargs): """Fully connected MLP parameterized via `real_nvp_template`.""" if condition_kwargs: raise NotImplementedError( 'Conditioning not implemented in the default template.') if tensorshape_util.rank(x.shape) == 1: x = x[tf.newaxis, ...] reshape_output = lambda x: x[0] else: reshape_output = lambda x: x for units in hidden_layers: x = tf1.layers.dense( inputs=x, units=units, activation=activation, *args, # pylint: disable=keyword-arg-before-vararg **kwargs) x = tf1.layers.dense( inputs=x, units=(1 if shift_only else 2) * output_units, activation=None, *args, # pylint: disable=keyword-arg-before-vararg **kwargs) if shift_only: return reshape_output(x), None shift, log_scale = tf.split(x, 2, axis=-1) # ** Here is the second modification. return reshape_output(shift), 1e-7 + tf.nn.softplus(reshape_output(log_scale)) return tf1.make_template('real_nvp_default_template', _fn) class RealNVPBijector(tf.keras.Model): def __init__(self, dimensions, n_couplings, hidden_layers, dense_kwargs): super(RealNVPBijector, self).__init__() permutations = [np.random.permutation(dimensions) for _ in range(n_couplings)] bijectors = [] for permutation in permutations: bijectors.append(RealNVP( dimensions // 2, real_nvp_default_template(hidden_layers, **dense_kwargs))) bijectors.append(tfb.Permute(permutation)) self._bijector = tfb.Chain(bijectors) def call(self, inputs): return self._bijector.forward(inputs) class InnerProdCritic(tf.keras.Model): def call(self, x, y): return tf.matmul(x, y, transpose_b=True) class BilinearCritic(tf.keras.Model): def __init__(self, feature_dim=100, **kwargs): super(BilinearCritic, self).__init__(**kwargs) self._W = tfkl.Dense(feature_dim, use_bias=False) def call(self, x, y): return tf.matmul(x, self._W(y), transpose_b=True) # https://colab.research.google.com/github/google-research/google-research/blob/master/vbmi/vbmi_demo.ipynb class ConcatCritic(tf.keras.Model): def __init__(self, hidden_dim=200, layers=1, activation='relu', **kwargs): super(ConcatCritic, self).__init__(**kwargs) # output is scalar score self._f = MLP([hidden_dim for _ in range(layers)]+[1], False, {"activation": "relu"}) def call(self, x, y): batch_size = tf.shape(input=x)[0] # Tile all possible combinations of x and y x_tiled = tf.tile(x[None, :], (batch_size, 1, 1)) y_tiled = tf.tile(y[:, None], (1, batch_size, 1)) # xy is [batch_size * batch_size, x_dim + y_dim] xy_pairs = tf.reshape(tf.concat((x_tiled, y_tiled), axis=2), [batch_size * batch_size, -1]) # Compute scores for each x_i, y_j pair. scores = self._f(xy_pairs) return tf.transpose(a=tf.reshape(scores, [batch_size, batch_size])) class SeparableCritic(tf.keras.Model): def __init__(self, hidden_dim=100, output_dim=100, layers=1, activation='relu', **kwargs): super(SeparableCritic, self).__init__(**kwargs) self._f_x = MLP([hidden_dim for _ in range(layers)] + [output_dim], False, {"activation": activation}) self._f_y = MLP([hidden_dim for _ in range(layers)] + [output_dim], False, {"activation": activation}) def call(self, x, y): x_mapped = self._f_x(x) y_mapped = self._f_y(y) return tf.matmul(x_mapped, y_mapped, transpose_b=True) def train(g1, g2, critic, loss_fn, learning_rate, batch_size=TRAIN_BATCH_SIZE, n_iters=15000, n_evals=15, compute_jacobian=False, noise_std=0.0, data_dimensions=DIMS//2, n_iter=1, loss_name='InfoNCE', ): """Runs the training loop for a fixed model. Args: g1: Function, maps input1 to representation. g2: Function, maps input2 to representation. critic: Function, maps two representations to scalar. loss_fn: Function, mutual information estimator. learning_rate: Learning rate. batch_size: Training batch size. n_iters: Number of optimization iterations. n_evals: Number of model evaluations. compute_jacobian: Whether to estimate the singular values of the Jacobian. noise_std: Standard deviation for the Gaussian noise. Default is 0.0. data_dimensions: The dimension of the data. By default it's half of the original data dimension. Returns: Returns and instance of `Results` tuple. """ x_1, x_2, _ = processed_train_data(data_dimensions, batch_size) if noise_std > 0.0: assert x_1.shape == x_2.shape, "X1 and X2 shapes must agree to add noise!" noise = noise_std * tf.random.normal(x_1.shape) x_1 += noise x_2 += noise # Compute the representations. code_1, code_2 = g1(x_1), g2(x_2) critic_matrix = critic(code_1, code_2) # Compute the Jacobian of g1 if needed. if compute_jacobian: jacobian = gradients.batch_jacobian(code_1, x_1, use_pfor=False) singular_values = tf.linalg.svd(jacobian, compute_uv=False) # Optimizer setup. loss = loss_fn(critic_matrix) optimizer = tf.compat.v1.train.AdamOptimizer(learning_rate=learning_rate) if not loss_name == 'wpc': optimizer_op = optimizer.minimize(loss) else: gvs = optimizer.compute_gradients(loss) capped_gvs = [(tf.clip_by_value(grad, -1., 1.), var) for grad, var in gvs] optimizer_op = optimizer.apply_gradients(capped_gvs) with tf.compat.v1.Session() as session: session.run(tf.compat.v1.global_variables_initializer()) # Subgraph for eval (add noise to input if necessary) data_ph = tf.compat.v1.placeholder(tf.float32, shape=[None, data_dimensions]) data_ph_noisy = data_ph + noise_std * tf.random.normal(tf.shape(input=data_ph)) codes = g1(data_ph_noisy) training_losses, testing_losses, classification_accuracies, iters, sigmas \ = [], [], [], [], [] # Main training loop. for iter_n in range(n_iters): # Evaluate the model performance. if iter_n % (n_iters // n_evals) == 0: iters.append(iter_n) accuracy = get_classification_accuracy(session, codes, data_ph, data_dimensions) classification_accuracies.append(accuracy) testing_losses.append( get_testing_loss(x_test, session, loss, data_ph, data_dimensions)) if compute_jacobian: sigmas.append(session.run(singular_values)) print("{:d}th iter Loss_name {} Step {:>10d} fit {:>.5f} DS {} B {:d} lr {:f}".format(\ n_iter, loss_name, iter_n, accuracy, args.dataset, args.batch_size, args.lr)) # Run one optimization step. loss_np, _ = session.run([loss, optimizer_op]) training_losses.append(loss_np) return Results(iterations=iters, training_losses=training_losses, testing_losses=testing_losses, classification_accuracies=classification_accuracies, singular_values=sigmas) def run_sweep(nets, critics, loss_fns, exp_name, **kwargs): """Runs the sweep across encoder networks, critics, and the estimators.""" grid = itertools.product(nets, critics, loss_fns) data_frames = [] results_with_singular_values = []
for nets_name, critic_name, loss_name in grid: print("[New experiment] encoder: {}, critic: {}, loss: {}".format( nets_name, critic_name, loss_name))
random_line_split
preprocessing.py
higher than point n. # (i.e. contacts can only decrease when increasing distance) if smooth and mat_n > 2: ir = IsotonicRegression(increasing=False) dist[~np.isfinite(dist)] = 0 dist = ir.fit_transform(range(len(dist)), dist) return dist def get_detectable_bins(mat, n_mads=3, inter=False): """ Returns lists of detectable indices after excluding low interacting bins based on the proportion of zero pixel values in the matrix bins. Parameters ---------- mat : scipy.sparse.coo_matrix A Hi-C matrix in tihe form of a 2D numpy array or coo matrix n_mads : int Number of median absolute deviation below the median required to consider bins non-detectable. inter : bool Whether the matrix is interchromosomal. Default is to consider the matrix is intrachromosomal (i.e. upper symmetric). Returns ------- numpy ndarray : tuple of 2 1D arrays containing indices of low interacting rows and columns, respectively. """ matrix = mat.copy() matrix.eliminate_zeros() def
(x): return ss.median_abs_deviation(x, nan_policy="omit") if not inter: if matrix.shape[0] != matrix.shape[1]: raise ValueError("Intrachromosomal matrices must be symmetric.") # Replace nonzero pixels by ones to work on prop. of nonzero pixels matrix.data = np.ones(matrix.data.shape) # Compute number of nonzero values in each bin sum_bins = sum_mat_bins(matrix) # Compute variation in the number of nonzero pixels sum_mad = mad(sum_bins) # Find poor interacting rows and columns sum_med = np.median(sum_bins) detect_threshold = max(1, sum_med - sum_mad * n_mads) # Removal of poor interacting rows and columns good_bins = np.flatnonzero(sum_bins >= detect_threshold) good_bins = (good_bins, good_bins) else: # Adapted for asymetric matrices (need to compute rows and columns) sum_rows, sum_cols = matrix.sum(axis=1).A1, matrix.sum(axis=0).A1 mad_rows, mad_cols = mad(sum_rows), mad(sum_cols) med_rows, med_cols = np.median(sum_rows), np.median(sum_cols) detect_threshold_rows = max(1, med_rows - mad_rows * n_mads) detect_threshold_cols = max(1, med_cols - mad_cols * n_mads) good_rows = np.flatnonzero(sum_rows > detect_threshold_rows) good_cols = np.flatnonzero(sum_cols > detect_threshold_cols) good_bins = (good_rows, good_cols) return good_bins def detrend( matrix, detectable_bins=None, max_dist=None, smooth=False, fun=np.nanmean, max_val=10, ): """ Detrends a Hi-C matrix by the distance law. The input matrix should have been normalised beforehandand. Parameters ---------- matrix : scipy.sparse.csr_matrix The normalised intrachromosomal Hi-C matrix to detrend. detectable_bins : tuple Tuple containing a list of detectable rows and a list of columns on which to perform detrending. Poorly interacting indices have been excluded. max_dist : int Maximum number of bins from the diagonal at which to compute trend. smooth : bool Whether to use isotonic regression to smooth the trend. fun : callable Function to use on each diagonal to compute the trend. max_val : float or None Maximum value in the detrended matrix. Set to None to disable Returns ------- numpy.ndarray : The detrended Hi-C matrix. """ matrix = matrix.tocsr() y = distance_law( matrix, detectable_bins=detectable_bins, max_dist=max_dist, smooth=smooth, fun=fun, ) y[np.isnan(y)] = 0.0 # Detrending by the distance law clean_mat = matrix.tocoo() # clean_mat.data /= y_savgol[abs(clean_mat.row - clean_mat.col)] try: clean_mat.data = clean_mat.data / y[abs(clean_mat.row - clean_mat.col)] # If no nonzero value in matrix, do nothing except TypeError: pass clean_mat = clean_mat.tocsr() if max_val is not None: clean_mat[clean_mat >= max_val] = 1 return clean_mat def ztransform(matrix): """ Z transformation for Hi-C matrices. Parameters ---------- matrix : scipy.sparse.coo_matrix A Hi-C matrix in sparse format. Returns ------- scipy.sparse.coo_matrix: The detrended Hi-C matrix """ mat = matrix.copy() mu = np.mean(mat.data) sd = np.std(mat.data) mat.data -= mu mat.data /= sd return mat def sum_mat_bins(mat): """ Compute the sum of matrices bins (i.e. rows or columns) using only the upper triangle, assuming symmetrical matrices. Parameters ---------- mat : scipy.sparse.coo_matrix Contact map in sparse format, either in upper triangle or full matrix. Returns ------- numpy.ndarray : 1D array of bin sums. """ # Equivalaent to row or col sum on a full matrix # Note: mat.sum returns a 'matrix' object. A1 extracts the 1D flat array # from the matrix return mat.sum(axis=0).A1 + mat.sum(axis=1).A1 - mat.diagonal(0) def subsample_contacts(M, n_contacts): """Bootstrap sampling of contacts in a sparse Hi-C map. Parameters ---------- M : scipy.sparse.coo_matrix The input Hi-C contact map in sparse format. n_contacts : int The number of contacts to sample. Returns ------- scipy.sparse.coo_matrix A new matrix with a fraction of the original contacts. """ S = M.data.copy() # Match cell idx to cumulative number of contacts cum_counts = np.cumsum(S) # Total number of contacts to sample tot_contacts = int(cum_counts[-1]) # Sample desired number of contacts from the range(0, n_contacts) array sampled_contacts = np.random.choice( int(tot_contacts), size=(n_contacts), replace=False ) # Get indices of sampled contacts in the cum_counts array idx = np.searchsorted(cum_counts, sampled_contacts, side="right") # Bin those indices to the same dimensions as matrix data to get counts sampled_counts = np.bincount(idx, minlength=S.shape[0]) # Get nonzero values to build new sparse matrix nnz_mask = sampled_counts > 0 sampled_counts = sampled_counts[nnz_mask].astype(np.float64) sampled_rows = M.row[nnz_mask] sampled_cols = M.col[nnz_mask] return sp.coo_matrix( (sampled_counts, (sampled_rows, sampled_cols)), shape=(M.shape[0], M.shape[1]), ) def frame_missing_mask(mask, kernel_shape, sym_upper=False, max_dist=None): """ Adds a frame around input mask, given a kernel. The goal of this frame is define margins around the matrix where the kernel will not perform convolution (denoted by 1). If the matrix is upper symmetric, a margin of half the kernel's width is added below the diagonal and a maximum distance from the diagonal above which margins need not be drawn can be considered. Otherwise Margins are simply added on all 4 sides of the matrix. :: signal kernel _________ ______ ____ |#######| | | | | ==> |# #| | | |___| |# #| | | |# #| |_____| |# #| |#######| -------- Parameters ---------- mask : scipy.sparse.csr_matrix of bool The mask around which to add margins. kernels_shape : tuple of ints The number of rows and kernel in the input kernel. Margins will be half these values. sym_upper : bool Whether the signal is a symmetric upper triangle matrix. If so, values on a margin below the diagonal will be masked. max_dist : int or None Number of diagonals to keep Returns ------- framed_mask : scipy.sparse.csr_matrix of bool The input mask with a padding of True around the edges. If sym_upper is True, a padding is also added below the diagonal. """ if mask.dtype != bool: raise ValueError("Mask must contain boolean values") if not sp.issparse(mask): raise ValueError("Mask must be a sparse matrix") framed_mask = mask.copy() ms, ns = mask.shape mk, nk = kernel_shape if sym_upper and (max_dist is not None): # Remove diagonals further than scan distance in the input mask framed_mask = diag_trim(framed_mask, max_dist + max(nk
mad
identifier_name
preprocessing.py
higher than point n. # (i.e. contacts can only decrease when increasing distance) if smooth and mat_n > 2: ir = IsotonicRegression(increasing=False) dist[~np.isfinite(dist)] = 0 dist = ir.fit_transform(range(len(dist)), dist) return dist def get_detectable_bins(mat, n_mads=3, inter=False): """ Returns lists of detectable indices after excluding low interacting bins based on the proportion of zero pixel values in the matrix bins. Parameters ---------- mat : scipy.sparse.coo_matrix A Hi-C matrix in tihe form of a 2D numpy array or coo matrix n_mads : int Number of median absolute deviation below the median required to consider bins non-detectable. inter : bool Whether the matrix is interchromosomal. Default is to consider the matrix is intrachromosomal (i.e. upper symmetric). Returns ------- numpy ndarray : tuple of 2 1D arrays containing indices of low interacting rows and columns, respectively. """ matrix = mat.copy() matrix.eliminate_zeros() def mad(x): return ss.median_abs_deviation(x, nan_policy="omit") if not inter: if matrix.shape[0] != matrix.shape[1]: raise ValueError("Intrachromosomal matrices must be symmetric.") # Replace nonzero pixels by ones to work on prop. of nonzero pixels matrix.data = np.ones(matrix.data.shape) # Compute number of nonzero values in each bin sum_bins = sum_mat_bins(matrix) # Compute variation in the number of nonzero pixels sum_mad = mad(sum_bins) # Find poor interacting rows and columns sum_med = np.median(sum_bins) detect_threshold = max(1, sum_med - sum_mad * n_mads) # Removal of poor interacting rows and columns good_bins = np.flatnonzero(sum_bins >= detect_threshold) good_bins = (good_bins, good_bins) else: # Adapted for asymetric matrices (need to compute rows and columns)
return good_bins def detrend( matrix, detectable_bins=None, max_dist=None, smooth=False, fun=np.nanmean, max_val=10, ): """ Detrends a Hi-C matrix by the distance law. The input matrix should have been normalised beforehandand. Parameters ---------- matrix : scipy.sparse.csr_matrix The normalised intrachromosomal Hi-C matrix to detrend. detectable_bins : tuple Tuple containing a list of detectable rows and a list of columns on which to perform detrending. Poorly interacting indices have been excluded. max_dist : int Maximum number of bins from the diagonal at which to compute trend. smooth : bool Whether to use isotonic regression to smooth the trend. fun : callable Function to use on each diagonal to compute the trend. max_val : float or None Maximum value in the detrended matrix. Set to None to disable Returns ------- numpy.ndarray : The detrended Hi-C matrix. """ matrix = matrix.tocsr() y = distance_law( matrix, detectable_bins=detectable_bins, max_dist=max_dist, smooth=smooth, fun=fun, ) y[np.isnan(y)] = 0.0 # Detrending by the distance law clean_mat = matrix.tocoo() # clean_mat.data /= y_savgol[abs(clean_mat.row - clean_mat.col)] try: clean_mat.data = clean_mat.data / y[abs(clean_mat.row - clean_mat.col)] # If no nonzero value in matrix, do nothing except TypeError: pass clean_mat = clean_mat.tocsr() if max_val is not None: clean_mat[clean_mat >= max_val] = 1 return clean_mat def ztransform(matrix): """ Z transformation for Hi-C matrices. Parameters ---------- matrix : scipy.sparse.coo_matrix A Hi-C matrix in sparse format. Returns ------- scipy.sparse.coo_matrix: The detrended Hi-C matrix """ mat = matrix.copy() mu = np.mean(mat.data) sd = np.std(mat.data) mat.data -= mu mat.data /= sd return mat def sum_mat_bins(mat): """ Compute the sum of matrices bins (i.e. rows or columns) using only the upper triangle, assuming symmetrical matrices. Parameters ---------- mat : scipy.sparse.coo_matrix Contact map in sparse format, either in upper triangle or full matrix. Returns ------- numpy.ndarray : 1D array of bin sums. """ # Equivalaent to row or col sum on a full matrix # Note: mat.sum returns a 'matrix' object. A1 extracts the 1D flat array # from the matrix return mat.sum(axis=0).A1 + mat.sum(axis=1).A1 - mat.diagonal(0) def subsample_contacts(M, n_contacts): """Bootstrap sampling of contacts in a sparse Hi-C map. Parameters ---------- M : scipy.sparse.coo_matrix The input Hi-C contact map in sparse format. n_contacts : int The number of contacts to sample. Returns ------- scipy.sparse.coo_matrix A new matrix with a fraction of the original contacts. """ S = M.data.copy() # Match cell idx to cumulative number of contacts cum_counts = np.cumsum(S) # Total number of contacts to sample tot_contacts = int(cum_counts[-1]) # Sample desired number of contacts from the range(0, n_contacts) array sampled_contacts = np.random.choice( int(tot_contacts), size=(n_contacts), replace=False ) # Get indices of sampled contacts in the cum_counts array idx = np.searchsorted(cum_counts, sampled_contacts, side="right") # Bin those indices to the same dimensions as matrix data to get counts sampled_counts = np.bincount(idx, minlength=S.shape[0]) # Get nonzero values to build new sparse matrix nnz_mask = sampled_counts > 0 sampled_counts = sampled_counts[nnz_mask].astype(np.float64) sampled_rows = M.row[nnz_mask] sampled_cols = M.col[nnz_mask] return sp.coo_matrix( (sampled_counts, (sampled_rows, sampled_cols)), shape=(M.shape[0], M.shape[1]), ) def frame_missing_mask(mask, kernel_shape, sym_upper=False, max_dist=None): """ Adds a frame around input mask, given a kernel. The goal of this frame is define margins around the matrix where the kernel will not perform convolution (denoted by 1). If the matrix is upper symmetric, a margin of half the kernel's width is added below the diagonal and a maximum distance from the diagonal above which margins need not be drawn can be considered. Otherwise Margins are simply added on all 4 sides of the matrix. :: signal kernel _________ ______ ____ |#######| | | | | ==> |# #| | | |___| |# #| | | |# #| |_____| |# #| |#######| -------- Parameters ---------- mask : scipy.sparse.csr_matrix of bool The mask around which to add margins. kernels_shape : tuple of ints The number of rows and kernel in the input kernel. Margins will be half these values. sym_upper : bool Whether the signal is a symmetric upper triangle matrix. If so, values on a margin below the diagonal will be masked. max_dist : int or None Number of diagonals to keep Returns ------- framed_mask : scipy.sparse.csr_matrix of bool The input mask with a padding of True around the edges. If sym_upper is True, a padding is also added below the diagonal. """ if mask.dtype != bool: raise ValueError("Mask must contain boolean values") if not sp.issparse(mask): raise ValueError("Mask must be a sparse matrix") framed_mask = mask.copy() ms, ns = mask.shape mk, nk = kernel_shape if sym_upper and (max_dist is not None): # Remove diagonals further than scan distance in the input mask framed_mask = diag_trim(framed_mask, max_dist + max(nk
sum_rows, sum_cols = matrix.sum(axis=1).A1, matrix.sum(axis=0).A1 mad_rows, mad_cols = mad(sum_rows), mad(sum_cols) med_rows, med_cols = np.median(sum_rows), np.median(sum_cols) detect_threshold_rows = max(1, med_rows - mad_rows * n_mads) detect_threshold_cols = max(1, med_cols - mad_cols * n_mads) good_rows = np.flatnonzero(sum_rows > detect_threshold_rows) good_cols = np.flatnonzero(sum_cols > detect_threshold_cols) good_bins = (good_rows, good_cols)
conditional_block
preprocessing.py
higher than point n. # (i.e. contacts can only decrease when increasing distance) if smooth and mat_n > 2: ir = IsotonicRegression(increasing=False) dist[~np.isfinite(dist)] = 0 dist = ir.fit_transform(range(len(dist)), dist) return dist def get_detectable_bins(mat, n_mads=3, inter=False): """ Returns lists of detectable indices after excluding low interacting bins based on the proportion of zero pixel values in the matrix bins. Parameters ---------- mat : scipy.sparse.coo_matrix A Hi-C matrix in tihe form of a 2D numpy array or coo matrix n_mads : int Number of median absolute deviation below the median required to consider bins non-detectable. inter : bool Whether the matrix is interchromosomal. Default is to consider the matrix is intrachromosomal (i.e. upper symmetric). Returns ------- numpy ndarray : tuple of 2 1D arrays containing indices of low interacting rows and columns, respectively. """ matrix = mat.copy() matrix.eliminate_zeros() def mad(x): return ss.median_abs_deviation(x, nan_policy="omit") if not inter: if matrix.shape[0] != matrix.shape[1]: raise ValueError("Intrachromosomal matrices must be symmetric.") # Replace nonzero pixels by ones to work on prop. of nonzero pixels matrix.data = np.ones(matrix.data.shape) # Compute number of nonzero values in each bin sum_bins = sum_mat_bins(matrix) # Compute variation in the number of nonzero pixels sum_mad = mad(sum_bins) # Find poor interacting rows and columns sum_med = np.median(sum_bins) detect_threshold = max(1, sum_med - sum_mad * n_mads) # Removal of poor interacting rows and columns good_bins = np.flatnonzero(sum_bins >= detect_threshold) good_bins = (good_bins, good_bins) else: # Adapted for asymetric matrices (need to compute rows and columns) sum_rows, sum_cols = matrix.sum(axis=1).A1, matrix.sum(axis=0).A1 mad_rows, mad_cols = mad(sum_rows), mad(sum_cols) med_rows, med_cols = np.median(sum_rows), np.median(sum_cols) detect_threshold_rows = max(1, med_rows - mad_rows * n_mads) detect_threshold_cols = max(1, med_cols - mad_cols * n_mads) good_rows = np.flatnonzero(sum_rows > detect_threshold_rows) good_cols = np.flatnonzero(sum_cols > detect_threshold_cols) good_bins = (good_rows, good_cols) return good_bins def detrend( matrix, detectable_bins=None, max_dist=None, smooth=False, fun=np.nanmean, max_val=10, ): """ Detrends a Hi-C matrix by the distance law. The input matrix should have been normalised beforehandand. Parameters ---------- matrix : scipy.sparse.csr_matrix The normalised intrachromosomal Hi-C matrix to detrend. detectable_bins : tuple Tuple containing a list of detectable rows and a list of columns on which to perform detrending. Poorly interacting indices have been excluded. max_dist : int Maximum number of bins from the diagonal at which to compute trend. smooth : bool Whether to use isotonic regression to smooth the trend. fun : callable Function to use on each diagonal to compute the trend. max_val : float or None Maximum value in the detrended matrix. Set to None to disable Returns ------- numpy.ndarray : The detrended Hi-C matrix. """ matrix = matrix.tocsr() y = distance_law( matrix, detectable_bins=detectable_bins, max_dist=max_dist, smooth=smooth, fun=fun, ) y[np.isnan(y)] = 0.0 # Detrending by the distance law clean_mat = matrix.tocoo() # clean_mat.data /= y_savgol[abs(clean_mat.row - clean_mat.col)] try: clean_mat.data = clean_mat.data / y[abs(clean_mat.row - clean_mat.col)] # If no nonzero value in matrix, do nothing except TypeError: pass clean_mat = clean_mat.tocsr() if max_val is not None: clean_mat[clean_mat >= max_val] = 1 return clean_mat def ztransform(matrix): """ Z transformation for Hi-C matrices. Parameters ---------- matrix : scipy.sparse.coo_matrix A Hi-C matrix in sparse format.
Returns ------- scipy.sparse.coo_matrix: The detrended Hi-C matrix """ mat = matrix.copy() mu = np.mean(mat.data) sd = np.std(mat.data) mat.data -= mu mat.data /= sd return mat def sum_mat_bins(mat): """ Compute the sum of matrices bins (i.e. rows or columns) using only the upper triangle, assuming symmetrical matrices. Parameters ---------- mat : scipy.sparse.coo_matrix Contact map in sparse format, either in upper triangle or full matrix. Returns ------- numpy.ndarray : 1D array of bin sums. """ # Equivalaent to row or col sum on a full matrix # Note: mat.sum returns a 'matrix' object. A1 extracts the 1D flat array # from the matrix return mat.sum(axis=0).A1 + mat.sum(axis=1).A1 - mat.diagonal(0) def subsample_contacts(M, n_contacts): """Bootstrap sampling of contacts in a sparse Hi-C map. Parameters ---------- M : scipy.sparse.coo_matrix The input Hi-C contact map in sparse format. n_contacts : int The number of contacts to sample. Returns ------- scipy.sparse.coo_matrix A new matrix with a fraction of the original contacts. """ S = M.data.copy() # Match cell idx to cumulative number of contacts cum_counts = np.cumsum(S) # Total number of contacts to sample tot_contacts = int(cum_counts[-1]) # Sample desired number of contacts from the range(0, n_contacts) array sampled_contacts = np.random.choice( int(tot_contacts), size=(n_contacts), replace=False ) # Get indices of sampled contacts in the cum_counts array idx = np.searchsorted(cum_counts, sampled_contacts, side="right") # Bin those indices to the same dimensions as matrix data to get counts sampled_counts = np.bincount(idx, minlength=S.shape[0]) # Get nonzero values to build new sparse matrix nnz_mask = sampled_counts > 0 sampled_counts = sampled_counts[nnz_mask].astype(np.float64) sampled_rows = M.row[nnz_mask] sampled_cols = M.col[nnz_mask] return sp.coo_matrix( (sampled_counts, (sampled_rows, sampled_cols)), shape=(M.shape[0], M.shape[1]), ) def frame_missing_mask(mask, kernel_shape, sym_upper=False, max_dist=None): """ Adds a frame around input mask, given a kernel. The goal of this frame is define margins around the matrix where the kernel will not perform convolution (denoted by 1). If the matrix is upper symmetric, a margin of half the kernel's width is added below the diagonal and a maximum distance from the diagonal above which margins need not be drawn can be considered. Otherwise Margins are simply added on all 4 sides of the matrix. :: signal kernel _________ ______ ____ |#######| | | | | ==> |# #| | | |___| |# #| | | |# #| |_____| |# #| |#######| -------- Parameters ---------- mask : scipy.sparse.csr_matrix of bool The mask around which to add margins. kernels_shape : tuple of ints The number of rows and kernel in the input kernel. Margins will be half these values. sym_upper : bool Whether the signal is a symmetric upper triangle matrix. If so, values on a margin below the diagonal will be masked. max_dist : int or None Number of diagonals to keep Returns ------- framed_mask : scipy.sparse.csr_matrix of bool The input mask with a padding of True around the edges. If sym_upper is True, a padding is also added below the diagonal. """ if mask.dtype != bool: raise ValueError("Mask must contain boolean values") if not sp.issparse(mask): raise ValueError("Mask must be a sparse matrix") framed_mask = mask.copy() ms, ns = mask.shape mk, nk = kernel_shape if sym_upper and (max_dist is not None): # Remove diagonals further than scan distance in the input mask framed_mask = diag_trim(framed_mask, max_dist + max(nk
random_line_split
preprocessing.py
raise ValueError("input type must be scipy.sparse.csr_matrix") # Trim diagonals by removing all elements further than n in the # upper triangle trimmed = sp.tril(mat, n, format="csr") trimmed = sp.triu(trimmed, format="csr") else: trimmed = mat.copy() n_diags = trimmed.shape[0] for diag in range(n, n_diags): set_mat_diag(trimmed, diag, 0) return trimmed return trimmed def distance_law( matrix, detectable_bins=None, max_dist=None, smooth=True, fun=np.nanmean ): """ Computes genomic distance law by averaging over each diagonal in the upper triangle matrix. If a list of detectable bins is provided, pixels in missing bins will be excluded from the averages. A maximum distance can be specified to define how many diagonals should be computed. parameters ---------- matrix: scipy.sparse.csr_matrix the input matrix to compute distance law from. detectable_bins : numpy.ndarray of ints An array of detectable bins indices to consider when computing distance law. max_dist : int Maximum distance from diagonal, in number of bins in which to compute distance law smooth : bool Whether to use isotonic regression to smooth the distance law. fun : callable A function to apply on each diagonal. Defaults to mean. Returns ------- dist: np.ndarray the output genomic distance law. example ------- >>> m = np.ones((3,3)) >>> m += np.array([1,2,3]) >>> m array([[2., 3., 4.], [2., 3., 4.], [2., 3., 4.]]) >>> distance_law(csr_matrix(m)) array([3. , 3.5, 4. ]) """ mat_n = matrix.shape[0] if max_dist is None: max_dist = mat_n n_diags = min(mat_n, max_dist + 1) dist = np.zeros(mat_n) if detectable_bins is None: detectable_bins = np.array(range(mat_n)) for diag in range(n_diags): # Find detectable which fall in diagonal detect_mask = np.zeros(mat_n, dtype=bool) detect_mask[detectable_bins] = 1 # Find bins which are detectable in the diagonal (intersect of # hori and verti) detect_mask_h = detect_mask[: (mat_n - diag)] detect_mask_v = detect_mask[mat_n - (mat_n - diag) :] detect_mask_diag = detect_mask_h & detect_mask_v detect_diag = matrix.diagonal(diag)[detect_mask_diag] dist[diag] = fun(detect_diag[detect_diag > 0]) # Smooth the curve using isotonic regression: Find closest approximation # with the condition that point n+1 cannot be higher than point n. # (i.e. contacts can only decrease when increasing distance) if smooth and mat_n > 2: ir = IsotonicRegression(increasing=False) dist[~np.isfinite(dist)] = 0 dist = ir.fit_transform(range(len(dist)), dist) return dist def get_detectable_bins(mat, n_mads=3, inter=False): """ Returns lists of detectable indices after excluding low interacting bins based on the proportion of zero pixel values in the matrix bins. Parameters ---------- mat : scipy.sparse.coo_matrix A Hi-C matrix in tihe form of a 2D numpy array or coo matrix n_mads : int Number of median absolute deviation below the median required to consider bins non-detectable. inter : bool Whether the matrix is interchromosomal. Default is to consider the matrix is intrachromosomal (i.e. upper symmetric). Returns ------- numpy ndarray : tuple of 2 1D arrays containing indices of low interacting rows and columns, respectively. """ matrix = mat.copy() matrix.eliminate_zeros() def mad(x): return ss.median_abs_deviation(x, nan_policy="omit") if not inter: if matrix.shape[0] != matrix.shape[1]: raise ValueError("Intrachromosomal matrices must be symmetric.") # Replace nonzero pixels by ones to work on prop. of nonzero pixels matrix.data = np.ones(matrix.data.shape) # Compute number of nonzero values in each bin sum_bins = sum_mat_bins(matrix) # Compute variation in the number of nonzero pixels sum_mad = mad(sum_bins) # Find poor interacting rows and columns sum_med = np.median(sum_bins) detect_threshold = max(1, sum_med - sum_mad * n_mads) # Removal of poor interacting rows and columns good_bins = np.flatnonzero(sum_bins >= detect_threshold) good_bins = (good_bins, good_bins) else: # Adapted for asymetric matrices (need to compute rows and columns) sum_rows, sum_cols = matrix.sum(axis=1).A1, matrix.sum(axis=0).A1 mad_rows, mad_cols = mad(sum_rows), mad(sum_cols) med_rows, med_cols = np.median(sum_rows), np.median(sum_cols) detect_threshold_rows = max(1, med_rows - mad_rows * n_mads) detect_threshold_cols = max(1, med_cols - mad_cols * n_mads) good_rows = np.flatnonzero(sum_rows > detect_threshold_rows) good_cols = np.flatnonzero(sum_cols > detect_threshold_cols) good_bins = (good_rows, good_cols) return good_bins def detrend( matrix, detectable_bins=None, max_dist=None, smooth=False, fun=np.nanmean, max_val=10, ): """ Detrends a Hi-C matrix by the distance law. The input matrix should have been normalised beforehandand. Parameters ---------- matrix : scipy.sparse.csr_matrix The normalised intrachromosomal Hi-C matrix to detrend. detectable_bins : tuple Tuple containing a list of detectable rows and a list of columns on which to perform detrending. Poorly interacting indices have been excluded. max_dist : int Maximum number of bins from the diagonal at which to compute trend. smooth : bool Whether to use isotonic regression to smooth the trend. fun : callable Function to use on each diagonal to compute the trend. max_val : float or None Maximum value in the detrended matrix. Set to None to disable Returns ------- numpy.ndarray : The detrended Hi-C matrix. """ matrix = matrix.tocsr() y = distance_law( matrix, detectable_bins=detectable_bins, max_dist=max_dist, smooth=smooth, fun=fun, ) y[np.isnan(y)] = 0.0 # Detrending by the distance law clean_mat = matrix.tocoo() # clean_mat.data /= y_savgol[abs(clean_mat.row - clean_mat.col)] try: clean_mat.data = clean_mat.data / y[abs(clean_mat.row - clean_mat.col)] # If no nonzero value in matrix, do nothing except TypeError: pass clean_mat = clean_mat.tocsr() if max_val is not None: clean_mat[clean_mat >= max_val] = 1 return clean_mat def ztransform(matrix): """ Z transformation for Hi-C matrices. Parameters ---------- matrix : scipy.sparse.coo_matrix A Hi-C matrix in sparse format. Returns ------- scipy.sparse.coo_matrix: The detrended Hi-C matrix """ mat = matrix.copy() mu = np.mean(mat.data) sd = np.std(mat.data) mat.data -= mu mat.data /= sd return mat def sum_mat_bins(mat): """ Compute the sum of matrices bins (i.e. rows or columns) using only the upper triangle, assuming symmetrical matrices. Parameters ---------- mat : scipy.sparse.coo_matrix Contact map in sparse format, either in upper triangle or full matrix. Returns ------- numpy.ndarray : 1D array of bin sums. """ # Equivalaent to row or col sum on a full matrix # Note: mat.sum returns a 'matrix' object. A1 extracts the 1D flat array # from the matrix return mat.sum(axis=0).A1 + mat.sum(axis=1).A1 - mat.diagonal(0) def subsample_contacts(M, n_contacts): """Bootstrap
""" Trim an upper triangle sparse matrix so that only the first n diagonals are kept. Parameters ---------- mat : scipy.sparse.csr_matrix or numpy.ndarray The sparse matrix to be trimmed n : int The number of diagonals from the center to keep (0-based). Returns ------- scipy.sparse.dia_matrix or numpy.ndarray: The diagonally trimmed upper triangle matrix with only the first n diagonal. """ if sp.issparse(mat): if mat.format != "csr":
identifier_body
graph.service.ts
key.SankeyNode<SNodeExtra, SLinkExtra>; export type SLink = d3Sankey.SankeyLink<SNodeExtra, SLinkExtra>; export interface DAG { nodes: SNode[]; links: SLink[]; } // -- Dag Node -- @Injectable({ providedIn: 'root' }) export class GraphService { private docGuids = {}; private docDb: Db = null; private docs = {}; private nextDocGuid = 0; private filterOrder = []; private runningFilters = false; public updateSubject = new Rx.BehaviorSubject(0); public updateViewSubject = new Rx.BehaviorSubject(0); public visualStyle = true; public visualZoom = 1; constructor( private messageService: MessageService, private http: HttpClient) { this.addTab("ISO"); } public runFilters(changedTab: GraphTab, parentChanged: boolean) { if (this.runningFilters) return; // prevent re-entry this.runningFilters = true; var tabs = this.graphTabs; var anyChanged = false; for (var i of this.filterOrder) {
if (anyChanged || t.column.autoFilterSrc == changedTab.column || (parentChanged && t == changedTab)) { anyChanged = true; // filter child tree GraphFilter.runFilter(t.column); } } //if (anyChanged) this.updateSubject.next(0); this.runningFilters = false; } getGuid(id: string, type: string, rev: string, createMissing: boolean = true): number { var key = `${type}-${id}-${rev}`; if (key in this.docGuids) return this.docGuids[key]; if (!createMissing) return null; var value = this.nextDocGuid++; this.docGuids[key] = value; return value; } getDbIndex() : Observable<Db> { if (this.docDb) return of(this.docDb); return this.http.get<Db>('assets/output/docs-index.json', {responseType: 'json'}) .pipe( tap( data => { this.docDb = data; }, error => this.handleError("getDbIndex", []) ) ); } getDoc(id: string) : Observable<Doc2> { if (this.docs[id]) return of(this.docs[id]); return this.http.get<Doc2>('assets/output/docs-' + id + '.json', {responseType: 'json'}) .pipe( tap( data => { this.docs[id] = data; }, error => this.handleError("getDoc", []) ) ); } getDocTypes() : Observable<CategoryList> { return this.getDbIndex().pipe( map( data => { return data.docs.map(v => { return { id: v.id, title: v.type }; }); } ) ); } private addToDoc(parent: FullDocNode, input: DocNode2) { var child = new FullDocNode(input); if (parent) { parent.children.push(child); } // Recurse for (var c of input.children) { this.addToDoc(child, c); } return child; } getFullDocByType(id: string) : Observable<FullDocNode> { return this.getDoc(id).pipe( map( data => { return this.addToDoc(null, data); } ) ); } getChangeLog(): Observable<Change[]> { return this.getDbIndex().pipe( map( data => { return data.changelog; } ) ); } // Live state management: maybe move this to a different service. public graphTabs: GraphTab[] = [ ]; public selectedTab: number = 0; public get canAdd(): boolean { return this.graphTabs.length < 3; } public addTab(id: string, filterTopLevelKeys: string[] = null, customName: string = null) { if (!this.canAdd) return; this.getFullDocByType(id) .subscribe(doc => { var newTab = new GraphTab(this, null, doc); newTab.nodes = doc.children; // Filter the top level nodes if desired if (filterTopLevelKeys) { newTab.nodes = doc.children.filter(c => { var rootKey = c.id.replace(/_/g, ''); // The 'all' structure has _, where the menu choice has those stripped. Remove them for comparison. return filterTopLevelKeys.includes(rootKey); }); } // Override the tab name if desired if (customName) newTab.title = customName; // Reference this data from the column view. newTab.column.nodes = newTab.nodes; this.graphTabs.push(newTab); this.ensureISOIsInMiddle(); // Coverage calculation is disabled to save time. //if (id != "ISO") //{ // // compare with iso. // newTab.coverage = this.compareDocs(newTab.column, this.graphTabs[1]); //} var selectTab = newTab; if (newTab.isAll) { selectTab = this.graphTabs.find(t => t.isIso); } // The current request is to NOT activate the newly added tab. So only activate index 0 if (this.graphTabs.length != 1) { selectTab = this.graphTabs[this.selectedTab]; // reselect current selection } // even if we dont change tabs, we still have to reactive it to configure filters this.activateTab(selectTab); }); } private ensureISOIsInMiddle() { var isoTab = this.graphTabs.find(t => t.isIso); if (this.graphTabs.length > 1) { this.graphTabs = this.graphTabs.filter(t => t != isoTab); this.graphTabs.splice(1, 0, isoTab); } } public configureFilterStack() { switch (this.selectedTab) { case 0: this.filterOrder = [0, 1, 2]; break; case 1: this.filterOrder = [1, 0, 2]; break; case 2: this.filterOrder = [2, 1, 0]; break; } // setup filters var isoTab = this.graphTabs.find(t => t.isIso); var primary = this.graphTabs[this.filterOrder[0]]; if (!primary) return; // clear auto filter of left tab primary.column.autoFilterSrc = null; primary.column.autoFilterParent = null; primary.column.autoFilterSelf = false; var secondary = this.graphTabs[this.filterOrder[1]]; if (secondary) { if (secondary == isoTab) { // assure iso filters from the primary: "auto filter" isoTab.column.autoFilterSrc = primary.column; isoTab.column.autoFilterParent = primary.column.parent; isoTab.column.autoFilterSelf = false; } else { // auto filter with this tabs connections to iso secondary.column.autoFilterSrc = isoTab.column; secondary.column.autoFilterParent = primary.column.parent; // the primary tab always drives the selection secondary.column.autoFilterSelf = true; } } var third = this.graphTabs[this.filterOrder[2]]; if (third) { // auto filter with this tabs connections to iso third.column.autoFilterSrc = isoTab.column; third.column.autoFilterParent = primary.column.parent; // the primary tab always drives the selection third.column.autoFilterSelf = true; } } public tabChanged() { this.configureFilterStack(); if (this.selectedTab >= 0 && this.selectedTab < this.graphTabs.length) { this.graphTabs[this.selectedTab].parentTabTreeChanged(); } } public removeTab(tab) { this.graphTabs = this.graphTabs.filter(t => t!=tab); this.ensureISOIsInMiddle(); this.activateTab(this.graphTabs[0]); } public activateTab(tab: GraphTab): Promise<boolean> { return new Promise<boolean>((resolve, reject) => { var newIndex = this.graphTabs.indexOf(tab); var finalize = () => { this.selectedTab = newIndex; this.tabChanged(); setTimeout(() => { resolve(true); }, 1000); }; // if the index is the same this.selectedTab = -1; // set it to non-value so change is detected setTimeout(finalize, 1000); // give dom time to stabilize }); } public getNodesWithLinks(children: FullDocNode[], result: FullDocNode[]) { for (var c of children) { if (c.node.links && c.node.links.length > 0) result.push(c); this.getNodesWithLinks(c.children, result); } return result; } public flattenSections(children: FullDocNode[], result: string[]) { for (var c of children) { if (c.getBody()) result.push(c.id); this.flattenSections(c.children, result); } return result; } public flattenLinks(children: FullDocNode[], result: Link
var t = tabs[i]; if (!t) continue;
random_line_split
graph.service.ts
key.SankeyNode<SNodeExtra, SLinkExtra>; export type SLink = d3Sankey.SankeyLink<SNodeExtra, SLinkExtra>; export interface DAG { nodes: SNode[]; links: SLink[]; } // -- Dag Node -- @Injectable({ providedIn: 'root' }) export class GraphService { private docGuids = {}; private docDb: Db = null; private docs = {}; private nextDocGuid = 0; private filterOrder = []; private runningFilters = false; public updateSubject = new Rx.BehaviorSubject(0); public updateViewSubject = new Rx.BehaviorSubject(0); public visualStyle = true; public visualZoom = 1;
( private messageService: MessageService, private http: HttpClient) { this.addTab("ISO"); } public runFilters(changedTab: GraphTab, parentChanged: boolean) { if (this.runningFilters) return; // prevent re-entry this.runningFilters = true; var tabs = this.graphTabs; var anyChanged = false; for (var i of this.filterOrder) { var t = tabs[i]; if (!t) continue; if (anyChanged || t.column.autoFilterSrc == changedTab.column || (parentChanged && t == changedTab)) { anyChanged = true; // filter child tree GraphFilter.runFilter(t.column); } } //if (anyChanged) this.updateSubject.next(0); this.runningFilters = false; } getGuid(id: string, type: string, rev: string, createMissing: boolean = true): number { var key = `${type}-${id}-${rev}`; if (key in this.docGuids) return this.docGuids[key]; if (!createMissing) return null; var value = this.nextDocGuid++; this.docGuids[key] = value; return value; } getDbIndex() : Observable<Db> { if (this.docDb) return of(this.docDb); return this.http.get<Db>('assets/output/docs-index.json', {responseType: 'json'}) .pipe( tap( data => { this.docDb = data; }, error => this.handleError("getDbIndex", []) ) ); } getDoc(id: string) : Observable<Doc2> { if (this.docs[id]) return of(this.docs[id]); return this.http.get<Doc2>('assets/output/docs-' + id + '.json', {responseType: 'json'}) .pipe( tap( data => { this.docs[id] = data; }, error => this.handleError("getDoc", []) ) ); } getDocTypes() : Observable<CategoryList> { return this.getDbIndex().pipe( map( data => { return data.docs.map(v => { return { id: v.id, title: v.type }; }); } ) ); } private addToDoc(parent: FullDocNode, input: DocNode2) { var child = new FullDocNode(input); if (parent) { parent.children.push(child); } // Recurse for (var c of input.children) { this.addToDoc(child, c); } return child; } getFullDocByType(id: string) : Observable<FullDocNode> { return this.getDoc(id).pipe( map( data => { return this.addToDoc(null, data); } ) ); } getChangeLog(): Observable<Change[]> { return this.getDbIndex().pipe( map( data => { return data.changelog; } ) ); } // Live state management: maybe move this to a different service. public graphTabs: GraphTab[] = [ ]; public selectedTab: number = 0; public get canAdd(): boolean { return this.graphTabs.length < 3; } public addTab(id: string, filterTopLevelKeys: string[] = null, customName: string = null) { if (!this.canAdd) return; this.getFullDocByType(id) .subscribe(doc => { var newTab = new GraphTab(this, null, doc); newTab.nodes = doc.children; // Filter the top level nodes if desired if (filterTopLevelKeys) { newTab.nodes = doc.children.filter(c => { var rootKey = c.id.replace(/_/g, ''); // The 'all' structure has _, where the menu choice has those stripped. Remove them for comparison. return filterTopLevelKeys.includes(rootKey); }); } // Override the tab name if desired if (customName) newTab.title = customName; // Reference this data from the column view. newTab.column.nodes = newTab.nodes; this.graphTabs.push(newTab); this.ensureISOIsInMiddle(); // Coverage calculation is disabled to save time. //if (id != "ISO") //{ // // compare with iso. // newTab.coverage = this.compareDocs(newTab.column, this.graphTabs[1]); //} var selectTab = newTab; if (newTab.isAll) { selectTab = this.graphTabs.find(t => t.isIso); } // The current request is to NOT activate the newly added tab. So only activate index 0 if (this.graphTabs.length != 1) { selectTab = this.graphTabs[this.selectedTab]; // reselect current selection } // even if we dont change tabs, we still have to reactive it to configure filters this.activateTab(selectTab); }); } private ensureISOIsInMiddle() { var isoTab = this.graphTabs.find(t => t.isIso); if (this.graphTabs.length > 1) { this.graphTabs = this.graphTabs.filter(t => t != isoTab); this.graphTabs.splice(1, 0, isoTab); } } public configureFilterStack() { switch (this.selectedTab) { case 0: this.filterOrder = [0, 1, 2]; break; case 1: this.filterOrder = [1, 0, 2]; break; case 2: this.filterOrder = [2, 1, 0]; break; } // setup filters var isoTab = this.graphTabs.find(t => t.isIso); var primary = this.graphTabs[this.filterOrder[0]]; if (!primary) return; // clear auto filter of left tab primary.column.autoFilterSrc = null; primary.column.autoFilterParent = null; primary.column.autoFilterSelf = false; var secondary = this.graphTabs[this.filterOrder[1]]; if (secondary) { if (secondary == isoTab) { // assure iso filters from the primary: "auto filter" isoTab.column.autoFilterSrc = primary.column; isoTab.column.autoFilterParent = primary.column.parent; isoTab.column.autoFilterSelf = false; } else { // auto filter with this tabs connections to iso secondary.column.autoFilterSrc = isoTab.column; secondary.column.autoFilterParent = primary.column.parent; // the primary tab always drives the selection secondary.column.autoFilterSelf = true; } } var third = this.graphTabs[this.filterOrder[2]]; if (third) { // auto filter with this tabs connections to iso third.column.autoFilterSrc = isoTab.column; third.column.autoFilterParent = primary.column.parent; // the primary tab always drives the selection third.column.autoFilterSelf = true; } } public tabChanged() { this.configureFilterStack(); if (this.selectedTab >= 0 && this.selectedTab < this.graphTabs.length) { this.graphTabs[this.selectedTab].parentTabTreeChanged(); } } public removeTab(tab) { this.graphTabs = this.graphTabs.filter(t => t!=tab); this.ensureISOIsInMiddle(); this.activateTab(this.graphTabs[0]); } public activateTab(tab: GraphTab): Promise<boolean> { return new Promise<boolean>((resolve, reject) => { var newIndex = this.graphTabs.indexOf(tab); var finalize = () => { this.selectedTab = newIndex; this.tabChanged(); setTimeout(() => { resolve(true); }, 1000); }; // if the index is the same this.selectedTab = -1; // set it to non-value so change is detected setTimeout(finalize, 1000); // give dom time to stabilize }); } public getNodesWithLinks(children: FullDocNode[], result: FullDocNode[]) { for (var c of children) { if (c.node.links && c.node.links.length > 0) result.push(c); this.getNodesWithLinks(c.children, result); } return result; } public flattenSections(children: FullDocNode[], result: string[]) { for (var c of children) { if (c.getBody()) result.push(c.id); this.flattenSections(c.children, result); } return result; } public flattenLinks(children: FullDocNode[], result: Link
constructor
identifier_name
graph.service.ts
, title: v.type }; }); } ) ); } private addToDoc(parent: FullDocNode, input: DocNode2) { var child = new FullDocNode(input); if (parent) { parent.children.push(child); } // Recurse for (var c of input.children) { this.addToDoc(child, c); } return child; } getFullDocByType(id: string) : Observable<FullDocNode> { return this.getDoc(id).pipe( map( data => { return this.addToDoc(null, data); } ) ); } getChangeLog(): Observable<Change[]> { return this.getDbIndex().pipe( map( data => { return data.changelog; } ) ); } // Live state management: maybe move this to a different service. public graphTabs: GraphTab[] = [ ]; public selectedTab: number = 0; public get canAdd(): boolean { return this.graphTabs.length < 3; } public addTab(id: string, filterTopLevelKeys: string[] = null, customName: string = null) { if (!this.canAdd) return; this.getFullDocByType(id) .subscribe(doc => { var newTab = new GraphTab(this, null, doc); newTab.nodes = doc.children; // Filter the top level nodes if desired if (filterTopLevelKeys) { newTab.nodes = doc.children.filter(c => { var rootKey = c.id.replace(/_/g, ''); // The 'all' structure has _, where the menu choice has those stripped. Remove them for comparison. return filterTopLevelKeys.includes(rootKey); }); } // Override the tab name if desired if (customName) newTab.title = customName; // Reference this data from the column view. newTab.column.nodes = newTab.nodes; this.graphTabs.push(newTab); this.ensureISOIsInMiddle(); // Coverage calculation is disabled to save time. //if (id != "ISO") //{ // // compare with iso. // newTab.coverage = this.compareDocs(newTab.column, this.graphTabs[1]); //} var selectTab = newTab; if (newTab.isAll) { selectTab = this.graphTabs.find(t => t.isIso); } // The current request is to NOT activate the newly added tab. So only activate index 0 if (this.graphTabs.length != 1) { selectTab = this.graphTabs[this.selectedTab]; // reselect current selection } // even if we dont change tabs, we still have to reactive it to configure filters this.activateTab(selectTab); }); } private ensureISOIsInMiddle() { var isoTab = this.graphTabs.find(t => t.isIso); if (this.graphTabs.length > 1) { this.graphTabs = this.graphTabs.filter(t => t != isoTab); this.graphTabs.splice(1, 0, isoTab); } } public configureFilterStack() { switch (this.selectedTab) { case 0: this.filterOrder = [0, 1, 2]; break; case 1: this.filterOrder = [1, 0, 2]; break; case 2: this.filterOrder = [2, 1, 0]; break; } // setup filters var isoTab = this.graphTabs.find(t => t.isIso); var primary = this.graphTabs[this.filterOrder[0]]; if (!primary) return; // clear auto filter of left tab primary.column.autoFilterSrc = null; primary.column.autoFilterParent = null; primary.column.autoFilterSelf = false; var secondary = this.graphTabs[this.filterOrder[1]]; if (secondary) { if (secondary == isoTab) { // assure iso filters from the primary: "auto filter" isoTab.column.autoFilterSrc = primary.column; isoTab.column.autoFilterParent = primary.column.parent; isoTab.column.autoFilterSelf = false; } else { // auto filter with this tabs connections to iso secondary.column.autoFilterSrc = isoTab.column; secondary.column.autoFilterParent = primary.column.parent; // the primary tab always drives the selection secondary.column.autoFilterSelf = true; } } var third = this.graphTabs[this.filterOrder[2]]; if (third) { // auto filter with this tabs connections to iso third.column.autoFilterSrc = isoTab.column; third.column.autoFilterParent = primary.column.parent; // the primary tab always drives the selection third.column.autoFilterSelf = true; } } public tabChanged() { this.configureFilterStack(); if (this.selectedTab >= 0 && this.selectedTab < this.graphTabs.length) { this.graphTabs[this.selectedTab].parentTabTreeChanged(); } } public removeTab(tab) { this.graphTabs = this.graphTabs.filter(t => t!=tab); this.ensureISOIsInMiddle(); this.activateTab(this.graphTabs[0]); } public activateTab(tab: GraphTab): Promise<boolean> { return new Promise<boolean>((resolve, reject) => { var newIndex = this.graphTabs.indexOf(tab); var finalize = () => { this.selectedTab = newIndex; this.tabChanged(); setTimeout(() => { resolve(true); }, 1000); }; // if the index is the same this.selectedTab = -1; // set it to non-value so change is detected setTimeout(finalize, 1000); // give dom time to stabilize }); } public getNodesWithLinks(children: FullDocNode[], result: FullDocNode[]) { for (var c of children) { if (c.node.links && c.node.links.length > 0) result.push(c); this.getNodesWithLinks(c.children, result); } return result; } public flattenSections(children: FullDocNode[], result: string[]) { for (var c of children) { if (c.getBody()) result.push(c.id); this.flattenSections(c.children, result); } return result; } public flattenLinks(children: FullDocNode[], result: Link[], linkData: any) { for (var c of children) { if (c.shouldBeMapped) { linkData.total++; if (!c.isUnmapped) { linkData.linked++; result = result.concat(c.node.links); } } result = this.flattenLinks(c.children, result, linkData); } return result; } public compareDocs(aTab: GraphTab, bTab: GraphTab): any { var bSections = []; this.flattenSections(bTab.nodes, bSections); var bCopy = bSections.slice(); var linkData = { total: 0, linked: 0 }; var aLinks = this.flattenLinks(aTab.nodes, [], linkData); var found = 0; var checked = 0; for (var a of aLinks) { ++checked; var b = bCopy.find(x => x == a.id) if (b) { bCopy = bCopy.filter(x => x != b); ++found; } } return { coverage: found + "/" + bSections.length, mapped: linkData.linked + "/" + linkData.total, uniqueconnections: found + "/" + checked, uncoveredIds: bCopy //"coverage": (found / bSections.length * 100).toFixed(1) + "% (" + found + "/" + bSections.length + ")", //"mapped": (linkData.linked / linkData.total * 100).toFixed(1) + "% (" + linkData.linked + "/" + linkData.total + ")", //"uniqueconnections": (found / checked * 100).toFixed(1) + "% (" + found + "/" + checked + ")" }; } /** * Handle Http operation that failed. * Let the app continue. * @param operation - name of the operation that failed * @param result - optional value to return as the observable result */ private handleError<T> (operation = 'operation', result?: T) { return (error: any): Observable<T> => { // TODO: send the error to remote logging infrastructure console.error(error); // log to console instead // TODO: better job of transforming error for user consumption this.log(`${operation} failed: ${error.message}`); // Let the app keep running by returning an empty result. return of(result as T); }; } /** Log a GraphService message with the MessageService */ private log(message: string) { this.messageService.add(`GraphService: ${message}`); } public get errorStrings(): string[] { return this.graphTabs.reduce((a: string[], v: GraphTab) => { a.concat(v.errors()); return a; }, []); } public get anyErrors(): boolean
{ for (var t of this.graphTabs) if (t.anyErrors) return true; return false; }
identifier_body
binomial.rs
algorithm based on the inverse // transformation of the binomial distribution is efficient. Otherwise, // the BTPE algorithm is used. // // Voratas Kachitvichyanukul and Bruce W. Schmeiser. 1988. Binomial // random variate generation. Commun. ACM 31, 2 (February 1988), // 216-222. http://dx.doi.org/10.1145/42372.42381 // Threshold for prefering the BINV algorithm. The paper suggests 10, // Ranlib uses 30, and GSL uses 14. const BINV_THRESHOLD: f64 = 10.; if (self.n as f64) * p < BINV_THRESHOLD && self.n <= (::std::i32::MAX as u64) { // Use the BINV algorithm. let s = p / q; let a = ((self.n + 1) as f64) * s; let mut r = q.powi(self.n as i32); let mut u: f64 = rng.gen(); let mut x = 0; while u > r as f64 { u -= r; x += 1; r *= a / (x as f64) - s; } result = x; } else { // Use the BTPE algorithm. // Threshold for using the squeeze algorithm. This can be freely // chosen based on performance. Ranlib and GSL use 20. const SQUEEZE_THRESHOLD: i64 = 20; // Step 0: Calculate constants as functions of `n` and `p`. let n = self.n as f64; let np = n * p; let npq = np * q; let f_m = np + p; let m = f64_to_i64(f_m); // radius of triangle region, since height=1 also area of region let p1 = (2.195 * npq.sqrt() - 4.6 * q).floor() + 0.5; // tip of triangle let x_m = (m as f64) + 0.5; // left edge of triangle let x_l = x_m - p1; // right edge of triangle let x_r = x_m + p1; let c = 0.134 + 20.5 / (15.3 + (m as f64)); // p1 + area of parallelogram region let p2 = p1 * (1. + 2. * c); fn lambda(a: f64) -> f64 { a * (1. + 0.5 * a) } let lambda_l = lambda((f_m - x_l) / (f_m - x_l * p)); let lambda_r = lambda((x_r - f_m) / (x_r * q)); // p1 + area of left tail let p3 = p2 + c / lambda_l; // p1 + area of right tail let p4 = p3 + c / lambda_r; // return value let mut y: i64; let gen_u = Uniform::new(0., p4); let gen_v = Uniform::new(0., 1.); loop { // Step 1: Generate `u` for selecting the region. If region 1 is // selected, generate a triangularly distributed variate. let u = gen_u.sample(rng); let mut v = gen_v.sample(rng); if !(u > p1) { y = f64_to_i64(x_m - p1 * v + u); break; } if !(u > p2) { // Step 2: Region 2, parallelograms. Check if region 2 is // used. If so, generate `y`. let x = x_l + (u - p1) / c; v = v * c + 1.0 - (x - x_m).abs() / p1; if v > 1. { continue; } else { y = f64_to_i64(x); } } else if !(u > p3) { // Step 3: Region 3, left exponential tail. y = f64_to_i64(x_l + v.ln() / lambda_l); if y < 0 { continue; } else { v *= (u - p2) * lambda_l; } } else { // Step 4: Region 4, right exponential tail. y = f64_to_i64(x_r - v.ln() / lambda_r); if y > 0 && (y as u64) > self.n { continue; } else { v *= (u - p3) * lambda_r; } } // Step 5: Acceptance/rejection comparison. // Step 5.0: Test for appropriate method of evaluating f(y). let k = (y - m).abs(); if !(k > SQUEEZE_THRESHOLD && (k as f64) < 0.5 * npq - 1.) { // Step 5.1: Evaluate f(y) via the recursive relationship. Start the // search from the mode. let s = p / q; let a = s * (n + 1.); let mut f = 1.0; if m < y { let mut i = m; loop { i += 1; f *= a / (i as f64) - s; if i == y { break; } } } else if m > y { let mut i = y; loop { i += 1; f /= a / (i as f64) - s; if i == m { break; } } } if v > f { continue; } else { break; } } // Step 5.2: Squeezing. Check the value of ln(v) againts upper and // lower bound of ln(f(y)). let k = k as f64; let rho = (k / npq) * ((k * (k / 3. + 0.625) + 1. / 6.) / npq + 0.5); let t = -0.5 * k * k / npq; let alpha = v.ln(); if alpha < t - rho { break; } if alpha > t + rho { continue; } // Step 5.3: Final acceptance/rejection test. let x1 = (y + 1) as f64; let f1 = (m + 1) as f64; let z = (f64_to_i64(n) + 1 - m) as f64; let w = (f64_to_i64(n) - y + 1) as f64; fn stirling(a: f64) -> f64 { let a2 = a * a; (13860. - (462. - (132. - (99. - 140. / a2) / a2) / a2) / a2) / a / 166320. } if alpha > x_m * (f1 / x1).ln() + (n - (m as f64) + 0.5) * (z / w).ln() + ((y - m) as f64) * (w * p / (x1 * q)).ln() // We use the signs from the GSL implementation, which are // different than the ones in the reference. According to // the GSL authors, the new signs were verified to be // correct by one of the original designers of the // algorithm. + stirling(f1) + stirling(z) - stirling(x1) - stirling(w) { continue; } break; } assert!(y >= 0); result = y as u64; } // Invert the result for p < 0.5. if p != self.p { self.n - result } else { result } } } #[cfg(test)] mod test { use super::Binomial; use crate::distributions::Distribution; use crate::Rng; fn test_binomial_mean_and_variance<R: Rng>(n: u64, p: f64, rng: &mut R) { let binomial = Binomial::new(n, p); let expected_mean = n as f64 * p; let expected_variance = n as f64 * p * (1.0 - p);
let mut results = [0.0; 1000];
random_line_split
binomial.rs
; } else if self.p == 1.0 { return self.n; } // The binomial distribution is symmetrical with respect to p -> 1-p, // k -> n-k switch p so that it is less than 0.5 - this allows for lower // expected values we will just invert the result at the end let p = if self.p <= 0.5 { self.p } else { 1.0 - self.p }; let result; let q = 1. - p; // For small n * min(p, 1 - p), the BINV algorithm based on the inverse // transformation of the binomial distribution is efficient. Otherwise, // the BTPE algorithm is used. // // Voratas Kachitvichyanukul and Bruce W. Schmeiser. 1988. Binomial // random variate generation. Commun. ACM 31, 2 (February 1988), // 216-222. http://dx.doi.org/10.1145/42372.42381 // Threshold for prefering the BINV algorithm. The paper suggests 10, // Ranlib uses 30, and GSL uses 14. const BINV_THRESHOLD: f64 = 10.; if (self.n as f64) * p < BINV_THRESHOLD && self.n <= (::std::i32::MAX as u64) { // Use the BINV algorithm. let s = p / q; let a = ((self.n + 1) as f64) * s; let mut r = q.powi(self.n as i32); let mut u: f64 = rng.gen(); let mut x = 0; while u > r as f64 { u -= r; x += 1; r *= a / (x as f64) - s; } result = x; } else { // Use the BTPE algorithm. // Threshold for using the squeeze algorithm. This can be freely // chosen based on performance. Ranlib and GSL use 20. const SQUEEZE_THRESHOLD: i64 = 20; // Step 0: Calculate constants as functions of `n` and `p`. let n = self.n as f64; let np = n * p; let npq = np * q; let f_m = np + p; let m = f64_to_i64(f_m); // radius of triangle region, since height=1 also area of region let p1 = (2.195 * npq.sqrt() - 4.6 * q).floor() + 0.5; // tip of triangle let x_m = (m as f64) + 0.5; // left edge of triangle let x_l = x_m - p1; // right edge of triangle let x_r = x_m + p1; let c = 0.134 + 20.5 / (15.3 + (m as f64)); // p1 + area of parallelogram region let p2 = p1 * (1. + 2. * c); fn lambda(a: f64) -> f64 { a * (1. + 0.5 * a) } let lambda_l = lambda((f_m - x_l) / (f_m - x_l * p)); let lambda_r = lambda((x_r - f_m) / (x_r * q)); // p1 + area of left tail let p3 = p2 + c / lambda_l; // p1 + area of right tail let p4 = p3 + c / lambda_r; // return value let mut y: i64; let gen_u = Uniform::new(0., p4); let gen_v = Uniform::new(0., 1.); loop { // Step 1: Generate `u` for selecting the region. If region 1 is // selected, generate a triangularly distributed variate. let u = gen_u.sample(rng); let mut v = gen_v.sample(rng); if !(u > p1) { y = f64_to_i64(x_m - p1 * v + u); break; } if !(u > p2) { // Step 2: Region 2, parallelograms. Check if region 2 is // used. If so, generate `y`. let x = x_l + (u - p1) / c; v = v * c + 1.0 - (x - x_m).abs() / p1; if v > 1. { continue; } else { y = f64_to_i64(x); } } else if !(u > p3) { // Step 3: Region 3, left exponential tail. y = f64_to_i64(x_l + v.ln() / lambda_l); if y < 0 { continue; } else { v *= (u - p2) * lambda_l; } } else { // Step 4: Region 4, right exponential tail. y = f64_to_i64(x_r - v.ln() / lambda_r); if y > 0 && (y as u64) > self.n { continue; } else { v *= (u - p3) * lambda_r; } } // Step 5: Acceptance/rejection comparison. // Step 5.0: Test for appropriate method of evaluating f(y). let k = (y - m).abs(); if !(k > SQUEEZE_THRESHOLD && (k as f64) < 0.5 * npq - 1.) { // Step 5.1: Evaluate f(y) via the recursive relationship. Start the // search from the mode. let s = p / q; let a = s * (n + 1.); let mut f = 1.0; if m < y { let mut i = m; loop { i += 1; f *= a / (i as f64) - s; if i == y { break; } } } else if m > y { let mut i = y; loop { i += 1; f /= a / (i as f64) - s; if i == m { break; } } } if v > f { continue; } else { break; } } // Step 5.2: Squeezing. Check the value of ln(v) againts upper and // lower bound of ln(f(y)). let k = k as f64; let rho = (k / npq) * ((k * (k / 3. + 0.625) + 1. / 6.) / npq + 0.5); let t = -0.5 * k * k / npq; let alpha = v.ln(); if alpha < t - rho { break; } if alpha > t + rho { continue; } // Step 5.3: Final acceptance/rejection test. let x1 = (y + 1) as f64; let f1 = (m + 1) as f64; let z = (f64_to_i64(n) + 1 - m) as f64; let w = (f64_to_i64(n) - y + 1) as f64; fn stirling(a: f64) -> f64 { let a2 = a * a; (13860. - (462. - (132. - (99. - 140. / a2) / a2) / a2) / a2) / a / 166320. } if alpha > x_m * (f1 / x1).ln() + (n - (m as f64) + 0.5) * (z / w).ln() + ((y - m) as f64) * (w * p / (x1 * q)).ln() // We use the signs from the GSL implementation, which are // different than the ones in the reference. According to // the GSL authors, the new signs were verified to be // correct by one of the original designers of the // algorithm. + stirling(f1) + stirling(z) - stirling(x1) - stirling(w) { continue; } break; } assert!(y >= 0); result = y as u64; } // Invert the result for p < 0.5. if p != self.p
{ self.n - result }
conditional_block
binomial.rs
4) { // Use the BINV algorithm. let s = p / q; let a = ((self.n + 1) as f64) * s; let mut r = q.powi(self.n as i32); let mut u: f64 = rng.gen(); let mut x = 0; while u > r as f64 { u -= r; x += 1; r *= a / (x as f64) - s; } result = x; } else { // Use the BTPE algorithm. // Threshold for using the squeeze algorithm. This can be freely // chosen based on performance. Ranlib and GSL use 20. const SQUEEZE_THRESHOLD: i64 = 20; // Step 0: Calculate constants as functions of `n` and `p`. let n = self.n as f64; let np = n * p; let npq = np * q; let f_m = np + p; let m = f64_to_i64(f_m); // radius of triangle region, since height=1 also area of region let p1 = (2.195 * npq.sqrt() - 4.6 * q).floor() + 0.5; // tip of triangle let x_m = (m as f64) + 0.5; // left edge of triangle let x_l = x_m - p1; // right edge of triangle let x_r = x_m + p1; let c = 0.134 + 20.5 / (15.3 + (m as f64)); // p1 + area of parallelogram region let p2 = p1 * (1. + 2. * c); fn lambda(a: f64) -> f64 { a * (1. + 0.5 * a) } let lambda_l = lambda((f_m - x_l) / (f_m - x_l * p)); let lambda_r = lambda((x_r - f_m) / (x_r * q)); // p1 + area of left tail let p3 = p2 + c / lambda_l; // p1 + area of right tail let p4 = p3 + c / lambda_r; // return value let mut y: i64; let gen_u = Uniform::new(0., p4); let gen_v = Uniform::new(0., 1.); loop { // Step 1: Generate `u` for selecting the region. If region 1 is // selected, generate a triangularly distributed variate. let u = gen_u.sample(rng); let mut v = gen_v.sample(rng); if !(u > p1) { y = f64_to_i64(x_m - p1 * v + u); break; } if !(u > p2) { // Step 2: Region 2, parallelograms. Check if region 2 is // used. If so, generate `y`. let x = x_l + (u - p1) / c; v = v * c + 1.0 - (x - x_m).abs() / p1; if v > 1. { continue; } else { y = f64_to_i64(x); } } else if !(u > p3) { // Step 3: Region 3, left exponential tail. y = f64_to_i64(x_l + v.ln() / lambda_l); if y < 0 { continue; } else { v *= (u - p2) * lambda_l; } } else { // Step 4: Region 4, right exponential tail. y = f64_to_i64(x_r - v.ln() / lambda_r); if y > 0 && (y as u64) > self.n { continue; } else { v *= (u - p3) * lambda_r; } } // Step 5: Acceptance/rejection comparison. // Step 5.0: Test for appropriate method of evaluating f(y). let k = (y - m).abs(); if !(k > SQUEEZE_THRESHOLD && (k as f64) < 0.5 * npq - 1.) { // Step 5.1: Evaluate f(y) via the recursive relationship. Start the // search from the mode. let s = p / q; let a = s * (n + 1.); let mut f = 1.0; if m < y { let mut i = m; loop { i += 1; f *= a / (i as f64) - s; if i == y { break; } } } else if m > y { let mut i = y; loop { i += 1; f /= a / (i as f64) - s; if i == m { break; } } } if v > f { continue; } else { break; } } // Step 5.2: Squeezing. Check the value of ln(v) againts upper and // lower bound of ln(f(y)). let k = k as f64; let rho = (k / npq) * ((k * (k / 3. + 0.625) + 1. / 6.) / npq + 0.5); let t = -0.5 * k * k / npq; let alpha = v.ln(); if alpha < t - rho { break; } if alpha > t + rho { continue; } // Step 5.3: Final acceptance/rejection test. let x1 = (y + 1) as f64; let f1 = (m + 1) as f64; let z = (f64_to_i64(n) + 1 - m) as f64; let w = (f64_to_i64(n) - y + 1) as f64; fn stirling(a: f64) -> f64 { let a2 = a * a; (13860. - (462. - (132. - (99. - 140. / a2) / a2) / a2) / a2) / a / 166320. } if alpha > x_m * (f1 / x1).ln() + (n - (m as f64) + 0.5) * (z / w).ln() + ((y - m) as f64) * (w * p / (x1 * q)).ln() // We use the signs from the GSL implementation, which are // different than the ones in the reference. According to // the GSL authors, the new signs were verified to be // correct by one of the original designers of the // algorithm. + stirling(f1) + stirling(z) - stirling(x1) - stirling(w) { continue; } break; } assert!(y >= 0); result = y as u64; } // Invert the result for p < 0.5. if p != self.p { self.n - result } else { result } } } #[cfg(test)] mod test { use super::Binomial; use crate::distributions::Distribution; use crate::Rng; fn test_binomial_mean_and_variance<R: Rng>(n: u64, p: f64, rng: &mut R) { let binomial = Binomial::new(n, p); let expected_mean = n as f64 * p; let expected_variance = n as f64 * p * (1.0 - p); let mut results = [0.0; 1000]; for i in results.iter_mut() { *i = binomial.sample(rng) as f64; } let mean = results.iter().sum::<f64>() / results.len() as f64; assert!( (mean as f64 - expected_mean).abs() < expected_mean / 50.0, "mean: {}, expected_mean: {}", mean, expected_mean ); let variance = results.iter().map(|x| (x - mean) * (x - mean)).sum::<f64>() / results.len() as f64; assert!( (variance - expected_variance).abs() < expected_variance / 10.0, "variance: {}, expected_variance: {}", variance, expected_variance ); } #[test] #[cfg_attr(miri, ignore)] // Miri is too slow fn
test_binomial
identifier_name
binomial.rs
) -> Binomial { assert!(p >= 0.0, "Binomial::new called with p < 0"); assert!(p <= 1.0, "Binomial::new called with p > 1"); Binomial { n, p } } } /// Convert a `f64` to an `i64`, panicing on overflow. // In the future (Rust 1.34), this might be replaced with `TryFrom`. fn f64_to_i64(x: f64) -> i64 { assert!(x < (::std::i64::MAX as f64)); x as i64 } impl Distribution<u64> for Binomial { fn sample<R: Rng + ?Sized>(&self, rng: &mut R) -> u64 { // Handle these values directly. if self.p == 0.0 { return 0; } else if self.p == 1.0 { return self.n; } // The binomial distribution is symmetrical with respect to p -> 1-p, // k -> n-k switch p so that it is less than 0.5 - this allows for lower // expected values we will just invert the result at the end let p = if self.p <= 0.5 { self.p } else { 1.0 - self.p }; let result; let q = 1. - p; // For small n * min(p, 1 - p), the BINV algorithm based on the inverse // transformation of the binomial distribution is efficient. Otherwise, // the BTPE algorithm is used. // // Voratas Kachitvichyanukul and Bruce W. Schmeiser. 1988. Binomial // random variate generation. Commun. ACM 31, 2 (February 1988), // 216-222. http://dx.doi.org/10.1145/42372.42381 // Threshold for prefering the BINV algorithm. The paper suggests 10, // Ranlib uses 30, and GSL uses 14. const BINV_THRESHOLD: f64 = 10.; if (self.n as f64) * p < BINV_THRESHOLD && self.n <= (::std::i32::MAX as u64) { // Use the BINV algorithm. let s = p / q; let a = ((self.n + 1) as f64) * s; let mut r = q.powi(self.n as i32); let mut u: f64 = rng.gen(); let mut x = 0; while u > r as f64 { u -= r; x += 1; r *= a / (x as f64) - s; } result = x; } else { // Use the BTPE algorithm. // Threshold for using the squeeze algorithm. This can be freely // chosen based on performance. Ranlib and GSL use 20. const SQUEEZE_THRESHOLD: i64 = 20; // Step 0: Calculate constants as functions of `n` and `p`. let n = self.n as f64; let np = n * p; let npq = np * q; let f_m = np + p; let m = f64_to_i64(f_m); // radius of triangle region, since height=1 also area of region let p1 = (2.195 * npq.sqrt() - 4.6 * q).floor() + 0.5; // tip of triangle let x_m = (m as f64) + 0.5; // left edge of triangle let x_l = x_m - p1; // right edge of triangle let x_r = x_m + p1; let c = 0.134 + 20.5 / (15.3 + (m as f64)); // p1 + area of parallelogram region let p2 = p1 * (1. + 2. * c); fn lambda(a: f64) -> f64
let lambda_l = lambda((f_m - x_l) / (f_m - x_l * p)); let lambda_r = lambda((x_r - f_m) / (x_r * q)); // p1 + area of left tail let p3 = p2 + c / lambda_l; // p1 + area of right tail let p4 = p3 + c / lambda_r; // return value let mut y: i64; let gen_u = Uniform::new(0., p4); let gen_v = Uniform::new(0., 1.); loop { // Step 1: Generate `u` for selecting the region. If region 1 is // selected, generate a triangularly distributed variate. let u = gen_u.sample(rng); let mut v = gen_v.sample(rng); if !(u > p1) { y = f64_to_i64(x_m - p1 * v + u); break; } if !(u > p2) { // Step 2: Region 2, parallelograms. Check if region 2 is // used. If so, generate `y`. let x = x_l + (u - p1) / c; v = v * c + 1.0 - (x - x_m).abs() / p1; if v > 1. { continue; } else { y = f64_to_i64(x); } } else if !(u > p3) { // Step 3: Region 3, left exponential tail. y = f64_to_i64(x_l + v.ln() / lambda_l); if y < 0 { continue; } else { v *= (u - p2) * lambda_l; } } else { // Step 4: Region 4, right exponential tail. y = f64_to_i64(x_r - v.ln() / lambda_r); if y > 0 && (y as u64) > self.n { continue; } else { v *= (u - p3) * lambda_r; } } // Step 5: Acceptance/rejection comparison. // Step 5.0: Test for appropriate method of evaluating f(y). let k = (y - m).abs(); if !(k > SQUEEZE_THRESHOLD && (k as f64) < 0.5 * npq - 1.) { // Step 5.1: Evaluate f(y) via the recursive relationship. Start the // search from the mode. let s = p / q; let a = s * (n + 1.); let mut f = 1.0; if m < y { let mut i = m; loop { i += 1; f *= a / (i as f64) - s; if i == y { break; } } } else if m > y { let mut i = y; loop { i += 1; f /= a / (i as f64) - s; if i == m { break; } } } if v > f { continue; } else { break; } } // Step 5.2: Squeezing. Check the value of ln(v) againts upper and // lower bound of ln(f(y)). let k = k as f64; let rho = (k / npq) * ((k * (k / 3. + 0.625) + 1. / 6.) / npq + 0.5); let t = -0.5 * k * k / npq; let alpha = v.ln(); if alpha < t - rho { break; } if alpha > t + rho { continue; } // Step 5.3: Final acceptance/rejection test. let x1 = (y + 1) as f64; let f1 = (m + 1) as f64; let z = (f64_to_i64(n) + 1 - m) as f64; let w = (f64_to_i64(n) - y + 1) as f64; fn stirling(a: f64) -> f64 { let a2 = a * a; (13860. - (462. - (132. - (99. - 140. / a2) / a2) / a2) / a2) / a / 166320. } if alpha > x_m *
{ a * (1. + 0.5 * a) }
identifier_body
main.rs
( game: &Arc<RwLock<Game>>, user_id: usize, msg: WsMsg, tx: &mpsc::UnboundedSender<WsMsg> ) -> Result<()> { match msg { WsMsg::Login(username) => { if game.read().await.players.len() >= MAX_N_PLAYERS { tx.send(WsMsg::LoginRejected(LoginRejectedReason::GameIsFull))?; return Ok(()) } if game.read().await.players.values().any(|player| player.name == username) { tx.send(WsMsg::LoginRejected(LoginRejectedReason::UsernameIsTaken))?; return Ok(()) } tx.send(WsMsg::LoginAccepted)?; let hand = game.write().await.answers.draw(N_CARDS_IN_HAND); let player = Player { name: username.clone(), hand: hand.clone(), score: 0, }; game.write().await.players.insert(user_id, player); // Notify other players game.write().await.broadcast_to_players(&WsMsg::PlayerJoined { name: username })?; // Only start new round if there are enough players if game.read().await.players.len() >= MIN_N_PLAYERS { let game = &mut game.write().await; let round = if let Some(round) = &game.round { round } else { // TODO lobby println!("Starting new round"); game.new_round()?; game.round.as_ref().unwrap() }; // If in judgement, don't send NewRound if round.state == RoundState::Answering { let role = if round.czar == user_id { Role::Czar } else { Role::Player }; tx.send(WsMsg::NewRound { role, prompt: round.prompt.clone(), hand: hand, })?; } } Ok(()) }, // WsMsg::Register(name) => todo!(), // WsMsg::Ready => todo!(), // WsMsg::NotReady => todo!(), WsMsg::SubmitAnswer(answers) => { if let Game { clients, players, round: Some(round), .. } = &mut *game.write().await { if round.state != RoundState::Answering { eprintln!("invalid query SubmitAnswer: round is in judgement phase"); return Ok(()) } if round.czar == user_id { eprintln!("invalid query SubmitAnswer: player is Czar"); return Ok(()) } match round.answers.entry(user_id) { hash_map::Entry::Occupied(_) => { eprintln!("invalid query SubmitAnswer: player already submitted answer") }, hash_map::Entry::Vacant(entry) => { let hand = &mut players.get_mut(&user_id).unwrap().hand; if !answers.iter().all(|x| hand.contains(x)) { eprintln!("invalid query SubmitAnswer: cards are not in player's deck"); return Ok(()) } println!("SubmitAnswer({})", answers.iter().map(Answer::to_string).collect::<Vec<_>>().join(", ")); // Remove cards from player's hand hand.retain(|x| !answers.contains(x)); // Insert cards into submitted answers entry.insert(answers); tx.send(WsMsg::AnswerAccepted)?; }, } // Check whether all players have answered if round.answers.len() == players.len() - 1 { round.state = RoundState::Judging; // If so, notify them that JUDGEMENT HAS BEGUN // TODO maybe obfuscate the player IDs before sending for id in players.keys() { clients[id].send(WsMsg::ReadyToJudge(round.answers.clone()))?; } } } else { eprintln!("invalid query SubmitAnswer: there is no ongoing round"); } // TODO send AnswerAccepted/Rejected messages Ok(()) }, WsMsg::SubmitJudgement(answer_id) => { let mut new_round = false; if let Game { clients, players, round: Some(round), .. } = &mut *game.write().await { if round.state != RoundState::Judging { eprintln!("invalid query SubmitAnswer: round is in judgement phase"); return Ok(()) } if round.czar != user_id { eprintln!("invalid query SubmitJudgement: player isn't Czar"); return Ok(()) } match round.answers.get(&answer_id) { None => { eprintln!("invalid query SubmitJudgement: user ID does not exist"); }, Some(winning_answers) => { let winner = { // Increment winner's scores let winner = players.get_mut(&answer_id).unwrap(); winner.score += 1; // Get winner's name winner.name.clone() }; let scores = players.values().map(|player| (player.name.clone(), player.score)).collect(); let msg = WsMsg::RoundEnded { winner, winning_answers: winning_answers.clone(), scores, }; // Notify end of round, provide winner and scores for id in players.keys() { clients[id].send(msg.clone())?; } new_round = true; } } } else { eprintln!("invalid query SubmitAnswer: there is no ongoing round"); } if new_round { game.write().await.new_round()?; } // TODO send JudgementAccepted/Rejected messages Ok(()) }, _ => unreachable!(), } } async fn user_connected(game: Arc<RwLock<Game>>, socket: WebSocket) { let my_id = NEXT_USER_ID.fetch_add(1, Ordering::Relaxed); println!("User connected: #{}", my_id); let (tx, mut rx) = socket.split(); // Manage outgoing messages to this user let tx = { let (tx2, rx) = mpsc::unbounded_channel(); tokio::task::spawn(rx.map(|msg| { Ok(Message::text(serde_json::to_string(&msg).unwrap())) }).forward(tx).map(move |result| { if let Err(e) = result { eprintln!("websocket send error: {}", e); } })); tx2 }; game.write().await.clients.insert(my_id, tx.clone()); // Manage incoming messages from this user while let Some(result) = rx.next().await { let msg = match result { Ok(msg) => msg, Err(e) => { eprintln!("websocket error with user {}: {}", my_id, e); break; } }; if let Ok(text) = msg.to_str() { if let Ok(response) = serde_json::from_str::<WsMsg>(text) { if let Err(_) = process_message(&game, my_id, response, &tx).await { eprintln!("Error while processing message from player #{}", my_id); break; } } else { eprintln!("cannot read message"); } } } println!("Client #{} disconnected", my_id); user_disconnected(game, my_id).await; } async fn user_disconnected(game: Arc<RwLock<Game>>, user_id: usize) { let game = &mut *game.write().await; game.clients.remove(&user_id); if let Some(player) = game.players.remove(&user_id) { // Discard player's answers game.answers.discard(&player.hand); // Discard player's submitted answers, if any let mut user_is_czar = false; if let Game { answers, round: Some(Round { answers: submitted_answers, czar, .. }), .. } = game { if let Some(cards) = submitted_answers.remove(&user_id) { answers.discard(&cards); } user_is_czar = *czar == user_id; } // If player is Czar, return submitted answers to owners and restart round if user_is_czar { let mut round = game.round.take().unwrap(); game.prompts.discard(&[round.prompt]); for (id, player) in game.players.iter_mut() { player.hand.extend(round.answers.remove(id).into_iter().flatten()); } if game.players.len() > 0 { game.new_round().expect("Couldn't start new round"); } } // Notify other players game.broadcast_to_players(&WsMsg::PlayerLeft { name: player.name.clone() }); } // If not enough players, cancel round if game.players.len() < MIN_N_PLAYERS { game.round = None; game.answers.reset(); game.prompts.reset(); for id in game.players.keys() { game.clients[id].send(WsMsg::GameEnded); } // Clear player hands, to avoid double-discard for player in game.players.values_mut() { player.hand.clear(); } } } use ron; use std::fs::File; use serde::de::DeserializeOwned; fn
load_deck
identifier_name
main.rs
#{}", round.czar); // Distribute cards and notify players self.distribute_cards(); for (id, player) in &mut self.players { let role = if *id == round.czar { Role::Czar } else { Role::Player }; self.clients[id].send(WsMsg::NewRound { role, prompt: round.prompt.clone(), hand: player.hand.clone(), })?; } // Set new round self.round = Some(round); Ok(()) } fn broadcast_to_players(&mut self, msg: &WsMsg) -> Result<()> { for id in self.players.keys() { self.clients[id].send(msg.clone())?; } Ok(()) } } #[derive(PartialEq)] enum RoundState { Answering, Judging, } struct Round { prompt: Prompt, czar: usize, answers: HashMap<usize, Vec<Answer>>, state: RoundState, } struct Player { name: String, hand: Vec<Answer>, score: u64, } static NEXT_USER_ID: AtomicUsize = AtomicUsize::new(1); async fn process_message( game: &Arc<RwLock<Game>>, user_id: usize, msg: WsMsg, tx: &mpsc::UnboundedSender<WsMsg> ) -> Result<()> { match msg { WsMsg::Login(username) => { if game.read().await.players.len() >= MAX_N_PLAYERS { tx.send(WsMsg::LoginRejected(LoginRejectedReason::GameIsFull))?; return Ok(()) } if game.read().await.players.values().any(|player| player.name == username) { tx.send(WsMsg::LoginRejected(LoginRejectedReason::UsernameIsTaken))?; return Ok(()) } tx.send(WsMsg::LoginAccepted)?; let hand = game.write().await.answers.draw(N_CARDS_IN_HAND); let player = Player { name: username.clone(), hand: hand.clone(), score: 0, }; game.write().await.players.insert(user_id, player); // Notify other players game.write().await.broadcast_to_players(&WsMsg::PlayerJoined { name: username })?; // Only start new round if there are enough players if game.read().await.players.len() >= MIN_N_PLAYERS { let game = &mut game.write().await; let round = if let Some(round) = &game.round { round } else { // TODO lobby println!("Starting new round"); game.new_round()?; game.round.as_ref().unwrap() }; // If in judgement, don't send NewRound if round.state == RoundState::Answering { let role = if round.czar == user_id { Role::Czar } else { Role::Player }; tx.send(WsMsg::NewRound { role, prompt: round.prompt.clone(), hand: hand, })?; } } Ok(()) }, // WsMsg::Register(name) => todo!(), // WsMsg::Ready => todo!(), // WsMsg::NotReady => todo!(), WsMsg::SubmitAnswer(answers) => { if let Game { clients, players, round: Some(round), .. } = &mut *game.write().await { if round.state != RoundState::Answering { eprintln!("invalid query SubmitAnswer: round is in judgement phase"); return Ok(()) } if round.czar == user_id { eprintln!("invalid query SubmitAnswer: player is Czar"); return Ok(()) } match round.answers.entry(user_id) { hash_map::Entry::Occupied(_) => { eprintln!("invalid query SubmitAnswer: player already submitted answer") }, hash_map::Entry::Vacant(entry) => { let hand = &mut players.get_mut(&user_id).unwrap().hand; if !answers.iter().all(|x| hand.contains(x)) { eprintln!("invalid query SubmitAnswer: cards are not in player's deck"); return Ok(()) } println!("SubmitAnswer({})", answers.iter().map(Answer::to_string).collect::<Vec<_>>().join(", ")); // Remove cards from player's hand hand.retain(|x| !answers.contains(x)); // Insert cards into submitted answers entry.insert(answers); tx.send(WsMsg::AnswerAccepted)?; }, } // Check whether all players have answered if round.answers.len() == players.len() - 1 { round.state = RoundState::Judging; // If so, notify them that JUDGEMENT HAS BEGUN // TODO maybe obfuscate the player IDs before sending for id in players.keys() { clients[id].send(WsMsg::ReadyToJudge(round.answers.clone()))?; } } } else { eprintln!("invalid query SubmitAnswer: there is no ongoing round"); } // TODO send AnswerAccepted/Rejected messages Ok(()) }, WsMsg::SubmitJudgement(answer_id) => { let mut new_round = false; if let Game { clients, players, round: Some(round), .. } = &mut *game.write().await { if round.state != RoundState::Judging { eprintln!("invalid query SubmitAnswer: round is in judgement phase"); return Ok(()) } if round.czar != user_id { eprintln!("invalid query SubmitJudgement: player isn't Czar"); return Ok(()) } match round.answers.get(&answer_id) { None => { eprintln!("invalid query SubmitJudgement: user ID does not exist"); }, Some(winning_answers) => { let winner = { // Increment winner's scores let winner = players.get_mut(&answer_id).unwrap(); winner.score += 1; // Get winner's name winner.name.clone() }; let scores = players.values().map(|player| (player.name.clone(), player.score)).collect(); let msg = WsMsg::RoundEnded { winner, winning_answers: winning_answers.clone(), scores, }; // Notify end of round, provide winner and scores for id in players.keys() { clients[id].send(msg.clone())?; } new_round = true; } } } else { eprintln!("invalid query SubmitAnswer: there is no ongoing round"); } if new_round { game.write().await.new_round()?; } // TODO send JudgementAccepted/Rejected messages Ok(()) }, _ => unreachable!(), } } async fn user_connected(game: Arc<RwLock<Game>>, socket: WebSocket) { let my_id = NEXT_USER_ID.fetch_add(1, Ordering::Relaxed); println!("User connected: #{}", my_id); let (tx, mut rx) = socket.split(); // Manage outgoing messages to this user let tx = { let (tx2, rx) = mpsc::unbounded_channel(); tokio::task::spawn(rx.map(|msg| { Ok(Message::text(serde_json::to_string(&msg).unwrap())) }).forward(tx).map(move |result| { if let Err(e) = result { eprintln!("websocket send error: {}", e); } })); tx2 }; game.write().await.clients.insert(my_id, tx.clone()); // Manage incoming messages from this user while let Some(result) = rx.next().await { let msg = match result { Ok(msg) => msg, Err(e) => { eprintln!("websocket error with user {}: {}", my_id, e); break; } }; if let Ok(text) = msg.to_str() { if let Ok(response) = serde_json::from_str::<WsMsg>(text) { if let Err(_) = process_message(&game, my_id, response, &tx).await { eprintln!("Error while processing message from player #{}", my_id); break; } } else { eprintln!("cannot read message"); } } } println!("Client #{} disconnected", my_id); user_disconnected(game, my_id).await; } async fn user_disconnected(game: Arc<RwLock<Game>>, user_id: usize)
{ let game = &mut *game.write().await; game.clients.remove(&user_id); if let Some(player) = game.players.remove(&user_id) { // Discard player's answers game.answers.discard(&player.hand); // Discard player's submitted answers, if any let mut user_is_czar = false; if let Game { answers, round: Some(Round { answers: submitted_answers, czar, .. }), .. } = game { if let Some(cards) = submitted_answers.remove(&user_id) { answers.discard(&cards); } user_is_czar = *czar == user_id; }
identifier_body
main.rs
} // Find next czar let mut player_ids = self.players.keys().collect::<Vec<_>>(); player_ids.sort_unstable(); if let Err(idx) = player_ids.binary_search(&&next_czar) { // There's no player with ID next_czar if idx == player_ids.len() { // There isn't a greater key next_czar = *player_ids[0]; } else { // There is a key greater than next_czar next_czar = *player_ids[idx]; } } // Create new round println!("Players to choose from: {:?}", self.players.keys().map(|u| u.to_string()).collect::<Vec<_>>().join(", ")); let round = Round { prompt: self.prompts.draw_once(), // TODO cycle Czars czar: next_czar, answers: Default::default(), state: RoundState::Answering, }; println!("Next czar is Player #{}", round.czar); // Distribute cards and notify players self.distribute_cards(); for (id, player) in &mut self.players { let role = if *id == round.czar { Role::Czar } else { Role::Player }; self.clients[id].send(WsMsg::NewRound { role, prompt: round.prompt.clone(), hand: player.hand.clone(), })?; } // Set new round self.round = Some(round); Ok(()) } fn broadcast_to_players(&mut self, msg: &WsMsg) -> Result<()> { for id in self.players.keys() { self.clients[id].send(msg.clone())?; } Ok(()) } } #[derive(PartialEq)] enum RoundState { Answering, Judging, } struct Round { prompt: Prompt, czar: usize, answers: HashMap<usize, Vec<Answer>>, state: RoundState, } struct Player { name: String, hand: Vec<Answer>, score: u64, } static NEXT_USER_ID: AtomicUsize = AtomicUsize::new(1); async fn process_message( game: &Arc<RwLock<Game>>, user_id: usize, msg: WsMsg, tx: &mpsc::UnboundedSender<WsMsg> ) -> Result<()> { match msg { WsMsg::Login(username) => { if game.read().await.players.len() >= MAX_N_PLAYERS { tx.send(WsMsg::LoginRejected(LoginRejectedReason::GameIsFull))?; return Ok(()) } if game.read().await.players.values().any(|player| player.name == username) { tx.send(WsMsg::LoginRejected(LoginRejectedReason::UsernameIsTaken))?; return Ok(()) } tx.send(WsMsg::LoginAccepted)?; let hand = game.write().await.answers.draw(N_CARDS_IN_HAND); let player = Player { name: username.clone(), hand: hand.clone(), score: 0, }; game.write().await.players.insert(user_id, player); // Notify other players game.write().await.broadcast_to_players(&WsMsg::PlayerJoined { name: username })?; // Only start new round if there are enough players if game.read().await.players.len() >= MIN_N_PLAYERS { let game = &mut game.write().await; let round = if let Some(round) = &game.round { round } else { // TODO lobby println!("Starting new round"); game.new_round()?; game.round.as_ref().unwrap() }; // If in judgement, don't send NewRound if round.state == RoundState::Answering { let role = if round.czar == user_id { Role::Czar } else { Role::Player }; tx.send(WsMsg::NewRound { role, prompt: round.prompt.clone(), hand: hand, })?; } } Ok(()) }, // WsMsg::Register(name) => todo!(), // WsMsg::Ready => todo!(), // WsMsg::NotReady => todo!(), WsMsg::SubmitAnswer(answers) => { if let Game { clients, players, round: Some(round), .. } = &mut *game.write().await { if round.state != RoundState::Answering { eprintln!("invalid query SubmitAnswer: round is in judgement phase"); return Ok(()) } if round.czar == user_id { eprintln!("invalid query SubmitAnswer: player is Czar"); return Ok(()) } match round.answers.entry(user_id) { hash_map::Entry::Occupied(_) => { eprintln!("invalid query SubmitAnswer: player already submitted answer") }, hash_map::Entry::Vacant(entry) => { let hand = &mut players.get_mut(&user_id).unwrap().hand; if !answers.iter().all(|x| hand.contains(x)) { eprintln!("invalid query SubmitAnswer: cards are not in player's deck"); return Ok(()) } println!("SubmitAnswer({})", answers.iter().map(Answer::to_string).collect::<Vec<_>>().join(", ")); // Remove cards from player's hand hand.retain(|x| !answers.contains(x)); // Insert cards into submitted answers entry.insert(answers); tx.send(WsMsg::AnswerAccepted)?;
}, } // Check whether all players have answered if round.answers.len() == players.len() - 1 { round.state = RoundState::Judging; // If so, notify them that JUDGEMENT HAS BEGUN // TODO maybe obfuscate the player IDs before sending for id in players.keys() { clients[id].send(WsMsg::ReadyToJudge(round.answers.clone()))?; } } } else { eprintln!("invalid query SubmitAnswer: there is no ongoing round"); } // TODO send AnswerAccepted/Rejected messages Ok(()) }, WsMsg::SubmitJudgement(answer_id) => { let mut new_round = false; if let Game { clients, players, round: Some(round), .. } = &mut *game.write().await { if round.state != RoundState::Judging { eprintln!("invalid query SubmitAnswer: round is in judgement phase"); return Ok(()) } if round.czar != user_id { eprintln!("invalid query SubmitJudgement: player isn't Czar"); return Ok(()) } match round.answers.get(&answer_id) { None => { eprintln!("invalid query SubmitJudgement: user ID does not exist"); }, Some(winning_answers) => { let winner = { // Increment winner's scores let winner = players.get_mut(&answer_id).unwrap(); winner.score += 1; // Get winner's name winner.name.clone() }; let scores = players.values().map(|player| (player.name.clone(), player.score)).collect(); let msg = WsMsg::RoundEnded { winner, winning_answers: winning_answers.clone(), scores, }; // Notify end of round, provide winner and scores for id in players.keys() { clients[id].send(msg.clone())?; } new_round = true; } } } else { eprintln!("invalid query SubmitAnswer: there is no ongoing round"); } if new_round { game.write().await.new_round()?; } // TODO send JudgementAccepted/Rejected messages Ok(()) }, _ => unreachable!(), } } async fn user_connected(game: Arc<RwLock<Game>>, socket: WebSocket) { let my_id = NEXT_USER_ID.fetch_add(1, Ordering::Relaxed); println!("User connected: #{}", my_id); let (tx, mut rx) = socket.split(); // Manage outgoing messages to this user let tx = { let (tx2, rx) = mpsc::unbounded_channel(); tokio::task::spawn(rx.map(|msg| { Ok(Message::text(serde_json::to_string(&msg).unwrap())) }).forward(tx).map(move |result| { if let Err(e) = result { eprintln!("websocket send error: {}", e); } })); tx2 }; game.write().await.clients.insert(my_id, tx.clone()); // Manage incoming messages from this user while let Some(result) = rx.next().await { let msg = match result { Ok(msg) => msg, Err(e) => { eprintln!("websocket error with user {}: {}", my_id, e); break; } }; if let Ok(text) = msg.to_str() { if let Ok(response) = serde_json::from_str::<WsMsg>(text) { if let Err(_) = process_message(&game, my_id, response, &tx).await { eprintln!("Error while processing message
random_line_split
cougballoon_rcv.py
, 'Hours':RMChours, 'Minutes':RMCminutes, 'Seconds':RMCseconds},'External temperature(deg F)':extTemp, 'Internal temperature(deg F)':intTemp, 'Video Transmitter temperature(deg F)':vidTemp, 'Carbon Monoxide level(ppm)':COlevel, 'Methane level(ppm)':CH4level, 'HackHD camera statuses':HackStatus } ] data_string = json.dumps(JSONdata) JSONdataString = str(data_string) newJSONdata = re.match(r'\[(.*)', JSONdataString) newJSONdata2 = "," + newJSONdata.group(1) f = open('/Users/michaelhamilton/Desktop/json_data.html', 'r+') jumpBack = -1 #jump back 1 spot from the end f.seek(jumpBack, 2) #2 is the end of the file last = f.readline() #read the last line while (last != "]"): #if it's not a ], then keep jumping back jumpBack = jumpBack - 1 #decrement if (last == "]"): f.seek(-1, 2) f.write(newJSONdata2) f.close() #Send it to the server os.system("cd /Users/michaelhamilton/Desktop && scp json_data.html [email protected]:Sites/") print "Updated JSON information was sent to the server." return #Parse out the data from an RMC nmea string def RegExprNMEAdataRMC(line): #if it's an RMC string.... print line newRMCline = re.match( r'\$GPRMC,(\d\d)(\d\d)(\d\d).*,\D,(\d+.\d+),\D,(\d+.\d+),\D,(\d+.\d+),(\d+.\d+),(\d\d)(\d\d)(\d\d),.*,.*', line, re.I) #All data are strings, not integers if (newRMCline): global RMChours RMChours = newRMCline.group(1) #Convert UTC hours to PST(Daylight Savings Time) RMChours = UTCtoPSTDST(RMChours) global RMCminutes RMCminutes = newRMCline.group(2) global RMCseconds RMCseconds = newRMCline.group(3) global RMClatitude RMClatitude = newRMCline.group(4) RMClatitude = StringToFloatGPS(RMClatitude) global RMClongitude RMClongitude = newRMCline.group(5) RMClongitude = StringToFloatGPS(RMClongitude) global RMCspeed RMCspeed = newRMCline.group(6) RMCspeed = StringToFloatGPS(RMCspeed) global RMCheading RMCheading = newRMCline.group(7) RMCheading = StringToFloatGPS(RMCheading) global RMCday RMCday = newRMCline.group(8) global RMCmonth RMCmonth = newRMCline.group(9) global RMCyear RMCyear = newRMCline.group(10) return True else: return False #Parse out the data from an GGA nmea string def RegExprNMEAdataGGA(line): #if it's a GGA string.... print line newGGAline = re.match( r'\$GPGGA,(\d\d)(\d\d)(\d\d).*,(.*..*),\D,(.*..*),\D,\d,\d\d\,\d.\d\d,(\d+.\d),\D.*', line, re.I) #All data are strings, not integers if (newGGAline): global GGAhours GGAhours = newGGAline.group(1) #Convert UTC hours to PST(Daylight Savings Time) GGAhours = UTCtoPSTDST(GGAhours) global GGAminutes GGAminutes = newGGAline.group(2) global GGAseconds GGAseconds = newGGAline.group(3) global GGAlatitude GGAlatitude = newGGAline.group(4) GGAlatitude = StringToFloatGPS(GGAlatitude) global GGAlongitude GGAlongitude = newGGAline.group(5) GGAlongitude = StringToFloatGPS(GGAlongitude) global GGAaltitude GGAaltitude = newGGAline.group(6) GGAaltitude = StringToFloatGPS(GGAaltitude) s2.write(dict(x=datetime.datetime.now().strftime('%Y-%m-%d %H:%M:%S.%f'), y=GGAaltitude)) return True else: return False #Get my login and keys ready ro send data to plot.ly stream_ids = tls.get_credentials_file()['stream_ids'] #Set up the plotly streams stream_id1 = stream_ids[0]#External temperature #A stream_id2 = stream_ids[1]#GGAaltitude #B stream_id3 = stream_ids[2]#Internal temperature #C #stream_id4 = stream_ids[3]#Internal pressure #D stream_id4 = stream_ids[3]#pressureAltitude #D #stream_id5 = stream_ids[4]#Videolynx temperature #E stream_id5 = stream_ids[4]#10DOF temperature #E stream_id6 = stream_ids[5]#CO level in ppm #F stream_id7 = stream_ids[6]#CH4 level in ppm #G stream_id8 = stream_ids[7]#Humidity #J stream_id9 = stream_ids[8]#Roll #L stream_id10 = stream_ids[9]#Pitch #P #stream_id11 = stream_ids[10]#Heading #Q #stream_id12 = stream_ids[11]#Pressure #T stream_id13 = stream_ids[12]#PressureAltitude #U #Graph 1 data, stream names coincide with stream_ids for simplicity #External temperature #A stream1 = Stream( token=stream_id1, maxpoints=20 ) #GGAaltitude #A stream2 = Stream( token=stream_id2, maxpoints=4 ) #Internal temperature #C stream3 = Stream( token=stream_id3, maxpoints=20 ) #pressureAltitude #C stream4 = Stream( token=stream_id4, maxpoints=20 ) #10DOF temperature #E stream5 = Stream( token=stream_id5, maxpoints=20 ) #Graph 2 data, stream names coincide with stream_ids for simplicity #CO level in ppm #G stream6 = Stream( token=stream_id6, maxpoints=20 ) #CH4 level in ppm #G stream7 = Stream( token=stream_id7, maxpoints=20 ) #Roll #L stream9 = Stream( token=stream_id9, maxpoints=20 ) #Pitch #P stream10 = Stream( token=stream_id10, maxpoints=20 ) #Heading #Q #stream11 = Stream( # token=stream_id11, # maxpoints=20 #) #Pressure #T #stream12 = Stream( # token=stream_id12, # maxpoints=20 #) #PressureAltitude #U stream13 = Stream( token=stream_id13, maxpoints=20 ) #Trace names coincide with stream names trace1 = Scatter( x=[], y=[], mode='lines+markers', stream=stream1 ) trace2 = Scatter( x=[], y=[], mode='lines+markers', stream=stream2 ) trace3 = Scatter( x=[], y=[], mode='lines+markers', stream=stream3 ) trace4 = Scatter( x=[], y=[], mode='lines+markers', stream=stream4 ) trace5 = Scatter( x=[], y=[], mode='lines+markers', stream=stream5 ) trace6 = Scatter( x=[], y=[], mode='lines+markers', stream=stream6 ) trace7 = Scatter( x=[], y=[], mode='lines+markers', stream=stream7 ) trace9 = Scatter( x=[], y=[], mode='lines+markers', stream=stream9 ) trace10 = Scatter( x=[], y=[], mode='lines+markers', stream=stream10 )
#trace11 = Scatter( # x=[], # y=[], # mode='lines+markers', # stream=stream11
random_line_split
cougballoon_rcv.py
(nmeaString): #Commented out lines are for .docs, we are using .txt files instead. #f = open('/Users/michaelhamilton/gpsbabel/nmeaRawData.doc', 'a') f = open('/Users/michaelhamilton/gpsbabel/nmeaRawData.txt', 'a') f.write(nmeaString) f.close() saveAllIncomingData(nmeaString) #os.system("cd /Users/michaelhamilton/gpsbabel && ./gpsbabel -i nmea -f nmeaRawData.doc -o kml,deficon=http://encs.vancouver.wsu.edu/~mikehmbn/balloon-icon-map.png,line_color=FF321E98,floating=1 -F cougballoon.kml") os.system("cd /Users/michaelhamilton/gpsbabel && ./gpsbabel -i nmea -f nmeaRawData.txt -o kml,deficon=http://encs.vancouver.wsu.edu/~mikehmbn/balloon-icon-map.png,line_color=FF321E98,floating=1 -F cougballoon.kml") os.system("cd /Users/michaelhamilton/gpsbabel && scp cougballoon.kml [email protected]:Sites/") print "Updated KML file was sent to the server" return #Get JSON data and send it to the server def parseToJson(RMClongitude, RMClatitude, GGAaltitude, RMCspeed, RMCheading, RMCday, RMCmonth, RMCyear, RMChours, RMCminutes, RMCseconds, extTemp, intTemp, vidTemp, COlevel, CH4level, HackStatus): JSONdata2 = { 'cougballoon':[ { 'Longitude':RMClongitude, 'Latitude':RMClatitude, 'Altitude':GGAaltitude, 'Speed':RMCspeed, 'Heading':RMCheading, 'Time':{'Day':RMCday, 'Month':RMCmonth, 'Year':RMCyear, 'Hours':RMChours, 'Minutes':RMCminutes, 'Seconds':RMCseconds},'External temperature(deg F)':extTemp, 'Internal temperature(deg F)':intTemp, 'Video Transmitter temperature(deg F)':vidTemp, 'Carbon Monoxide level(ppm)':COlevel, 'Methane level(ppm)':CH4level, 'HackHD':HackStatus } ] } data_string2 = json.dumps(JSONdata2) #Now post it to json_data.json for the map legend f = open('/Users/michaelhamilton/Desktop/json_data.json', 'w') f.write(data_string2) f.close() os.system("scp /Users/michaelhamilton/Desktop/json_data.json [email protected]:Sites/") #Now to handle it for json_data.html JSONdata = [ { 'Longitude':RMClongitude, 'Latitude':RMClatitude, 'Altitude':GGAaltitude, 'Speed(mph)':RMCspeed, 'Heading':RMCheading, 'Time':{'Day':RMCday, 'Month':RMCmonth, 'Year':RMCyear, 'Hours':RMChours, 'Minutes':RMCminutes, 'Seconds':RMCseconds},'External temperature(deg F)':extTemp, 'Internal temperature(deg F)':intTemp, 'Video Transmitter temperature(deg F)':vidTemp, 'Carbon Monoxide level(ppm)':COlevel, 'Methane level(ppm)':CH4level, 'HackHD camera statuses':HackStatus } ] data_string = json.dumps(JSONdata) JSONdataString = str(data_string) newJSONdata = re.match(r'\[(.*)', JSONdataString) newJSONdata2 = "," + newJSONdata.group(1) f = open('/Users/michaelhamilton/Desktop/json_data.html', 'r+') jumpBack = -1 #jump back 1 spot from the end f.seek(jumpBack, 2) #2 is the end of the file last = f.readline() #read the last line while (last != "]"): #if it's not a ], then keep jumping back jumpBack = jumpBack - 1 #decrement if (last == "]"): f.seek(-1, 2) f.write(newJSONdata2) f.close() #Send it to the server os.system("cd /Users/michaelhamilton/Desktop && scp json_data.html [email protected]:Sites/") print "Updated JSON information was sent to the server." return #Parse out the data from an RMC nmea string def RegExprNMEAdataRMC(line): #if it's an RMC string.... print line newRMCline = re.match( r'\$GPRMC,(\d\d)(\d\d)(\d\d).*,\D,(\d+.\d+),\D,(\d+.\d+),\D,(\d+.\d+),(\d+.\d+),(\d\d)(\d\d)(\d\d),.*,.*', line, re.I) #All data are strings, not integers if (newRMCline): global RMChours RMChours = newRMCline.group(1) #Convert UTC hours to PST(Daylight Savings Time) RMChours = UTCtoPSTDST(RMChours) global RMCminutes RMCminutes = newRMCline.group(2) global RMCseconds RMCseconds = newRMCline.group(3) global RMClatitude RMClatitude = newRMCline.group(4) RMClatitude = StringToFloatGPS(RMClatitude) global RMClongitude RMClongitude = newRMCline.group(5) RMClongitude = StringToFloatGPS(RMClongitude) global RMCspeed RMCspeed = newRMCline.group(6) RMCspeed = StringToFloatGPS(RMCspeed) global RMCheading RMCheading = newRMCline.group(7) RMCheading = StringToFloatGPS(RMCheading) global RMCday RMCday = newRMCline.group(8) global RMCmonth RMCmonth = newRMCline.group(9) global RMCyear RMCyear = newRMCline.group(10) return True else: return False #Parse out the data from an GGA nmea string def RegExprNMEAdataGGA(line): #if it's a GGA string.... print line newGGAline = re.match( r'\$GPGGA,(\d\d)(\d\d)(\d\d).*,(.*..*),\D,(.*..*),\D,\d,\d\d\,\d.\d\d,(\d+.\d),\D.*', line, re.I) #All data are strings, not integers if (newGGAline): global GGAhours GGAhours = newGGAline.group(1) #Convert UTC hours to PST(Daylight Savings Time) GGAhours = UTCtoPSTDST(GGAhours) global GGAminutes GGAminutes = newGGAline.group(2) global GGAseconds GGAseconds = newGGAline.group(3) global GGAlatitude GGAlatitude = newGGAline.group(4) GGAlatitude = StringToFloatGPS(GGAlatitude) global GGAlongitude GGAlongitude = newGGAline.group(5) GGAlongitude = StringToFloatGPS(GGAlongitude) global GGAaltitude GGAaltitude = newGGAline.group(6) GGAaltitude = StringToFloatGPS(GGAaltitude) s2.write(dict(x=datetime.datetime.now().strftime('%Y-%m-%d %H:%M:%S.%f'), y=GGAaltitude)) return True else: return False #Get my login and keys ready ro send data to plot.ly stream_ids = tls.get_credentials_file()['stream_ids'] #Set up the plotly streams stream_id1 = stream_ids[0]#External temperature #A stream_id2 = stream_ids[1]#GGAaltitude #B stream_id3 = stream_ids[2]#Internal temperature #C #stream_id4 = stream_ids[3]#Internal pressure #D stream_id4 = stream_ids[3]#pressureAltitude #D #stream_id5 = stream_ids[4]#Videolynx temperature #E stream_id5 = stream_ids[4]#10DOF temperature #E stream_id6 = stream_ids[5]#CO level in ppm #F stream_id7 = stream_ids[6]#CH4 level in ppm #G stream_id8 = stream_ids[7]#Humidity #J stream_id9 = stream_ids[8]#Roll #L stream_id10 = stream_ids[9]#Pitch #P #stream_id11 = stream_ids
handleGPSdata
identifier_name
cougballoon_rcv.py
#Save all incoming data with a current date/time string def saveData(a): x = datetime.datetime.now().strftime('%Y-%m-%d %H:%M:%S.%f') saveAllIncomingData(x) saveAllIncomingData(a) #Convert GPS strings to floats (Couldn't get str.isnumeric() to work correctly) def StringToFloatGPS(a): a = a.rstrip('\n'); a = a.rstrip('\r'); a = float(a) return a #FIX SO IT DOES NOT RETURN A ZERO!!!!!!!! #COnvert data strings to floats def StringToFloat(a, b): #print len(a) if (len(a) < 4): print "Incomplete data, returning a zero." return b a = a[1:len(a)] a = a.rstrip('\n'); a = a.rstrip('\r'); if (a == "-"): print "Only a negative sign in string, returning a zero." return b a = float(a) return a #Saves all incoming data to a file on the desktop def saveAllIncomingData(c): f = open('/Users/michaelhamilton/Desktop/cougballoonData.txt', 'a') f.write(c) f.close #Convert nmea string to .kml file, send to server def handleGPSdata(nmeaString): #Commented out lines are for .docs, we are using .txt files instead. #f = open('/Users/michaelhamilton/gpsbabel/nmeaRawData.doc', 'a') f = open('/Users/michaelhamilton/gpsbabel/nmeaRawData.txt', 'a') f.write(nmeaString) f.close() saveAllIncomingData(nmeaString) #os.system("cd /Users/michaelhamilton/gpsbabel && ./gpsbabel -i nmea -f nmeaRawData.doc -o kml,deficon=http://encs.vancouver.wsu.edu/~mikehmbn/balloon-icon-map.png,line_color=FF321E98,floating=1 -F cougballoon.kml") os.system("cd /Users/michaelhamilton/gpsbabel && ./gpsbabel -i nmea -f nmeaRawData.txt -o kml,deficon=http://encs.vancouver.wsu.edu/~mikehmbn/balloon-icon-map.png,line_color=FF321E98,floating=1 -F cougballoon.kml") os.system("cd /Users/michaelhamilton/gpsbabel && scp cougballoon.kml [email protected]:Sites/") print "Updated KML file was sent to the server" return #Get JSON data and send it to the server def parseToJson(RMClongitude, RMClatitude, GGAaltitude, RMCspeed, RMCheading, RMCday, RMCmonth, RMCyear, RMChours, RMCminutes, RMCseconds, extTemp, intTemp, vidTemp, COlevel, CH4level, HackStatus): JSONdata2 = { 'cougballoon':[ { 'Longitude':RMClongitude, 'Latitude':RMClatitude, 'Altitude':GGAaltitude, 'Speed':RMCspeed, 'Heading':RMCheading, 'Time':{'Day':RMCday, 'Month':RMCmonth, 'Year':RMCyear, 'Hours':RMChours, 'Minutes':RMCminutes, 'Seconds':RMCseconds},'External temperature(deg F)':extTemp, 'Internal temperature(deg F)':intTemp, 'Video Transmitter temperature(deg F)':vidTemp, 'Carbon Monoxide level(ppm)':COlevel, 'Methane level(ppm)':CH4level, 'HackHD':HackStatus } ] } data_string2 = json.dumps(JSONdata2) #Now post it to json_data.json for the map legend f = open('/Users/michaelhamilton/Desktop/json_data.json', 'w') f.write(data_string2) f.close() os.system("scp /Users/michaelhamilton/Desktop/json_data.json [email protected]:Sites/") #Now to handle it for json_data.html JSONdata = [ { 'Longitude':RMClongitude, 'Latitude':RMClatitude, 'Altitude':GGAaltitude, 'Speed(mph)':RMCspeed, 'Heading':RMCheading, 'Time':{'Day':RMCday, 'Month':RMCmonth, 'Year':RMCyear, 'Hours':RMChours, 'Minutes':RMCminutes, 'Seconds':RMCseconds},'External temperature(deg F)':extTemp, 'Internal temperature(deg F)':intTemp, 'Video Transmitter temperature(deg F)':vidTemp, 'Carbon Monoxide level(ppm)':COlevel, 'Methane level(ppm)':CH4level, 'HackHD camera statuses':HackStatus } ] data_string = json.dumps(JSONdata) JSONdataString = str(data_string) newJSONdata = re.match(r'\[(.*)', JSONdataString) newJSONdata2 = "," + newJSONdata.group(1) f = open('/Users/michaelhamilton/Desktop/json_data.html', 'r+') jumpBack = -1 #jump back 1 spot from the end f.seek(jumpBack, 2) #2 is the end of the file last = f.readline() #read the last line while (last != "]"): #if it's not a ], then keep jumping back jumpBack = jumpBack - 1 #decrement if (last == "]"): f.seek(-1, 2) f.write(newJSONdata2) f.close() #Send it to the server os.system("cd /Users/michaelhamilton/Desktop && scp json_data.html [email protected]:Sites/") print "Updated JSON information was sent to the server." return #Parse out the data from an RMC nmea string def RegExprNMEAdataRMC(line): #if it's an RMC string.... print line newRMCline = re.match( r'\$GPRMC,(\d\d)(\d\d)(\d\d).*,\D,(\d+.\d+),\D,(\d+.\d+),\D,(\d+.\d+),(\d+.\d+),(\d\d)(\d\d)(\d\d),.*,.*', line, re.I) #All data are strings, not integers if (newRMCline): global RMChours RMChours = newRMCline.group(1) #Convert UTC hours to PST(Daylight Savings Time) RMChours = UTCtoPSTDST(RMChours) global RMCminutes RMCminutes = newRMCline.group(2) global RMCseconds RMCseconds = newRMCline.group(3) global RMClatitude RMClatitude = newRMCline.group(4) RMClatitude = StringToFloatGPS(RMClatitude) global RMClongitude RMClongitude = newRMCline.group(5) RMClongitude = StringToFloatGPS(RMClongitude) global RMCspeed RMCspeed = newRMCline.group(6) RMCspeed = StringToFloatGPS(RMCspeed) global RMCheading RMCheading = newRMCline.group(7) RMCheading = StringToFloatGPS(RMCheading) global RMCday RMCday = newRMCline.group(8) global RMCmonth RMCmonth = newRMCline.group(9) global RMCyear RMCyear = newRMCline.group(10) return True else: return False #Parse out the data from an GGA nmea string def RegExprNMEAdataGGA(line): #if it's a GGA string.... print line newGGAline = re.match( r'\$GPGGA,(\d\d)(\d\d)(\d\d).*,(.*..*),\D,(.*..*),\D,\d,\d\d\,\d.\d\d,(\d+.\d),\D.*', line, re.I) #All data are strings, not integers if (newGGAline): global GGAhours GGAhours = newGGAline.group(1) #Convert UTC hours to PST(Daylight Savings Time) GGAhours = UTCtoPSTDST(GGAhours) global GGAminutes GGAminutes = newGGAline.group(2) global GGAseconds GGAseconds = newGGAline.group(3) global GGAlatitude GGAlatitude = newGGAline.group(4) GGAlatitude = StringToFloatGPS(GGAlatitude) global GGAlongitude
hours = hours.rstrip('\n'); hours = hours.rstrip('\r'); hours = int(hours) + 17 if (hours > 24): hours = hours - 24 hours = str(hours) return hours
identifier_body
cougballoon_rcv.py
mode='lines+markers', stream=stream2 ) trace3 = Scatter( x=[], y=[], mode='lines+markers', stream=stream3 ) trace4 = Scatter( x=[], y=[], mode='lines+markers', stream=stream4 ) trace5 = Scatter( x=[], y=[], mode='lines+markers', stream=stream5 ) trace6 = Scatter( x=[], y=[], mode='lines+markers', stream=stream6 ) trace7 = Scatter( x=[], y=[], mode='lines+markers', stream=stream7 ) trace9 = Scatter( x=[], y=[], mode='lines+markers', stream=stream9 ) trace10 = Scatter( x=[], y=[], mode='lines+markers', stream=stream10 ) #trace11 = Scatter( # x=[], # y=[], # mode='lines+markers', # stream=stream11 #) #trace12 = Scatter( # x=[], # y=[], # mode='lines+markers', # stream=stream12 #) trace13 = Scatter( x=[], y=[], mode='lines+markers', stream=stream13 ) #Set up the plotly graphs data_graph_a = Data([trace1, trace3, trace5]) data_graph_b = Data([trace6, trace7]) data_graph_c = Data([trace9, trace10]) #data_graph_d = Data([trace2, trace4])#Does not work data_graph_e = Data([trace2, trace4]) layout_a = Layout(title='#cougballoon temperatures')#This is the name on the graph layout_b = Layout(title='#cougballoon air quality levels')#This is the name on the graph layout_c = Layout(title='#cougballoon payload pitch and roll data')#This is the name on the graph #layout_d = Layout(title='#cougballoon altitude')#This is the name on the graph layout_e = Layout(title='#cougballoon altitude')#This is the name on the graph fig_a = Figure(data=data_graph_a, layout=layout_a) fig_b = Figure(data=data_graph_b, layout=layout_b) fig_c = Figure(data=data_graph_c, layout=layout_c) #fig_d = Figure(data=data_graph_d, layout=layout_d) fig_e = Figure(data=data_graph_e, layout=layout_e) unique_url_a = py.plot(fig_a, filename='cougballoon1', fileopt='extend')#Name above the graph unique_url_b = py.plot(fig_b, filename='cougballoon2', fileopt='extend')#Name above the graph unique_url_c = py.plot(fig_c, filename='cougballoon3', fileopt='extend')#Name above the graph #unique_url_d = py.plot(fig_d, filename='cougballoon4', fileopt='extend')#Name above the graph unique_url_e = py.plot(fig_e, filename='cougballoon5', fileopt='extend')#Name above the graph #Print the plotly urls print unique_url_a print unique_url_b print unique_url_c #print unique_url_d print unique_url_e #Get the plotly streams ready s1 = py.Stream(stream_id1) s2 = py.Stream(stream_id2) s3 = py.Stream(stream_id3) s4 = py.Stream(stream_id4) s5 = py.Stream(stream_id5) s6 = py.Stream(stream_id6) s7 = py.Stream(stream_id7) s9 = py.Stream(stream_id9) s10 = py.Stream(stream_id10) #s11 = py.Stream(stream_id11) #s12 = py.Stream(stream_id12) #s13 = py.Stream(stream_id13) #Open the plotly streams s1.open() s2.open() s3.open() s4.open() s5.open() s6.open() s7.open() s9.open() s10.open() #s11.open() #s12.open() #s13.open() import datetime import time # Delay start of stream by 5 sec (time to switch tabs) time.sleep(5) #Clean out the buffers line = ser.readline() time.sleep(2) line = ser.readline() time.sleep(2) line = ser.readline() time.sleep(2) while True: # Current time on x-axis, values on y-axis x = datetime.datetime.now().strftime('%Y-%m-%d %H:%M:%S.%f') line = ser.readline() #properly captures incoming string #External temperature #A if ((line.find("A")) == 0): print "External temperature:" print line y = StringToFloat(line, extTemp) saveData(line) extTemp = y print y s1.write(dict(x=x, y=y)) #External pressure #B elif ((line.find("B")) == 0): print "External Pressure:" print line y = StringToFloat(line) saveData(line) print y #s2.write(dict(x=x, y=y)) #Internal temperature #C elif ((line.find("C")) == 0): print "Internal temperature:" print line y = StringToFloat(line, intTemp) saveData(line) intTemp = y print y s3.write(dict(x=x, y=y)) #Internal pressure #D #elif ((line.find("D")) == 0): #print "Internal pressure:" #print line #y = StringToFloat(line) #saveData(line) #print y #s4.write(dict(x=x, y=y)) #Videolynx temperature #E elif ((line.find("E")) == 0): print "Videolynx temperature:" print line y = StringToFloat(line) saveData(line) vidTemp = y print y #s5.write(dict(x=x, y=y)) #CO level in ppm #F elif ((line.find("F")) == 0): print "CO level (in ppm):" print line y = StringToFloat(line, COlevel) saveData(line) COlevel = y print y s6.write(dict(x=x, y=y)) #CH4 level in ppm #G elif ((line.find("G")) == 0): print "CH4 level (in ppm):" print line y = StringToFloat(line, CH4level) saveData(line) CH4level = y print y s7.write(dict(x=x, y=y)) #Humidity #J elif ((line.find("J")) == 0): print "Humidity:" print line y = StringToFloat(line) saveData(line) print y #What data do we want here? elif ((line.find("K")) == 0): print "FOUND A K!" print line y = StringToFloat(line) saveData(line) print y #What data do we want here? elif ((line.find("L")) == 0): print "Roll:" print line y = StringToFloat(line, roll) saveData(line) roll = y print y s9.write(dict(x=x, y=y)) #HACKHD INFO BELOW elif ((line.find("Hack")) == 0): print "HackHD information" print line saveData(line) HackStatus = line HackStatus = HackStatus[6:13] HackStatus = HackStatus.rstrip('\n'); HackStatus = HackStatus.rstrip('\r'); print HackStatus #What data do we want here? elif ((line.find("P")) == 0): print "Pitch:" print line y = StringToFloat(line, pitch) saveData(line) pitch = y print y s10.write(dict(x=x, y=y)) #What data do we want here? elif ((line.find("Q")) == 0): print "Heading:" print line y = StringToFloat(line, heading) saveData(line) heading = y print y #s11.write(dict(x=x, y=y)) #What data do we want here? elif ((line.find("T")) == 0): print "Pressure" print line y = StringToFloat(line, pressure) saveData(line) pressure = y print y #What data do we want here? elif ((line.find("U")) == 0):
print "Altitude(from press/temp):" print line y = StringToFloat(line, pressureAltitude) saveData(line) pressureAltitude = y print y s4.write(dict(x=x, y=y))
conditional_block
options.go
option() } // Options is a list of Option values that also satisfies the Option interface. // Helper comparison packages may return an Options value when packing multiple // Option values into a single Option. When this package processes an Options, // it will be implicitly expanded into a flat list. // // Applying a filter on an Options is equivalent to applying that same filter // on all individual options held within. type Options []Option func (Options) option() {} type ( pathFilter func(Path) bool valueFilter struct { in reflect.Type // T fnc reflect.Value // func(T, T) bool } ) type option struct { typeFilter reflect.Type pathFilters []pathFilter valueFilters []valueFilter // op is the operation to perform. If nil, then this acts as an ignore. op interface{} // nil | *transformer | *comparer } func (option)
() {} func (o option) String() string { // TODO: Add information about the caller? // TODO: Maintain the order that filters were added? var ss []string switch op := o.op.(type) { case *transformer: fn := getFuncName(op.fnc.Pointer()) ss = append(ss, fmt.Sprintf("Transformer(%s, %s)", op.name, fn)) case *comparer: fn := getFuncName(op.fnc.Pointer()) ss = append(ss, fmt.Sprintf("Comparer(%s)", fn)) default: ss = append(ss, "Ignore()") } for _, f := range o.pathFilters { fn := getFuncName(reflect.ValueOf(f).Pointer()) ss = append(ss, fmt.Sprintf("FilterPath(%s)", fn)) } for _, f := range o.valueFilters { fn := getFuncName(f.fnc.Pointer()) ss = append(ss, fmt.Sprintf("FilterValues(%s)", fn)) } return strings.Join(ss, "\n\t") } // getFuncName returns a short function name from the pointer. // The string parsing logic works up until Go1.9. func getFuncName(p uintptr) string { fnc := runtime.FuncForPC(p) if fnc == nil { return "<unknown>" } name := fnc.Name() // E.g., "long/path/name/mypkg.(mytype).(long/path/name/mypkg.myfunc)-fm" if strings.HasSuffix(name, ")-fm") || strings.HasSuffix(name, ")·fm") { // Strip the package name from method name. name = strings.TrimSuffix(name, ")-fm") name = strings.TrimSuffix(name, ")·fm") if i := strings.LastIndexByte(name, '('); i >= 0 { methodName := name[i+1:] // E.g., "long/path/name/mypkg.myfunc" if j := strings.LastIndexByte(methodName, '.'); j >= 0 { methodName = methodName[j+1:] // E.g., "myfunc" } name = name[:i] + methodName // E.g., "long/path/name/mypkg.(mytype)." + "myfunc" } } if i := strings.LastIndexByte(name, '/'); i >= 0 { // Strip the package name. name = name[i+1:] // E.g., "mypkg.(mytype).myfunc" } return name } // FilterPath returns a new Option where opt is only evaluated if filter f // returns true for the current Path in the value tree. // // The option passed in may be an Ignore, Transformer, Comparer, Options, or // a previously filtered Option. func FilterPath(f func(Path) bool, opt Option) Option { if f == nil { panic("invalid path filter function") } switch opt := opt.(type) { case Options: var opts []Option for _, o := range opt { opts = append(opts, FilterPath(f, o)) // Append to slice copy } return Options(opts) case option: n := len(opt.pathFilters) opt.pathFilters = append(opt.pathFilters[:n:n], f) // Append to copy return opt default: panic(fmt.Sprintf("unknown option type: %T", opt)) } } // FilterValues returns a new Option where opt is only evaluated if filter f, // which is a function of the form "func(T, T) bool", returns true for the // current pair of values being compared. If the type of the values is not // assignable to T, then this filter implicitly returns false. // // The filter function must be // symmetric (i.e., agnostic to the order of the inputs) and // deterministic (i.e., produces the same result when given the same inputs). // If T is an interface, it is possible that f is called with two values with // different concrete types that both implement T. // // The option passed in may be an Ignore, Transformer, Comparer, Options, or // a previously filtered Option. func FilterValues(f interface{}, opt Option) Option { v := reflect.ValueOf(f) if functionType(v.Type()) != valueFilterFunc || v.IsNil() { panic(fmt.Sprintf("invalid values filter function: %T", f)) } switch opt := opt.(type) { case Options: var opts []Option for _, o := range opt { opts = append(opts, FilterValues(f, o)) // Append to slice copy } return Options(opts) case option: n := len(opt.valueFilters) vf := valueFilter{v.Type().In(0), v} opt.valueFilters = append(opt.valueFilters[:n:n], vf) // Append to copy return opt default: panic(fmt.Sprintf("unknown option type: %T", opt)) } } // Ignore is an Option that causes all comparisons to be ignored. // This value is intended to be combined with FilterPath or FilterValues. // It is an error to pass an unfiltered Ignore option to Equal. func Ignore() Option { return option{} } // Transformer returns an Option that applies a transformation function that // converts values of a certain type into that of another. // // The transformer f must be a function "func(T) R" that converts values of // type T to those of type R and is implicitly filtered to input values // assignable to T. The transformer must not mutate T in any way. // If T and R are the same type, an additional filter must be applied to // act as the base case to prevent an infinite recursion applying the same // transform to itself (see the SortedSlice example). // // The name is a user provided label that is used as the Transform.Name in the // transformation PathStep. If empty, an arbitrary name is used. func Transformer(name string, f interface{}) Option { v := reflect.ValueOf(f) if functionType(v.Type()) != transformFunc || v.IsNil() { panic(fmt.Sprintf("invalid transformer function: %T", f)) } if name == "" { name = "λ" // Lambda-symbol as place-holder for anonymous transformer } if !isValid(name) { panic(fmt.Sprintf("invalid name: %q", name)) } opt := option{op: &transformer{name, reflect.ValueOf(f)}} if ti := v.Type().In(0); ti.Kind() != reflect.Interface || ti.NumMethod() > 0 { opt.typeFilter = ti } return opt } type transformer struct { name string fnc reflect.Value // func(T) R } // Comparer returns an Option that determines whether two values are equal // to each other. // // The comparer f must be a function "func(T, T) bool" and is implicitly // filtered to input values assignable to T. If T is an interface, it is // possible that f is called with two values of different concrete types that // both implement T. // // The equality function must be: // • Symmetric: equal(x, y) == equal(y, x) // • Deterministic: equal(x, y) == equal(x, y) // • Pure: equal(x, y) does not modify x or y func Comparer(f interface{}) Option { v := reflect.ValueOf(f) if functionType(v.Type()) != equalFunc || v.IsNil() { panic(fmt.Sprintf("invalid comparer function: %T", f)) } opt := option{op: &comparer{v}} if ti := v.Type().In(0); ti.Kind() != reflect.Interface || ti.NumMethod() > 0 { opt.typeFilter = ti } return opt } type comparer struct { fnc reflect.Value // func(T, T) bool } // AllowUnexported returns an Option that forcibly allows operations on // unexported fields in certain structs, which are specified by passing in a // value of each struct type. // // Users of this option must understand that comparing on unexported fields // from external packages is not safe since changes in the internal // implementation of some external package may cause the result of Equal // to unexpectedly change. However, it may be valid to use this option on types // defined in an internal package where the semantic meaning of an unexported // field is in the control of the user. // // For some cases, a custom Comparer should be used instead that defines // equality as a function of the public API of a type rather than the underlying // unexported implementation. // // For example, the reflect.Type documentation defines equality to be determined // by the == operator on the
option
identifier_name
options.go
option() } // Options is a list of Option values that also satisfies the Option interface. // Helper comparison packages may return an Options value when packing multiple // Option values into a single Option. When this package processes an Options, // it will be implicitly expanded into a flat list. // // Applying a filter on an Options is equivalent to applying that same filter // on all individual options held within. type Options []Option func (Options) option() {} type ( pathFilter func(Path) bool valueFilter struct { in reflect.Type // T fnc reflect.Value // func(T, T) bool } ) type option struct { typeFilter reflect.Type pathFilters []pathFilter valueFilters []valueFilter // op is the operation to perform. If nil, then this acts as an ignore. op interface{} // nil | *transformer | *comparer } func (option) option() {} func (o option) String() string { // TODO: Add information about the caller? // TODO: Maintain the order that filters were added? var ss []string switch op := o.op.(type) { case *transformer: fn := getFuncName(op.fnc.Pointer()) ss = append(ss, fmt.Sprintf("Transformer(%s, %s)", op.name, fn)) case *comparer: fn := getFuncName(op.fnc.Pointer()) ss = append(ss, fmt.Sprintf("Comparer(%s)", fn)) default: ss = append(ss, "Ignore()") } for _, f := range o.pathFilters { fn := getFuncName(reflect.ValueOf(f).Pointer()) ss = append(ss, fmt.Sprintf("FilterPath(%s)", fn)) } for _, f := range o.valueFilters { fn := getFuncName(f.fnc.Pointer()) ss = append(ss, fmt.Sprintf("FilterValues(%s)", fn)) } return strings.Join(ss, "\n\t") } // getFuncName returns a short function name from the pointer. // The string parsing logic works up until Go1.9. func getFuncName(p uintptr) string
name = name[i+1:] // E.g., "mypkg.(mytype).myfunc" } return name } // FilterPath returns a new Option where opt is only evaluated if filter f // returns true for the current Path in the value tree. // // The option passed in may be an Ignore, Transformer, Comparer, Options, or // a previously filtered Option. func FilterPath(f func(Path) bool, opt Option) Option { if f == nil { panic("invalid path filter function") } switch opt := opt.(type) { case Options: var opts []Option for _, o := range opt { opts = append(opts, FilterPath(f, o)) // Append to slice copy } return Options(opts) case option: n := len(opt.pathFilters) opt.pathFilters = append(opt.pathFilters[:n:n], f) // Append to copy return opt default: panic(fmt.Sprintf("unknown option type: %T", opt)) } } // FilterValues returns a new Option where opt is only evaluated if filter f, // which is a function of the form "func(T, T) bool", returns true for the // current pair of values being compared. If the type of the values is not // assignable to T, then this filter implicitly returns false. // // The filter function must be // symmetric (i.e., agnostic to the order of the inputs) and // deterministic (i.e., produces the same result when given the same inputs). // If T is an interface, it is possible that f is called with two values with // different concrete types that both implement T. // // The option passed in may be an Ignore, Transformer, Comparer, Options, or // a previously filtered Option. func FilterValues(f interface{}, opt Option) Option { v := reflect.ValueOf(f) if functionType(v.Type()) != valueFilterFunc || v.IsNil() { panic(fmt.Sprintf("invalid values filter function: %T", f)) } switch opt := opt.(type) { case Options: var opts []Option for _, o := range opt { opts = append(opts, FilterValues(f, o)) // Append to slice copy } return Options(opts) case option: n := len(opt.valueFilters) vf := valueFilter{v.Type().In(0), v} opt.valueFilters = append(opt.valueFilters[:n:n], vf) // Append to copy return opt default: panic(fmt.Sprintf("unknown option type: %T", opt)) } } // Ignore is an Option that causes all comparisons to be ignored. // This value is intended to be combined with FilterPath or FilterValues. // It is an error to pass an unfiltered Ignore option to Equal. func Ignore() Option { return option{} } // Transformer returns an Option that applies a transformation function that // converts values of a certain type into that of another. // // The transformer f must be a function "func(T) R" that converts values of // type T to those of type R and is implicitly filtered to input values // assignable to T. The transformer must not mutate T in any way. // If T and R are the same type, an additional filter must be applied to // act as the base case to prevent an infinite recursion applying the same // transform to itself (see the SortedSlice example). // // The name is a user provided label that is used as the Transform.Name in the // transformation PathStep. If empty, an arbitrary name is used. func Transformer(name string, f interface{}) Option { v := reflect.ValueOf(f) if functionType(v.Type()) != transformFunc || v.IsNil() { panic(fmt.Sprintf("invalid transformer function: %T", f)) } if name == "" { name = "λ" // Lambda-symbol as place-holder for anonymous transformer } if !isValid(name) { panic(fmt.Sprintf("invalid name: %q", name)) } opt := option{op: &transformer{name, reflect.ValueOf(f)}} if ti := v.Type().In(0); ti.Kind() != reflect.Interface || ti.NumMethod() > 0 { opt.typeFilter = ti } return opt } type transformer struct { name string fnc reflect.Value // func(T) R } // Comparer returns an Option that determines whether two values are equal // to each other. // // The comparer f must be a function "func(T, T) bool" and is implicitly // filtered to input values assignable to T. If T is an interface, it is // possible that f is called with two values of different concrete types that // both implement T. // // The equality function must be: // • Symmetric: equal(x, y) == equal(y, x) // • Deterministic: equal(x, y) == equal(x, y) // • Pure: equal(x, y) does not modify x or y func Comparer(f interface{}) Option { v := reflect.ValueOf(f) if functionType(v.Type()) != equalFunc || v.IsNil() { panic(fmt.Sprintf("invalid comparer function: %T", f)) } opt := option{op: &comparer{v}} if ti := v.Type().In(0); ti.Kind() != reflect.Interface || ti.NumMethod() > 0 { opt.typeFilter = ti } return opt } type comparer struct { fnc reflect.Value // func(T, T) bool } // AllowUnexported returns an Option that forcibly allows operations on // unexported fields in certain structs, which are specified by passing in a // value of each struct type. // // Users of this option must understand that comparing on unexported fields // from external packages is not safe since changes in the internal // implementation of some external package may cause the result of Equal // to unexpectedly change. However, it may be valid to use this option on types // defined in an internal package where the semantic meaning of an unexported // field is in the control of the user. // // For some cases, a custom Comparer should be used instead that defines // equality as a function of the public API of a type rather than the underlying // unexported implementation. // // For example, the reflect.Type documentation defines equality to be determined // by the == operator on the
{ fnc := runtime.FuncForPC(p) if fnc == nil { return "<unknown>" } name := fnc.Name() // E.g., "long/path/name/mypkg.(mytype).(long/path/name/mypkg.myfunc)-fm" if strings.HasSuffix(name, ")-fm") || strings.HasSuffix(name, ")·fm") { // Strip the package name from method name. name = strings.TrimSuffix(name, ")-fm") name = strings.TrimSuffix(name, ")·fm") if i := strings.LastIndexByte(name, '('); i >= 0 { methodName := name[i+1:] // E.g., "long/path/name/mypkg.myfunc" if j := strings.LastIndexByte(methodName, '.'); j >= 0 { methodName = methodName[j+1:] // E.g., "myfunc" } name = name[:i] + methodName // E.g., "long/path/name/mypkg.(mytype)." + "myfunc" } } if i := strings.LastIndexByte(name, '/'); i >= 0 { // Strip the package name.
identifier_body
options.go
"runtime" "strings" ) // Option configures for specific behavior of Equal and Diff. In particular, // the fundamental Option functions (Ignore, Transformer, and Comparer), // configure how equality is determined. // // The fundamental options may be composed with filters (FilterPath and // FilterValues) to control the scope over which they are applied. // // The cmp/cmpopts package provides helper functions for creating options that // may be used with Equal and Diff. type Option interface { // Prevent Option from being equivalent to interface{}, which provides // a small type checking benefit by preventing Equal(opt, x, y). option() } // Options is a list of Option values that also satisfies the Option interface. // Helper comparison packages may return an Options value when packing multiple // Option values into a single Option. When this package processes an Options, // it will be implicitly expanded into a flat list. // // Applying a filter on an Options is equivalent to applying that same filter // on all individual options held within. type Options []Option func (Options) option() {} type ( pathFilter func(Path) bool valueFilter struct { in reflect.Type // T fnc reflect.Value // func(T, T) bool } ) type option struct { typeFilter reflect.Type pathFilters []pathFilter valueFilters []valueFilter // op is the operation to perform. If nil, then this acts as an ignore. op interface{} // nil | *transformer | *comparer } func (option) option() {} func (o option) String() string { // TODO: Add information about the caller? // TODO: Maintain the order that filters were added? var ss []string switch op := o.op.(type) { case *transformer: fn := getFuncName(op.fnc.Pointer()) ss = append(ss, fmt.Sprintf("Transformer(%s, %s)", op.name, fn)) case *comparer: fn := getFuncName(op.fnc.Pointer()) ss = append(ss, fmt.Sprintf("Comparer(%s)", fn)) default: ss = append(ss, "Ignore()") } for _, f := range o.pathFilters { fn := getFuncName(reflect.ValueOf(f).Pointer()) ss = append(ss, fmt.Sprintf("FilterPath(%s)", fn)) } for _, f := range o.valueFilters { fn := getFuncName(f.fnc.Pointer()) ss = append(ss, fmt.Sprintf("FilterValues(%s)", fn)) } return strings.Join(ss, "\n\t") } // getFuncName returns a short function name from the pointer. // The string parsing logic works up until Go1.9. func getFuncName(p uintptr) string { fnc := runtime.FuncForPC(p) if fnc == nil { return "<unknown>" } name := fnc.Name() // E.g., "long/path/name/mypkg.(mytype).(long/path/name/mypkg.myfunc)-fm" if strings.HasSuffix(name, ")-fm") || strings.HasSuffix(name, ")·fm") { // Strip the package name from method name. name = strings.TrimSuffix(name, ")-fm") name = strings.TrimSuffix(name, ")·fm") if i := strings.LastIndexByte(name, '('); i >= 0 { methodName := name[i+1:] // E.g., "long/path/name/mypkg.myfunc" if j := strings.LastIndexByte(methodName, '.'); j >= 0 { methodName = methodName[j+1:] // E.g., "myfunc" } name = name[:i] + methodName // E.g., "long/path/name/mypkg.(mytype)." + "myfunc" } } if i := strings.LastIndexByte(name, '/'); i >= 0 { // Strip the package name. name = name[i+1:] // E.g., "mypkg.(mytype).myfunc" } return name } // FilterPath returns a new Option where opt is only evaluated if filter f // returns true for the current Path in the value tree. // // The option passed in may be an Ignore, Transformer, Comparer, Options, or // a previously filtered Option. func FilterPath(f func(Path) bool, opt Option) Option { if f == nil { panic("invalid path filter function") } switch opt := opt.(type) { case Options: var opts []Option for _, o := range opt { opts = append(opts, FilterPath(f, o)) // Append to slice copy } return Options(opts) case option: n := len(opt.pathFilters) opt.pathFilters = append(opt.pathFilters[:n:n], f) // Append to copy return opt default: panic(fmt.Sprintf("unknown option type: %T", opt)) } } // FilterValues returns a new Option where opt is only evaluated if filter f, // which is a function of the form "func(T, T) bool", returns true for the // current pair of values being compared. If the type of the values is not // assignable to T, then this filter implicitly returns false. // // The filter function must be // symmetric (i.e., agnostic to the order of the inputs) and // deterministic (i.e., produces the same result when given the same inputs). // If T is an interface, it is possible that f is called with two values with // different concrete types that both implement T. // // The option passed in may be an Ignore, Transformer, Comparer, Options, or // a previously filtered Option. func FilterValues(f interface{}, opt Option) Option { v := reflect.ValueOf(f) if functionType(v.Type()) != valueFilterFunc || v.IsNil() { panic(fmt.Sprintf("invalid values filter function: %T", f)) } switch opt := opt.(type) { case Options: var opts []Option for _, o := range opt { opts = append(opts, FilterValues(f, o)) // Append to slice copy } return Options(opts) case option: n := len(opt.valueFilters) vf := valueFilter{v.Type().In(0), v} opt.valueFilters = append(opt.valueFilters[:n:n], vf) // Append to copy return opt default: panic(fmt.Sprintf("unknown option type: %T", opt)) } } // Ignore is an Option that causes all comparisons to be ignored. // This value is intended to be combined with FilterPath or FilterValues. // It is an error to pass an unfiltered Ignore option to Equal. func Ignore() Option { return option{} } // Transformer returns an Option that applies a transformation function that // converts values of a certain type into that of another. // // The transformer f must be a function "func(T) R" that converts values of // type T to those of type R and is implicitly filtered to input values // assignable to T. The transformer must not mutate T in any way. // If T and R are the same type, an additional filter must be applied to // act as the base case to prevent an infinite recursion applying the same // transform to itself (see the SortedSlice example). // // The name is a user provided label that is used as the Transform.Name in the // transformation PathStep. If empty, an arbitrary name is used. func Transformer(name string, f interface{}) Option { v := reflect.ValueOf(f) if functionType(v.Type()) != transformFunc || v.IsNil() { panic(fmt.Sprintf("invalid transformer function: %T", f)) } if name == "" { name = "λ" // Lambda-symbol as place-holder for anonymous transformer } if !isValid(name) { panic(fmt.Sprintf("invalid name: %q", name)) } opt := option{op: &transformer{name, reflect.ValueOf(f)}} if ti := v.Type().In(0); ti.Kind() != reflect.Interface || ti.NumMethod() > 0 { opt.typeFilter = ti } return opt } type transformer struct { name string fnc reflect.Value // func(T) R } // Comparer returns an Option that determines whether two values are equal // to each other. // // The comparer f must be a function "func(T, T) bool" and is implicitly // filtered to input values assignable to T. If T is an interface, it is // possible that f is called with two values of different concrete types that // both implement T. // // The equality function must be: // • Symmetric: equal(x, y) == equal(y, x) // • Deterministic: equal(x, y) == equal(x, y) // • Pure: equal(x, y) does not modify x or y func Comparer(f interface{}) Option { v := reflect.ValueOf(f) if functionType(v.Type()) != equalFunc || v.IsNil() { panic(fmt.Sprintf("invalid comparer function: %T", f)) } opt := option{op: &comparer{v}} if ti := v.Type().In(0); ti.Kind() != reflect.Interface || ti.NumMethod() > 0 { opt.typeFilter = ti } return opt } type comparer struct { fnc reflect.Value // func(T, T) bool } // AllowUnexported returns an Option that forcibly allows operations on // unexported fields in certain structs, which are specified by passing in a // value of each struct type. // // Users of this option must understand that
import ( "fmt" "reflect"
random_line_split
options.go
option() } // Options is a list of Option values that also satisfies the Option interface. // Helper comparison packages may return an Options value when packing multiple // Option values into a single Option. When this package processes an Options, // it will be implicitly expanded into a flat list. // // Applying a filter on an Options is equivalent to applying that same filter // on all individual options held within. type Options []Option func (Options) option() {} type ( pathFilter func(Path) bool valueFilter struct { in reflect.Type // T fnc reflect.Value // func(T, T) bool } ) type option struct { typeFilter reflect.Type pathFilters []pathFilter valueFilters []valueFilter // op is the operation to perform. If nil, then this acts as an ignore. op interface{} // nil | *transformer | *comparer } func (option) option() {} func (o option) String() string { // TODO: Add information about the caller? // TODO: Maintain the order that filters were added? var ss []string switch op := o.op.(type) { case *transformer: fn := getFuncName(op.fnc.Pointer()) ss = append(ss, fmt.Sprintf("Transformer(%s, %s)", op.name, fn)) case *comparer: fn := getFuncName(op.fnc.Pointer()) ss = append(ss, fmt.Sprintf("Comparer(%s)", fn)) default: ss = append(ss, "Ignore()") } for _, f := range o.pathFilters { fn := getFuncName(reflect.ValueOf(f).Pointer()) ss = append(ss, fmt.Sprintf("FilterPath(%s)", fn)) } for _, f := range o.valueFilters { fn := getFuncName(f.fnc.Pointer()) ss = append(ss, fmt.Sprintf("FilterValues(%s)", fn)) } return strings.Join(ss, "\n\t") } // getFuncName returns a short function name from the pointer. // The string parsing logic works up until Go1.9. func getFuncName(p uintptr) string { fnc := runtime.FuncForPC(p) if fnc == nil
name := fnc.Name() // E.g., "long/path/name/mypkg.(mytype).(long/path/name/mypkg.myfunc)-fm" if strings.HasSuffix(name, ")-fm") || strings.HasSuffix(name, ")·fm") { // Strip the package name from method name. name = strings.TrimSuffix(name, ")-fm") name = strings.TrimSuffix(name, ")·fm") if i := strings.LastIndexByte(name, '('); i >= 0 { methodName := name[i+1:] // E.g., "long/path/name/mypkg.myfunc" if j := strings.LastIndexByte(methodName, '.'); j >= 0 { methodName = methodName[j+1:] // E.g., "myfunc" } name = name[:i] + methodName // E.g., "long/path/name/mypkg.(mytype)." + "myfunc" } } if i := strings.LastIndexByte(name, '/'); i >= 0 { // Strip the package name. name = name[i+1:] // E.g., "mypkg.(mytype).myfunc" } return name } // FilterPath returns a new Option where opt is only evaluated if filter f // returns true for the current Path in the value tree. // // The option passed in may be an Ignore, Transformer, Comparer, Options, or // a previously filtered Option. func FilterPath(f func(Path) bool, opt Option) Option { if f == nil { panic("invalid path filter function") } switch opt := opt.(type) { case Options: var opts []Option for _, o := range opt { opts = append(opts, FilterPath(f, o)) // Append to slice copy } return Options(opts) case option: n := len(opt.pathFilters) opt.pathFilters = append(opt.pathFilters[:n:n], f) // Append to copy return opt default: panic(fmt.Sprintf("unknown option type: %T", opt)) } } // FilterValues returns a new Option where opt is only evaluated if filter f, // which is a function of the form "func(T, T) bool", returns true for the // current pair of values being compared. If the type of the values is not // assignable to T, then this filter implicitly returns false. // // The filter function must be // symmetric (i.e., agnostic to the order of the inputs) and // deterministic (i.e., produces the same result when given the same inputs). // If T is an interface, it is possible that f is called with two values with // different concrete types that both implement T. // // The option passed in may be an Ignore, Transformer, Comparer, Options, or // a previously filtered Option. func FilterValues(f interface{}, opt Option) Option { v := reflect.ValueOf(f) if functionType(v.Type()) != valueFilterFunc || v.IsNil() { panic(fmt.Sprintf("invalid values filter function: %T", f)) } switch opt := opt.(type) { case Options: var opts []Option for _, o := range opt { opts = append(opts, FilterValues(f, o)) // Append to slice copy } return Options(opts) case option: n := len(opt.valueFilters) vf := valueFilter{v.Type().In(0), v} opt.valueFilters = append(opt.valueFilters[:n:n], vf) // Append to copy return opt default: panic(fmt.Sprintf("unknown option type: %T", opt)) } } // Ignore is an Option that causes all comparisons to be ignored. // This value is intended to be combined with FilterPath or FilterValues. // It is an error to pass an unfiltered Ignore option to Equal. func Ignore() Option { return option{} } // Transformer returns an Option that applies a transformation function that // converts values of a certain type into that of another. // // The transformer f must be a function "func(T) R" that converts values of // type T to those of type R and is implicitly filtered to input values // assignable to T. The transformer must not mutate T in any way. // If T and R are the same type, an additional filter must be applied to // act as the base case to prevent an infinite recursion applying the same // transform to itself (see the SortedSlice example). // // The name is a user provided label that is used as the Transform.Name in the // transformation PathStep. If empty, an arbitrary name is used. func Transformer(name string, f interface{}) Option { v := reflect.ValueOf(f) if functionType(v.Type()) != transformFunc || v.IsNil() { panic(fmt.Sprintf("invalid transformer function: %T", f)) } if name == "" { name = "λ" // Lambda-symbol as place-holder for anonymous transformer } if !isValid(name) { panic(fmt.Sprintf("invalid name: %q", name)) } opt := option{op: &transformer{name, reflect.ValueOf(f)}} if ti := v.Type().In(0); ti.Kind() != reflect.Interface || ti.NumMethod() > 0 { opt.typeFilter = ti } return opt } type transformer struct { name string fnc reflect.Value // func(T) R } // Comparer returns an Option that determines whether two values are equal // to each other. // // The comparer f must be a function "func(T, T) bool" and is implicitly // filtered to input values assignable to T. If T is an interface, it is // possible that f is called with two values of different concrete types that // both implement T. // // The equality function must be: // • Symmetric: equal(x, y) == equal(y, x) // • Deterministic: equal(x, y) == equal(x, y) // • Pure: equal(x, y) does not modify x or y func Comparer(f interface{}) Option { v := reflect.ValueOf(f) if functionType(v.Type()) != equalFunc || v.IsNil() { panic(fmt.Sprintf("invalid comparer function: %T", f)) } opt := option{op: &comparer{v}} if ti := v.Type().In(0); ti.Kind() != reflect.Interface || ti.NumMethod() > 0 { opt.typeFilter = ti } return opt } type comparer struct { fnc reflect.Value // func(T, T) bool } // AllowUnexported returns an Option that forcibly allows operations on // unexported fields in certain structs, which are specified by passing in a // value of each struct type. // // Users of this option must understand that comparing on unexported fields // from external packages is not safe since changes in the internal // implementation of some external package may cause the result of Equal // to unexpectedly change. However, it may be valid to use this option on types // defined in an internal package where the semantic meaning of an unexported // field is in the control of the user. // // For some cases, a custom Comparer should be used instead that defines // equality as a function of the public API of a type rather than the underlying // unexported implementation. // // For example, the reflect.Type documentation defines equality to be determined // by the == operator on
{ return "<unknown>" }
conditional_block
combat.rs
20); NewState::Player } pub fn ranged_attack( ecs: &mut World, map: &mut Map, attacker: Entity, victim: Entity, ranged_power: i32, ) { let mut attacker_pos = None; let mut victim_pos = None; // Find positions for the start and end if let Ok(ae) = ecs.entry_ref(attacker) { if let Ok(pos) = ae.get_component::<Position>() { attacker_pos = Some(pos.clone()); } } if let Ok(ae) = ecs.entry_ref(victim) { if let Ok(pos) = ae.get_component::<Position>() { victim_pos = Some(pos.clone()); } } if attacker_pos.is_none() || victim_pos.is_none() { return; } let attacker_pos = attacker_pos.unwrap(); let victim_pos = victim_pos.unwrap(); // Set state for the projectile path let mut power = ranged_power; let mut range = 0; let mut projectile_path = Vec::new(); let mut splatter = None; let mut commands = CommandBuffer::new(ecs); let current_layer = attacker_pos.layer; // Map of entity locations. Rebuilt every time because it might change. let pos_map = <(&Position, &Health)>::query() .iter(ecs) .map(|(pos, _)| pos.pt) .collect::<HashSet<Point>>(); // Plot the initial trajectory line2d_bresenham(attacker_pos.pt, victim_pos.pt) .iter() .skip(1) .for_each(|pt| { projectile_path.push(*pt); if pos_map.contains(&pt) { power -= hit_tile_contents(ecs, *pt, current_layer, &mut commands, &mut splatter, power); if power < 1 { power = 1; range += 200; } } if let Some(bsplatter) = &mut splatter { let idx = map.get_current().point2d_to_index(*pt); map.get_current_mut().tiles[idx].color.fg = bsplatter.to_rgba(1.0); bsplatter.r = f32::max(0.0, bsplatter.r - 0.1); bsplatter.g = f32::max(0.0, bsplatter.g - 0.1); bsplatter.b = f32::max(0.0, bsplatter.b - 0.1); if bsplatter.r + bsplatter.g + bsplatter.b < 0.1 { splatter = None; } } range += 1; if range > 5 { power -= 1; if power < 1 { power = 1; } } }); // The trajectory can continue if we have power left use ultraviolet::Vec2; let mut projectile_pos: Vec2 = Vec2::new(victim_pos.pt.x as f32, victim_pos.pt.y as f32); let slope = (projectile_pos - Vec2::new(attacker_pos.pt.x as f32, attacker_pos.pt.y as f32)) .normalized(); while range < 25 && power > 0 { projectile_pos += slope; let pt = Point::new(projectile_pos.x as i32, projectile_pos.y as i32); projectile_path.push(pt); if pos_map.contains(&pt) { power -= hit_tile_contents(ecs, pt, current_layer, &mut commands, &mut splatter, power); if power < 1 { power = 1; range += 200; } } if let Some(bsplatter) = &mut splatter { let idx = map.get_current().point2d_to_index(pt); map.get_current_mut().tiles[idx].color.fg = bsplatter.to_rgba(1.0); bsplatter.r = f32::max(0.0, bsplatter.r - 0.1); bsplatter.g = f32::max(0.0, bsplatter.g - 0.1); bsplatter.b = f32::max(0.0, bsplatter.b - 0.1); if bsplatter.r + bsplatter.g + bsplatter.b < 0.1 { splatter = None; } } let idx = map.get_current().point2d_to_index(pt); if map.get_current().tiles[idx].tile_type == TileType::Wall { range += 100; power = 0; } range += 1; if range > 5 { power -= 1; if power < 1 { power = 1; range += 100; } } } commands.push(( Projectile { path: projectile_path, layer: current_layer as usize, }, Glyph { glyph: to_cp437('*'), color: ColorPair::new(RED, BLACK), }, )); commands.flush(ecs); } pub fn hit_tile_contents( ecs: &mut World, pt: Point, layer: u32, commands: &mut CommandBuffer, splatter: &mut Option<RGB>, power: i32, ) -> i32 { let mut rng_lock = crate::RNG.lock(); let rng = rng_lock.as_mut().unwrap(); let mut power_loss = 0; let mut dead_entities = Vec::new(); <(Entity, &Position, &mut Health)>::query() .iter_mut(ecs) .filter(|(_, pos, _)| pos.layer == layer && pos.pt == pt) .for_each(|(entity, _, hp)| { power_loss += hp.current; if power_loss < 0 { power_loss = 0; } let damage = i32::max(0, power + rng.roll_dice(1, 4) - 2); //println!("{}", damage); hp.current -= damage; if hp.current < 0 { hp.current = 0; dead_entities.push(*entity); } }); dead_entities.iter().for_each(|e| { if let Ok(er) = ecs.entry_ref(*e) { if let Ok(boom) = er.get_component::<Explosive>() { if let Ok(pos) = er.get_component::<Position>() { commands.push(( Position::with_pt(pos.pt, pos.layer), Boom { range: boom.range }, )); } } } }); kill_things(ecs, commands, dead_entities, splatter); power_loss } pub fn melee(ecs: &mut World, map: &mut Map, attacker: Entity, victim: Entity, melee_power: i32) { // Check range and validity let mut attacker_pos = None; let mut defender_pos = None; if let Ok(e) = ecs.entry_ref(attacker) { if let Ok(pos) = e.get_component::<Position>() { attacker_pos = Some(*pos); } } if let Ok(e) = ecs.entry_ref(victim) { if let Ok(pos) = e.get_component::<Position>() { defender_pos = Some(*pos); } } if attacker_pos.is_none() || defender_pos.is_none() { return; // Bail out - invalid data arrived } let apos = attacker_pos.unwrap(); let dpos = defender_pos.unwrap(); if apos.layer != dpos.layer { return; // Bail out - can't attack across layers } let d = DistanceAlg::Pythagoras.distance2d(apos.pt, dpos.pt); if d > 1.5 { return; // Too far away, bail } // Inflict damage upon the hapless victim let mut dead_entities = Vec::new(); if let Ok(mut v) = ecs.entry_mut(victim) { if let Ok(hp) = v.get_component_mut::<Health>() { hp.current = i32::max(0, hp.current - melee_power); if hp.current == 0 { dead_entities.push(victim); } } if let Ok(blood) = v.get_component::<Blood>() { let idx = map.get_layer(dpos.layer as usize).point2d_to_index(dpos.pt); map.get_layer_mut(dpos.layer as usize).tiles[idx].color.fg = blood.0.into(); } } // If necessary, kill them. let mut commands = CommandBuffer::new(ecs); let mut splatter = None; kill_things(ecs, &mut commands, dead_entities, &mut splatter); // Splatter blood. It's good for you. } fn
( ecs: &mut World, commands: &mut CommandBuffer, dead_entities: Vec<Entity>, splatter: &mut Option<RGB>, ) { dead_entities.iter().for_each(|entity| { crate::stats::record_death(); let mut was_decor = false; let mut was_player = false; if let Ok(mut er) = ecs.entry_mut(*entity) { let mut was_colonist = false;
kill_things
identifier_name
combat.rs
20); NewState::Player } pub fn ranged_attack( ecs: &mut World, map: &mut Map, attacker: Entity, victim: Entity, ranged_power: i32, )
// Set state for the projectile path let mut power = ranged_power; let mut range = 0; let mut projectile_path = Vec::new(); let mut splatter = None; let mut commands = CommandBuffer::new(ecs); let current_layer = attacker_pos.layer; // Map of entity locations. Rebuilt every time because it might change. let pos_map = <(&Position, &Health)>::query() .iter(ecs) .map(|(pos, _)| pos.pt) .collect::<HashSet<Point>>(); // Plot the initial trajectory line2d_bresenham(attacker_pos.pt, victim_pos.pt) .iter() .skip(1) .for_each(|pt| { projectile_path.push(*pt); if pos_map.contains(&pt) { power -= hit_tile_contents(ecs, *pt, current_layer, &mut commands, &mut splatter, power); if power < 1 { power = 1; range += 200; } } if let Some(bsplatter) = &mut splatter { let idx = map.get_current().point2d_to_index(*pt); map.get_current_mut().tiles[idx].color.fg = bsplatter.to_rgba(1.0); bsplatter.r = f32::max(0.0, bsplatter.r - 0.1); bsplatter.g = f32::max(0.0, bsplatter.g - 0.1); bsplatter.b = f32::max(0.0, bsplatter.b - 0.1); if bsplatter.r + bsplatter.g + bsplatter.b < 0.1 { splatter = None; } } range += 1; if range > 5 { power -= 1; if power < 1 { power = 1; } } }); // The trajectory can continue if we have power left use ultraviolet::Vec2; let mut projectile_pos: Vec2 = Vec2::new(victim_pos.pt.x as f32, victim_pos.pt.y as f32); let slope = (projectile_pos - Vec2::new(attacker_pos.pt.x as f32, attacker_pos.pt.y as f32)) .normalized(); while range < 25 && power > 0 { projectile_pos += slope; let pt = Point::new(projectile_pos.x as i32, projectile_pos.y as i32); projectile_path.push(pt); if pos_map.contains(&pt) { power -= hit_tile_contents(ecs, pt, current_layer, &mut commands, &mut splatter, power); if power < 1 { power = 1; range += 200; } } if let Some(bsplatter) = &mut splatter { let idx = map.get_current().point2d_to_index(pt); map.get_current_mut().tiles[idx].color.fg = bsplatter.to_rgba(1.0); bsplatter.r = f32::max(0.0, bsplatter.r - 0.1); bsplatter.g = f32::max(0.0, bsplatter.g - 0.1); bsplatter.b = f32::max(0.0, bsplatter.b - 0.1); if bsplatter.r + bsplatter.g + bsplatter.b < 0.1 { splatter = None; } } let idx = map.get_current().point2d_to_index(pt); if map.get_current().tiles[idx].tile_type == TileType::Wall { range += 100; power = 0; } range += 1; if range > 5 { power -= 1; if power < 1 { power = 1; range += 100; } } } commands.push(( Projectile { path: projectile_path, layer: current_layer as usize, }, Glyph { glyph: to_cp437('*'), color: ColorPair::new(RED, BLACK), }, )); commands.flush(ecs); } pub fn hit_tile_contents( ecs: &mut World, pt: Point, layer: u32, commands: &mut CommandBuffer, splatter: &mut Option<RGB>, power: i32, ) -> i32 { let mut rng_lock = crate::RNG.lock(); let rng = rng_lock.as_mut().unwrap(); let mut power_loss = 0; let mut dead_entities = Vec::new(); <(Entity, &Position, &mut Health)>::query() .iter_mut(ecs) .filter(|(_, pos, _)| pos.layer == layer && pos.pt == pt) .for_each(|(entity, _, hp)| { power_loss += hp.current; if power_loss < 0 { power_loss = 0; } let damage = i32::max(0, power + rng.roll_dice(1, 4) - 2); //println!("{}", damage); hp.current -= damage; if hp.current < 0 { hp.current = 0; dead_entities.push(*entity); } }); dead_entities.iter().for_each(|e| { if let Ok(er) = ecs.entry_ref(*e) { if let Ok(boom) = er.get_component::<Explosive>() { if let Ok(pos) = er.get_component::<Position>() { commands.push(( Position::with_pt(pos.pt, pos.layer), Boom { range: boom.range }, )); } } } }); kill_things(ecs, commands, dead_entities, splatter); power_loss } pub fn melee(ecs: &mut World, map: &mut Map, attacker: Entity, victim: Entity, melee_power: i32) { // Check range and validity let mut attacker_pos = None; let mut defender_pos = None; if let Ok(e) = ecs.entry_ref(attacker) { if let Ok(pos) = e.get_component::<Position>() { attacker_pos = Some(*pos); } } if let Ok(e) = ecs.entry_ref(victim) { if let Ok(pos) = e.get_component::<Position>() { defender_pos = Some(*pos); } } if attacker_pos.is_none() || defender_pos.is_none() { return; // Bail out - invalid data arrived } let apos = attacker_pos.unwrap(); let dpos = defender_pos.unwrap(); if apos.layer != dpos.layer { return; // Bail out - can't attack across layers } let d = DistanceAlg::Pythagoras.distance2d(apos.pt, dpos.pt); if d > 1.5 { return; // Too far away, bail } // Inflict damage upon the hapless victim let mut dead_entities = Vec::new(); if let Ok(mut v) = ecs.entry_mut(victim) { if let Ok(hp) = v.get_component_mut::<Health>() { hp.current = i32::max(0, hp.current - melee_power); if hp.current == 0 { dead_entities.push(victim); } } if let Ok(blood) = v.get_component::<Blood>() { let idx = map.get_layer(dpos.layer as usize).point2d_to_index(dpos.pt); map.get_layer_mut(dpos.layer as usize).tiles[idx].color.fg = blood.0.into(); } } // If necessary, kill them. let mut commands = CommandBuffer::new(ecs); let mut splatter = None; kill_things(ecs, &mut commands, dead_entities, &mut splatter); // Splatter blood. It's good for you. } fn kill_things( ecs: &mut World, commands: &mut CommandBuffer, dead_entities: Vec<Entity>, splatter: &mut Option<RGB>, ) { dead_entities.iter().for_each(|entity| { crate::stats::record_death(); let mut was_decor = false; let mut was_player = false; if let Ok(mut er) = ecs.entry_mut(*entity) { let mut was_colonist = false;
{ let mut attacker_pos = None; let mut victim_pos = None; // Find positions for the start and end if let Ok(ae) = ecs.entry_ref(attacker) { if let Ok(pos) = ae.get_component::<Position>() { attacker_pos = Some(pos.clone()); } } if let Ok(ae) = ecs.entry_ref(victim) { if let Ok(pos) = ae.get_component::<Position>() { victim_pos = Some(pos.clone()); } } if attacker_pos.is_none() || victim_pos.is_none() { return; } let attacker_pos = attacker_pos.unwrap(); let victim_pos = victim_pos.unwrap();
identifier_body
combat.rs
20); NewState::Player } pub fn ranged_attack( ecs: &mut World, map: &mut Map, attacker: Entity, victim: Entity, ranged_power: i32, ) { let mut attacker_pos = None; let mut victim_pos = None; // Find positions for the start and end if let Ok(ae) = ecs.entry_ref(attacker) { if let Ok(pos) = ae.get_component::<Position>() { attacker_pos = Some(pos.clone()); } } if let Ok(ae) = ecs.entry_ref(victim) { if let Ok(pos) = ae.get_component::<Position>() { victim_pos = Some(pos.clone()); } } if attacker_pos.is_none() || victim_pos.is_none() { return; } let attacker_pos = attacker_pos.unwrap(); let victim_pos = victim_pos.unwrap(); // Set state for the projectile path let mut power = ranged_power; let mut range = 0; let mut projectile_path = Vec::new(); let mut splatter = None; let mut commands = CommandBuffer::new(ecs); let current_layer = attacker_pos.layer; // Map of entity locations. Rebuilt every time because it might change. let pos_map = <(&Position, &Health)>::query() .iter(ecs) .map(|(pos, _)| pos.pt) .collect::<HashSet<Point>>(); // Plot the initial trajectory line2d_bresenham(attacker_pos.pt, victim_pos.pt) .iter() .skip(1) .for_each(|pt| { projectile_path.push(*pt); if pos_map.contains(&pt) { power -= hit_tile_contents(ecs, *pt, current_layer, &mut commands, &mut splatter, power); if power < 1 { power = 1; range += 200; } } if let Some(bsplatter) = &mut splatter { let idx = map.get_current().point2d_to_index(*pt); map.get_current_mut().tiles[idx].color.fg = bsplatter.to_rgba(1.0); bsplatter.r = f32::max(0.0, bsplatter.r - 0.1); bsplatter.g = f32::max(0.0, bsplatter.g - 0.1); bsplatter.b = f32::max(0.0, bsplatter.b - 0.1); if bsplatter.r + bsplatter.g + bsplatter.b < 0.1 { splatter = None; } } range += 1; if range > 5 { power -= 1; if power < 1 { power = 1; } } }); // The trajectory can continue if we have power left use ultraviolet::Vec2; let mut projectile_pos: Vec2 = Vec2::new(victim_pos.pt.x as f32, victim_pos.pt.y as f32); let slope = (projectile_pos - Vec2::new(attacker_pos.pt.x as f32, attacker_pos.pt.y as f32)) .normalized(); while range < 25 && power > 0 { projectile_pos += slope; let pt = Point::new(projectile_pos.x as i32, projectile_pos.y as i32); projectile_path.push(pt); if pos_map.contains(&pt) { power -= hit_tile_contents(ecs, pt, current_layer, &mut commands, &mut splatter, power); if power < 1 { power = 1; range += 200; } } if let Some(bsplatter) = &mut splatter { let idx = map.get_current().point2d_to_index(pt); map.get_current_mut().tiles[idx].color.fg = bsplatter.to_rgba(1.0); bsplatter.r = f32::max(0.0, bsplatter.r - 0.1); bsplatter.g = f32::max(0.0, bsplatter.g - 0.1); bsplatter.b = f32::max(0.0, bsplatter.b - 0.1); if bsplatter.r + bsplatter.g + bsplatter.b < 0.1 { splatter = None; } } let idx = map.get_current().point2d_to_index(pt); if map.get_current().tiles[idx].tile_type == TileType::Wall { range += 100; power = 0; } range += 1; if range > 5 { power -= 1; if power < 1 { power = 1; range += 100; } } } commands.push(( Projectile { path: projectile_path, layer: current_layer as usize, }, Glyph { glyph: to_cp437('*'), color: ColorPair::new(RED, BLACK), }, )); commands.flush(ecs); } pub fn hit_tile_contents( ecs: &mut World, pt: Point, layer: u32, commands: &mut CommandBuffer, splatter: &mut Option<RGB>, power: i32, ) -> i32 { let mut rng_lock = crate::RNG.lock(); let rng = rng_lock.as_mut().unwrap(); let mut power_loss = 0; let mut dead_entities = Vec::new(); <(Entity, &Position, &mut Health)>::query() .iter_mut(ecs) .filter(|(_, pos, _)| pos.layer == layer && pos.pt == pt) .for_each(|(entity, _, hp)| { power_loss += hp.current; if power_loss < 0 { power_loss = 0; } let damage = i32::max(0, power + rng.roll_dice(1, 4) - 2); //println!("{}", damage); hp.current -= damage; if hp.current < 0 { hp.current = 0; dead_entities.push(*entity); } }); dead_entities.iter().for_each(|e| { if let Ok(er) = ecs.entry_ref(*e) { if let Ok(boom) = er.get_component::<Explosive>() { if let Ok(pos) = er.get_component::<Position>() { commands.push(( Position::with_pt(pos.pt, pos.layer), Boom { range: boom.range }, )); } } } }); kill_things(ecs, commands, dead_entities, splatter); power_loss } pub fn melee(ecs: &mut World, map: &mut Map, attacker: Entity, victim: Entity, melee_power: i32) { // Check range and validity let mut attacker_pos = None; let mut defender_pos = None; if let Ok(e) = ecs.entry_ref(attacker) { if let Ok(pos) = e.get_component::<Position>() { attacker_pos = Some(*pos); } } if let Ok(e) = ecs.entry_ref(victim) { if let Ok(pos) = e.get_component::<Position>() { defender_pos = Some(*pos); } } if attacker_pos.is_none() || defender_pos.is_none() { return; // Bail out - invalid data arrived } let apos = attacker_pos.unwrap(); let dpos = defender_pos.unwrap(); if apos.layer != dpos.layer { return; // Bail out - can't attack across layers } let d = DistanceAlg::Pythagoras.distance2d(apos.pt, dpos.pt); if d > 1.5 { return; // Too far away, bail } // Inflict damage upon the hapless victim let mut dead_entities = Vec::new(); if let Ok(mut v) = ecs.entry_mut(victim)
// If necessary, kill them. let mut commands = CommandBuffer::new(ecs); let mut splatter = None; kill_things(ecs, &mut commands, dead_entities, &mut splatter); // Splatter blood. It's good for you. } fn kill_things( ecs: &mut World, commands: &mut CommandBuffer, dead_entities: Vec<Entity>, splatter: &mut Option<RGB>, ) { dead_entities.iter().for_each(|entity| { crate::stats::record_death(); let mut was_decor = false; let mut was_player = false; if let Ok(mut er) = ecs.entry_mut(*entity) { let mut was_colonist = false;
{ if let Ok(hp) = v.get_component_mut::<Health>() { hp.current = i32::max(0, hp.current - melee_power); if hp.current == 0 { dead_entities.push(victim); } } if let Ok(blood) = v.get_component::<Blood>() { let idx = map.get_layer(dpos.layer as usize).point2d_to_index(dpos.pt); map.get_layer_mut(dpos.layer as usize).tiles[idx].color.fg = blood.0.into(); } }
conditional_block
combat.rs
20); NewState::Player } pub fn ranged_attack( ecs: &mut World, map: &mut Map, attacker: Entity, victim: Entity, ranged_power: i32, ) { let mut attacker_pos = None; let mut victim_pos = None; // Find positions for the start and end if let Ok(ae) = ecs.entry_ref(attacker) { if let Ok(pos) = ae.get_component::<Position>() { attacker_pos = Some(pos.clone()); } } if let Ok(ae) = ecs.entry_ref(victim) { if let Ok(pos) = ae.get_component::<Position>() { victim_pos = Some(pos.clone()); } } if attacker_pos.is_none() || victim_pos.is_none() { return; } let attacker_pos = attacker_pos.unwrap(); let victim_pos = victim_pos.unwrap(); // Set state for the projectile path let mut power = ranged_power; let mut range = 0; let mut projectile_path = Vec::new(); let mut splatter = None; let mut commands = CommandBuffer::new(ecs); let current_layer = attacker_pos.layer; // Map of entity locations. Rebuilt every time because it might change. let pos_map = <(&Position, &Health)>::query() .iter(ecs) .map(|(pos, _)| pos.pt) .collect::<HashSet<Point>>(); // Plot the initial trajectory line2d_bresenham(attacker_pos.pt, victim_pos.pt) .iter() .skip(1) .for_each(|pt| { projectile_path.push(*pt); if pos_map.contains(&pt) { power -= hit_tile_contents(ecs, *pt, current_layer, &mut commands, &mut splatter, power); if power < 1 { power = 1; range += 200; } } if let Some(bsplatter) = &mut splatter { let idx = map.get_current().point2d_to_index(*pt); map.get_current_mut().tiles[idx].color.fg = bsplatter.to_rgba(1.0); bsplatter.r = f32::max(0.0, bsplatter.r - 0.1); bsplatter.g = f32::max(0.0, bsplatter.g - 0.1); bsplatter.b = f32::max(0.0, bsplatter.b - 0.1); if bsplatter.r + bsplatter.g + bsplatter.b < 0.1 { splatter = None; } } range += 1; if range > 5 { power -= 1; if power < 1 { power = 1; } } }); // The trajectory can continue if we have power left use ultraviolet::Vec2; let mut projectile_pos: Vec2 = Vec2::new(victim_pos.pt.x as f32, victim_pos.pt.y as f32); let slope = (projectile_pos - Vec2::new(attacker_pos.pt.x as f32, attacker_pos.pt.y as f32)) .normalized(); while range < 25 && power > 0 { projectile_pos += slope; let pt = Point::new(projectile_pos.x as i32, projectile_pos.y as i32); projectile_path.push(pt); if pos_map.contains(&pt) { power -= hit_tile_contents(ecs, pt, current_layer, &mut commands, &mut splatter, power); if power < 1 { power = 1; range += 200; } } if let Some(bsplatter) = &mut splatter { let idx = map.get_current().point2d_to_index(pt); map.get_current_mut().tiles[idx].color.fg = bsplatter.to_rgba(1.0); bsplatter.r = f32::max(0.0, bsplatter.r - 0.1); bsplatter.g = f32::max(0.0, bsplatter.g - 0.1); bsplatter.b = f32::max(0.0, bsplatter.b - 0.1); if bsplatter.r + bsplatter.g + bsplatter.b < 0.1 { splatter = None; } } let idx = map.get_current().point2d_to_index(pt); if map.get_current().tiles[idx].tile_type == TileType::Wall { range += 100; power = 0; } range += 1; if range > 5 { power -= 1; if power < 1 { power = 1; range += 100; } } } commands.push(( Projectile { path: projectile_path,
}, Glyph { glyph: to_cp437('*'), color: ColorPair::new(RED, BLACK), }, )); commands.flush(ecs); } pub fn hit_tile_contents( ecs: &mut World, pt: Point, layer: u32, commands: &mut CommandBuffer, splatter: &mut Option<RGB>, power: i32, ) -> i32 { let mut rng_lock = crate::RNG.lock(); let rng = rng_lock.as_mut().unwrap(); let mut power_loss = 0; let mut dead_entities = Vec::new(); <(Entity, &Position, &mut Health)>::query() .iter_mut(ecs) .filter(|(_, pos, _)| pos.layer == layer && pos.pt == pt) .for_each(|(entity, _, hp)| { power_loss += hp.current; if power_loss < 0 { power_loss = 0; } let damage = i32::max(0, power + rng.roll_dice(1, 4) - 2); //println!("{}", damage); hp.current -= damage; if hp.current < 0 { hp.current = 0; dead_entities.push(*entity); } }); dead_entities.iter().for_each(|e| { if let Ok(er) = ecs.entry_ref(*e) { if let Ok(boom) = er.get_component::<Explosive>() { if let Ok(pos) = er.get_component::<Position>() { commands.push(( Position::with_pt(pos.pt, pos.layer), Boom { range: boom.range }, )); } } } }); kill_things(ecs, commands, dead_entities, splatter); power_loss } pub fn melee(ecs: &mut World, map: &mut Map, attacker: Entity, victim: Entity, melee_power: i32) { // Check range and validity let mut attacker_pos = None; let mut defender_pos = None; if let Ok(e) = ecs.entry_ref(attacker) { if let Ok(pos) = e.get_component::<Position>() { attacker_pos = Some(*pos); } } if let Ok(e) = ecs.entry_ref(victim) { if let Ok(pos) = e.get_component::<Position>() { defender_pos = Some(*pos); } } if attacker_pos.is_none() || defender_pos.is_none() { return; // Bail out - invalid data arrived } let apos = attacker_pos.unwrap(); let dpos = defender_pos.unwrap(); if apos.layer != dpos.layer { return; // Bail out - can't attack across layers } let d = DistanceAlg::Pythagoras.distance2d(apos.pt, dpos.pt); if d > 1.5 { return; // Too far away, bail } // Inflict damage upon the hapless victim let mut dead_entities = Vec::new(); if let Ok(mut v) = ecs.entry_mut(victim) { if let Ok(hp) = v.get_component_mut::<Health>() { hp.current = i32::max(0, hp.current - melee_power); if hp.current == 0 { dead_entities.push(victim); } } if let Ok(blood) = v.get_component::<Blood>() { let idx = map.get_layer(dpos.layer as usize).point2d_to_index(dpos.pt); map.get_layer_mut(dpos.layer as usize).tiles[idx].color.fg = blood.0.into(); } } // If necessary, kill them. let mut commands = CommandBuffer::new(ecs); let mut splatter = None; kill_things(ecs, &mut commands, dead_entities, &mut splatter); // Splatter blood. It's good for you. } fn kill_things( ecs: &mut World, commands: &mut CommandBuffer, dead_entities: Vec<Entity>, splatter: &mut Option<RGB>, ) { dead_entities.iter().for_each(|entity| { crate::stats::record_death(); let mut was_decor = false; let mut was_player = false; if let Ok(mut er) = ecs.entry_mut(*entity) { let mut was_colonist = false; if
layer: current_layer as usize,
random_line_split
tutor-details.component.ts
requestPermissions:{google:['https://www.googleapis.com/auth/calendar']}, forceApprovalPrompt: {google: true}, requestOfflineToken: {google: true}}); Stripe.setPublishableKey(Meteor.settings.public.stripe.livePublishableKey); var handler = StripeCheckout.configure({ key: Meteor.settings.public.stripe.testPublishableKey, token: function(token) {} }); }); @Component({ selector: 'tutor-details', template, styles: [ style ] }) @InjectUser('user') export class TutorDetailsComponentUser implements OnInit, OnDestroy { loggedIn:boolean=true; user_skype_email: string; today: Date = new Date(); today_show: Date = new Date(); tutorId: string; slot: number; color: string[]= new Array(24); day: number=1; tutorAsUserId: string; tutor: Tutor; paramsSub: Subscription; imagesSubs: Subscription; classesSub: Subscription; reqSub: Subscription; tutorSub: Subscription; requests: Observable<Request[]>; mailtoTutor: string; tutor_user_email: string; class: Class_; tutorClasses: Observable<Class_[]>; user: Meteor.User; checkout: boolean=false; a_day: number[] = new Array(24); tutorSchedule: number[][] = new Array(); colorsSched: string[][] = new Array(); range_for_disp: number[] = new Array(12); amount: number=0; checkDetails: string[]=new Array(3); payment_form: FormGroup; payment_form_2: FormGroup; submitted: boolean = false; g_calendar: boolean=false; // cc cardNumber: string; expiryMonth: string; expiryYear: string; cvc: string; constructor( private router: Router, private formBuilder: FormBuilder, private route: ActivatedRoute )
private myDatePickerOptions: IMyOptions = { // other options... dateFormat: 'dd.mm.yyyy', inline: true, disableDateRanges: [{ begin: {year: this.today.getFullYear(), month: this.today.getMonth()-2, day: this.today.getDate()}, end: {year: this.today.getFullYear(), month: this.today.getMonth()+1, day: this.today.getDate()-1} },{ begin: {year: this.today.getFullYear(),month: this.today.getMonth()+1, day: this.today.getDate()+7}, end: {year: this.today.getFullYear(),month: this.today.getMonth()+2, day: this.today.getDate()} }] }; // Initialized to specific date (09.10.2018). private model: Object = { date: { year: 2018, month: 10, day: 9 } }; ngOnInit() { console.log(Meteor.userId()); if(!Meteor.userId()){ this.loggedIn=false; Bert.alert( 'You need to be logged in to view this page', 'danger', 'fixed-bottom' ); this.router.navigate('/login'); } this.imagesSubs = MeteorObservable.subscribe('images').subscribe(); this.payment_form_2 = this.formBuilder.group({ creditCard: ['', [<any>CreditCardValidator.validateCCNumber]], expDate: ['', [<any>CreditCardValidator.validateExpDate]], cvc: ['', [<any>Validators.required, <any>Validators.minLength(3), <any>Validators.maxLength(4)]] }); this.payment_form = this.formBuilder.group({ cardNumber: ['', Validators.required], expiryMonth: ['', Validators.required], expiryYear: ['', Validators.required], cvc: ['', Validators.required] }); for (var i = 0; i < 24; i++) { this.color[i]='green'; } for (var i = 0; i < 7; i++) { this.colorsSched[i]=this.color; } this.paramsSub = this.route.params .map(params => params['tutorId']) .subscribe(tutorId => { this.tutorId = tutorId; if (this.tutorSub) { this.tutorSub.unsubscribe(); } }); this.tutorSub = MeteorObservable.subscribe('tutors').subscribe(() => { this.tutor=Tutors.findOne(this.tutorId); this.tutorAsUserId=this.tutor.userId; this.tutorSchedule=this.tutor.times; this.amount = this.tutor.hourly_rating; let _MS_PER_DAY = 1000 * 60 * 60 * 24; let utc1 = Date.UTC(this.today.getFullYear(), this.today.getMonth(), this.today.getDate()); let utc2 = Date.UTC(this.tutor.lastUpdateDate.getFullYear(), this.tutor.lastUpdateDate.getMonth(), this.tutor.lastUpdateDate.getDate()); let last_update_diff = Math.floor((utc2 - utc1) / _MS_PER_DAY); for (var i = 0; i < last_update_diff; i++) { for(var j = 0; j < 24; j++) { this.colorsSched[i][j]='blue'; } } for (var i = last_update_diff; i < 7; i++) { for(var j = 0; j < 24; j++) { // console.log(this.tutorSchedule[i][j]); if(this.tutorSchedule[i][j]==1){ this.colorsSched[i][j]='blue'; }else{ this.colorsSched[i][j]='green'; } } } // console.log(this.colorsSched); }); console.log(this.colorsSched); this.tutorSub = MeteorObservable.subscribe('users').subscribe(() => { this.tutor_user_email=Users.findOne(this.tutorAsUserId).emails[0].address; this.mailtoTutor="mailto:"+ this.tutor_user_email; }); //TODO only find classes that this tutor do 34064745 this.classesSub = MeteorObservable.subscribe('classes').subscribe(() => { this.tutorClasses = Classes.find({tutorId: {$eq: this.tutorAsUserId} }); }); } get isMe(): boolean { if(this.user) return this.user._id === this.tutorAsUserId; return false; } toggleSlot(i: number): void { this.today_show.setHours(i,0,0); this.slot = i; console.log(this.tutorSchedule[this.day][i]); if(this.tutorSchedule[this.day][i]==0){ this.tutorSchedule[this.day][i]=1; this.colorsSched[this.day][i]='green'; }else if(this.tutorSchedule[this.day][i]==1){ this.tutorSchedule[this.day][i]=0; this.colorsSched[this.day][i]='blue'; } // else if(this.colorsSched[this.day][i]='blue'){ // this.colorsSched[this.day][i]='green'; // } } onDateChanged(event: IMyDateModel) { let _MS_PER_DAY = 1000 * 60 * 60 * 24; let utc1 = Date.UTC(this.today.getFullYear(), this.today.getMonth(), this.today.getDate()); let utc2 = Date.UTC(event.jsdate.getFullYear(), event.jsdate.getMonth(), event.jsdate.getDate()); let diff = Math.floor((utc2 - utc1) / _MS_PER_DAY); let i = diff; this.day = diff; this.today_show.setDate(this.today.getDate()+i); for(var j = 0; j < 24; j++) { // console.log(this.tutorSchedule[i][j]); if(this.tutorSchedule[i][j]==1){ this.colorsSched[i][j]='green'; }else{ this.colorsSched[i][j]='red'; } } } CheckoutFn():void{ if (this.payment_form.valid) { console.log('payment form valid') Stripe.card.createToken({ number: this.payment_form.value.cardNumber, cvc: this.payment_form.value.cvc, exp_month: this.payment_form.value.expiryMonth, exp_year: this.payment_form.value.expiryYear }, function(status, response) { console.log(status); console.log(response.error.message); stripeToken = response.id; Meteor.call('chargeCard', stripeToken, amount); }); } } step1(): void{ this.checkout = false; } GoToCheckOut(): void{ // if(!this.user_skype_email){ // alert('Please enter your skype username so the teatch can contact you :)'); // }else{ // alert('you are now registered in this class :)'); // } console.log(this.slot); console.log(this.tutorSchedule) this.tutorSchedule[this.day][this.slot]=2; // Tutors.update(this.tutorId, { // $set:{times: this.tutorSchedule } // }); this.checkout=true; } onSubmit() { let success= false; let free= true; if (this.payment_form_2.valid) { this.submitted = true; let m = this.payment_form_2.value.expDate[0]+this.payment_form_2.value.expDate[1]; let y = this.payment_form_2.value.expDate[5
{}
identifier_body
tutor-details.component.ts
({requestPermissions:{google:['https://www.googleapis.com/auth/calendar']}, forceApprovalPrompt: {google: true}, requestOfflineToken: {google: true}}); Stripe.setPublishableKey(Meteor.settings.public.stripe.livePublishableKey); var handler = StripeCheckout.configure({ key: Meteor.settings.public.stripe.testPublishableKey, token: function(token) {} }); }); @Component({ selector: 'tutor-details', template, styles: [ style ] }) @InjectUser('user') export class TutorDetailsComponentUser implements OnInit, OnDestroy { loggedIn:boolean=true; user_skype_email: string; today: Date = new Date(); today_show: Date = new Date(); tutorId: string; slot: number; color: string[]= new Array(24); day: number=1; tutorAsUserId: string; tutor: Tutor; paramsSub: Subscription; imagesSubs: Subscription; classesSub: Subscription; reqSub: Subscription; tutorSub: Subscription; requests: Observable<Request[]>; mailtoTutor: string; tutor_user_email: string; class: Class_; tutorClasses: Observable<Class_[]>; user: Meteor.User; checkout: boolean=false; a_day: number[] = new Array(24); tutorSchedule: number[][] = new Array(); colorsSched: string[][] = new Array(); range_for_disp: number[] = new Array(12); amount: number=0; checkDetails: string[]=new Array(3); payment_form: FormGroup; payment_form_2: FormGroup; submitted: boolean = false; g_calendar: boolean=false; // cc cardNumber: string; expiryMonth: string; expiryYear: string; cvc: string; constructor( private router: Router, private formBuilder: FormBuilder, private route: ActivatedRoute ) {} private myDatePickerOptions: IMyOptions = { // other options... dateFormat: 'dd.mm.yyyy', inline: true, disableDateRanges: [{ begin: {year: this.today.getFullYear(), month: this.today.getMonth()-2, day: this.today.getDate()}, end: {year: this.today.getFullYear(), month: this.today.getMonth()+1, day: this.today.getDate()-1} },{ begin: {year: this.today.getFullYear(),month: this.today.getMonth()+1, day: this.today.getDate()+7}, end: {year: this.today.getFullYear(),month: this.today.getMonth()+2, day: this.today.getDate()} }] }; // Initialized to specific date (09.10.2018). private model: Object = { date: { year: 2018, month: 10, day: 9 } }; ngOnInit() { console.log(Meteor.userId()); if(!Meteor.userId()){ this.loggedIn=false; Bert.alert( 'You need to be logged in to view this page', 'danger', 'fixed-bottom' ); this.router.navigate('/login'); } this.imagesSubs = MeteorObservable.subscribe('images').subscribe(); this.payment_form_2 = this.formBuilder.group({ creditCard: ['', [<any>CreditCardValidator.validateCCNumber]], expDate: ['', [<any>CreditCardValidator.validateExpDate]], cvc: ['', [<any>Validators.required, <any>Validators.minLength(3), <any>Validators.maxLength(4)]] }); this.payment_form = this.formBuilder.group({ cardNumber: ['', Validators.required], expiryMonth: ['', Validators.required], expiryYear: ['', Validators.required], cvc: ['', Validators.required] }); for (var i = 0; i < 24; i++) { this.color[i]='green'; } for (var i = 0; i < 7; i++) { this.colorsSched[i]=this.color; } this.paramsSub = this.route.params .map(params => params['tutorId']) .subscribe(tutorId => { this.tutorId = tutorId; if (this.tutorSub) { this.tutorSub.unsubscribe(); } }); this.tutorSub = MeteorObservable.subscribe('tutors').subscribe(() => { this.tutor=Tutors.findOne(this.tutorId); this.tutorAsUserId=this.tutor.userId; this.tutorSchedule=this.tutor.times; this.amount = this.tutor.hourly_rating; let _MS_PER_DAY = 1000 * 60 * 60 * 24; let utc1 = Date.UTC(this.today.getFullYear(), this.today.getMonth(), this.today.getDate()); let utc2 = Date.UTC(this.tutor.lastUpdateDate.getFullYear(), this.tutor.lastUpdateDate.getMonth(), this.tutor.lastUpdateDate.getDate()); let last_update_diff = Math.floor((utc2 - utc1) / _MS_PER_DAY); for (var i = 0; i < last_update_diff; i++) { for(var j = 0; j < 24; j++) { this.colorsSched[i][j]='blue'; } } for (var i = last_update_diff; i < 7; i++) { for(var j = 0; j < 24; j++) { // console.log(this.tutorSchedule[i][j]); if(this.tutorSchedule[i][j]==1){ this.colorsSched[i][j]='blue'; }else{ this.colorsSched[i][j]='green'; } } } // console.log(this.colorsSched); }); console.log(this.colorsSched); this.tutorSub = MeteorObservable.subscribe('users').subscribe(() => { this.tutor_user_email=Users.findOne(this.tutorAsUserId).emails[0].address; this.mailtoTutor="mailto:"+ this.tutor_user_email; }); //TODO only find classes that this tutor do 34064745 this.classesSub = MeteorObservable.subscribe('classes').subscribe(() => { this.tutorClasses = Classes.find({tutorId: {$eq: this.tutorAsUserId} }); }); } get isMe(): boolean { if(this.user) return this.user._id === this.tutorAsUserId; return false; } toggleSlot(i: number): void { this.today_show.setHours(i,0,0); this.slot = i; console.log(this.tutorSchedule[this.day][i]); if(this.tutorSchedule[this.day][i]==0){ this.tutorSchedule[this.day][i]=1; this.colorsSched[this.day][i]='green'; }else if(this.tutorSchedule[this.day][i]==1){ this.tutorSchedule[this.day][i]=0; this.colorsSched[this.day][i]='blue'; } // else if(this.colorsSched[this.day][i]='blue'){ // this.colorsSched[this.day][i]='green'; // } } onDateChanged(event: IMyDateModel) { let _MS_PER_DAY = 1000 * 60 * 60 * 24; let utc1 = Date.UTC(this.today.getFullYear(), this.today.getMonth(), this.today.getDate()); let utc2 = Date.UTC(event.jsdate.getFullYear(), event.jsdate.getMonth(), event.jsdate.getDate()); let diff = Math.floor((utc2 - utc1) / _MS_PER_DAY); let i = diff; this.day = diff; this.today_show.setDate(this.today.getDate()+i); for(var j = 0; j < 24; j++) { // console.log(this.tutorSchedule[i][j]); if(this.tutorSchedule[i][j]==1){ this.colorsSched[i][j]='green'; }else{ this.colorsSched[i][j]='red'; } } } CheckoutFn():void{ if (this.payment_form.valid) { console.log('payment form valid') Stripe.card.createToken({ number: this.payment_form.value.cardNumber, cvc: this.payment_form.value.cvc, exp_month: this.payment_form.value.expiryMonth, exp_year: this.payment_form.value.expiryYear }, function(status, response) { console.log(status); console.log(response.error.message); stripeToken = response.id; Meteor.call('chargeCard', stripeToken, amount); }); } }
(): void{ this.checkout = false; } GoToCheckOut(): void{ // if(!this.user_skype_email){ // alert('Please enter your skype username so the teatch can contact you :)'); // }else{ // alert('you are now registered in this class :)'); // } console.log(this.slot); console.log(this.tutorSchedule) this.tutorSchedule[this.day][this.slot]=2; // Tutors.update(this.tutorId, { // $set:{times: this.tutorSchedule } // }); this.checkout=true; } onSubmit() { let success= false; let free= true; if (this.payment_form_2.valid) { this.submitted = true; let m = this.payment_form_2.value.expDate[0]+this.payment_form_2.value.expDate[1]; let y = this.payment_form_2.value.expDate[5]+
step1
identifier_name
tutor-details.component.ts
requestPermissions:{google:['https://www.googleapis.com/auth/calendar']}, forceApprovalPrompt: {google: true}, requestOfflineToken: {google: true}}); Stripe.setPublishableKey(Meteor.settings.public.stripe.livePublishableKey); var handler = StripeCheckout.configure({ key: Meteor.settings.public.stripe.testPublishableKey, token: function(token) {} }); }); @Component({ selector: 'tutor-details', template, styles: [ style ] }) @InjectUser('user') export class TutorDetailsComponentUser implements OnInit, OnDestroy { loggedIn:boolean=true; user_skype_email: string; today: Date = new Date(); today_show: Date = new Date(); tutorId: string; slot: number; color: string[]= new Array(24); day: number=1; tutorAsUserId: string; tutor: Tutor; paramsSub: Subscription; imagesSubs: Subscription; classesSub: Subscription; reqSub: Subscription; tutorSub: Subscription; requests: Observable<Request[]>; mailtoTutor: string; tutor_user_email: string; class: Class_; tutorClasses: Observable<Class_[]>; user: Meteor.User; checkout: boolean=false; a_day: number[] = new Array(24); tutorSchedule: number[][] = new Array(); colorsSched: string[][] = new Array(); range_for_disp: number[] = new Array(12); amount: number=0; checkDetails: string[]=new Array(3); payment_form: FormGroup; payment_form_2: FormGroup; submitted: boolean = false; g_calendar: boolean=false; // cc cardNumber: string; expiryMonth: string; expiryYear: string; cvc: string; constructor( private router: Router, private formBuilder: FormBuilder, private route: ActivatedRoute ) {} private myDatePickerOptions: IMyOptions = { // other options... dateFormat: 'dd.mm.yyyy', inline: true, disableDateRanges: [{ begin: {year: this.today.getFullYear(), month: this.today.getMonth()-2, day: this.today.getDate()}, end: {year: this.today.getFullYear(), month: this.today.getMonth()+1, day: this.today.getDate()-1} },{ begin: {year: this.today.getFullYear(),month: this.today.getMonth()+1, day: this.today.getDate()+7}, end: {year: this.today.getFullYear(),month: this.today.getMonth()+2, day: this.today.getDate()} }] }; // Initialized to specific date (09.10.2018). private model: Object = { date: { year: 2018, month: 10, day: 9 } }; ngOnInit() { console.log(Meteor.userId()); if(!Meteor.userId()){ this.loggedIn=false; Bert.alert( 'You need to be logged in to view this page', 'danger', 'fixed-bottom' ); this.router.navigate('/login'); } this.imagesSubs = MeteorObservable.subscribe('images').subscribe(); this.payment_form_2 = this.formBuilder.group({ creditCard: ['', [<any>CreditCardValidator.validateCCNumber]], expDate: ['', [<any>CreditCardValidator.validateExpDate]], cvc: ['', [<any>Validators.required, <any>Validators.minLength(3), <any>Validators.maxLength(4)]] }); this.payment_form = this.formBuilder.group({ cardNumber: ['', Validators.required], expiryMonth: ['', Validators.required], expiryYear: ['', Validators.required], cvc: ['', Validators.required] }); for (var i = 0; i < 24; i++) { this.color[i]='green'; } for (var i = 0; i < 7; i++) { this.colorsSched[i]=this.color; } this.paramsSub = this.route.params .map(params => params['tutorId']) .subscribe(tutorId => { this.tutorId = tutorId; if (this.tutorSub) { this.tutorSub.unsubscribe(); } }); this.tutorSub = MeteorObservable.subscribe('tutors').subscribe(() => { this.tutor=Tutors.findOne(this.tutorId); this.tutorAsUserId=this.tutor.userId; this.tutorSchedule=this.tutor.times; this.amount = this.tutor.hourly_rating; let _MS_PER_DAY = 1000 * 60 * 60 * 24; let utc1 = Date.UTC(this.today.getFullYear(), this.today.getMonth(), this.today.getDate()); let utc2 = Date.UTC(this.tutor.lastUpdateDate.getFullYear(), this.tutor.lastUpdateDate.getMonth(), this.tutor.lastUpdateDate.getDate()); let last_update_diff = Math.floor((utc2 - utc1) / _MS_PER_DAY); for (var i = 0; i < last_update_diff; i++) { for(var j = 0; j < 24; j++) { this.colorsSched[i][j]='blue'; } } for (var i = last_update_diff; i < 7; i++) { for(var j = 0; j < 24; j++) { // console.log(this.tutorSchedule[i][j]); if(this.tutorSchedule[i][j]==1){ this.colorsSched[i][j]='blue'; }else{ this.colorsSched[i][j]='green'; } } } // console.log(this.colorsSched); }); console.log(this.colorsSched); this.tutorSub = MeteorObservable.subscribe('users').subscribe(() => { this.tutor_user_email=Users.findOne(this.tutorAsUserId).emails[0].address; this.mailtoTutor="mailto:"+ this.tutor_user_email; }); //TODO only find classes that this tutor do 34064745 this.classesSub = MeteorObservable.subscribe('classes').subscribe(() => { this.tutorClasses = Classes.find({tutorId: {$eq: this.tutorAsUserId} }); }); } get isMe(): boolean { if(this.user) return this.user._id === this.tutorAsUserId; return false; } toggleSlot(i: number): void { this.today_show.setHours(i,0,0); this.slot = i; console.log(this.tutorSchedule[this.day][i]); if(this.tutorSchedule[this.day][i]==0)
else if(this.tutorSchedule[this.day][i]==1){ this.tutorSchedule[this.day][i]=0; this.colorsSched[this.day][i]='blue'; } // else if(this.colorsSched[this.day][i]='blue'){ // this.colorsSched[this.day][i]='green'; // } } onDateChanged(event: IMyDateModel) { let _MS_PER_DAY = 1000 * 60 * 60 * 24; let utc1 = Date.UTC(this.today.getFullYear(), this.today.getMonth(), this.today.getDate()); let utc2 = Date.UTC(event.jsdate.getFullYear(), event.jsdate.getMonth(), event.jsdate.getDate()); let diff = Math.floor((utc2 - utc1) / _MS_PER_DAY); let i = diff; this.day = diff; this.today_show.setDate(this.today.getDate()+i); for(var j = 0; j < 24; j++) { // console.log(this.tutorSchedule[i][j]); if(this.tutorSchedule[i][j]==1){ this.colorsSched[i][j]='green'; }else{ this.colorsSched[i][j]='red'; } } } CheckoutFn():void{ if (this.payment_form.valid) { console.log('payment form valid') Stripe.card.createToken({ number: this.payment_form.value.cardNumber, cvc: this.payment_form.value.cvc, exp_month: this.payment_form.value.expiryMonth, exp_year: this.payment_form.value.expiryYear }, function(status, response) { console.log(status); console.log(response.error.message); stripeToken = response.id; Meteor.call('chargeCard', stripeToken, amount); }); } } step1(): void{ this.checkout = false; } GoToCheckOut(): void{ // if(!this.user_skype_email){ // alert('Please enter your skype username so the teatch can contact you :)'); // }else{ // alert('you are now registered in this class :)'); // } console.log(this.slot); console.log(this.tutorSchedule) this.tutorSchedule[this.day][this.slot]=2; // Tutors.update(this.tutorId, { // $set:{times: this.tutorSchedule } // }); this.checkout=true; } onSubmit() { let success= false; let free= true; if (this.payment_form_2.valid) { this.submitted = true; let m = this.payment_form_2.value.expDate[0]+this.payment_form_2.value.expDate[1]; let y = this.payment_form_2.value.expDate[5]+
{ this.tutorSchedule[this.day][i]=1; this.colorsSched[this.day][i]='green'; }
conditional_block
tutor-details.component.ts
({requestPermissions:{google:['https://www.googleapis.com/auth/calendar']}, forceApprovalPrompt: {google: true}, requestOfflineToken: {google: true}}); Stripe.setPublishableKey(Meteor.settings.public.stripe.livePublishableKey); var handler = StripeCheckout.configure({ key: Meteor.settings.public.stripe.testPublishableKey, token: function(token) {} }); }); @Component({ selector: 'tutor-details', template, styles: [ style ] }) @InjectUser('user') export class TutorDetailsComponentUser implements OnInit, OnDestroy { loggedIn:boolean=true; user_skype_email: string; today: Date = new Date(); today_show: Date = new Date(); tutorId: string; slot: number; color: string[]= new Array(24); day: number=1; tutorAsUserId: string; tutor: Tutor; paramsSub: Subscription; imagesSubs: Subscription; classesSub: Subscription; reqSub: Subscription; tutorSub: Subscription; requests: Observable<Request[]>; mailtoTutor: string; tutor_user_email: string; class: Class_; tutorClasses: Observable<Class_[]>; user: Meteor.User; checkout: boolean=false; a_day: number[] = new Array(24); tutorSchedule: number[][] = new Array(); colorsSched: string[][] = new Array(); range_for_disp: number[] = new Array(12); amount: number=0; checkDetails: string[]=new Array(3); payment_form: FormGroup; payment_form_2: FormGroup; submitted: boolean = false; g_calendar: boolean=false; // cc cardNumber: string; expiryMonth: string; expiryYear: string; cvc: string; constructor( private router: Router, private formBuilder: FormBuilder, private route: ActivatedRoute ) {} private myDatePickerOptions: IMyOptions = { // other options... dateFormat: 'dd.mm.yyyy', inline: true, disableDateRanges: [{ begin: {year: this.today.getFullYear(), month: this.today.getMonth()-2, day: this.today.getDate()}, end: {year: this.today.getFullYear(), month: this.today.getMonth()+1, day: this.today.getDate()-1} },{ begin: {year: this.today.getFullYear(),month: this.today.getMonth()+1, day: this.today.getDate()+7}, end: {year: this.today.getFullYear(),month: this.today.getMonth()+2, day: this.today.getDate()} }] }; // Initialized to specific date (09.10.2018). private model: Object = { date: { year: 2018, month: 10, day: 9 } }; ngOnInit() { console.log(Meteor.userId()); if(!Meteor.userId()){ this.loggedIn=false; Bert.alert( 'You need to be logged in to view this page', 'danger', 'fixed-bottom' ); this.router.navigate('/login'); } this.imagesSubs = MeteorObservable.subscribe('images').subscribe(); this.payment_form_2 = this.formBuilder.group({ creditCard: ['', [<any>CreditCardValidator.validateCCNumber]], expDate: ['', [<any>CreditCardValidator.validateExpDate]], cvc: ['', [<any>Validators.required, <any>Validators.minLength(3), <any>Validators.maxLength(4)]] }); this.payment_form = this.formBuilder.group({ cardNumber: ['', Validators.required], expiryMonth: ['', Validators.required], expiryYear: ['', Validators.required], cvc: ['', Validators.required] }); for (var i = 0; i < 24; i++) { this.color[i]='green'; } for (var i = 0; i < 7; i++) { this.colorsSched[i]=this.color; } this.paramsSub = this.route.params .map(params => params['tutorId']) .subscribe(tutorId => { this.tutorId = tutorId; if (this.tutorSub) { this.tutorSub.unsubscribe(); } }); this.tutorSub = MeteorObservable.subscribe('tutors').subscribe(() => { this.tutor=Tutors.findOne(this.tutorId); this.tutorAsUserId=this.tutor.userId; this.tutorSchedule=this.tutor.times; this.amount = this.tutor.hourly_rating; let _MS_PER_DAY = 1000 * 60 * 60 * 24; let utc1 = Date.UTC(this.today.getFullYear(), this.today.getMonth(), this.today.getDate()); let utc2 = Date.UTC(this.tutor.lastUpdateDate.getFullYear(), this.tutor.lastUpdateDate.getMonth(), this.tutor.lastUpdateDate.getDate()); let last_update_diff = Math.floor((utc2 - utc1) / _MS_PER_DAY); for (var i = 0; i < last_update_diff; i++) { for(var j = 0; j < 24; j++) { this.colorsSched[i][j]='blue'; } } for (var i = last_update_diff; i < 7; i++) { for(var j = 0; j < 24; j++) { // console.log(this.tutorSchedule[i][j]); if(this.tutorSchedule[i][j]==1){ this.colorsSched[i][j]='blue'; }else{ this.colorsSched[i][j]='green'; } } } // console.log(this.colorsSched); }); console.log(this.colorsSched); this.tutorSub = MeteorObservable.subscribe('users').subscribe(() => { this.tutor_user_email=Users.findOne(this.tutorAsUserId).emails[0].address; this.mailtoTutor="mailto:"+ this.tutor_user_email; }); //TODO only find classes that this tutor do 34064745 this.classesSub = MeteorObservable.subscribe('classes').subscribe(() => { this.tutorClasses = Classes.find({tutorId: {$eq: this.tutorAsUserId} }); }); } get isMe(): boolean { if(this.user) return this.user._id === this.tutorAsUserId; return false; } toggleSlot(i: number): void { this.today_show.setHours(i,0,0); this.slot = i; console.log(this.tutorSchedule[this.day][i]); if(this.tutorSchedule[this.day][i]==0){ this.tutorSchedule[this.day][i]=1; this.colorsSched[this.day][i]='green'; }else if(this.tutorSchedule[this.day][i]==1){ this.tutorSchedule[this.day][i]=0; this.colorsSched[this.day][i]='blue'; } // else if(this.colorsSched[this.day][i]='blue'){ // this.colorsSched[this.day][i]='green'; // } } onDateChanged(event: IMyDateModel) { let _MS_PER_DAY = 1000 * 60 * 60 * 24; let utc1 = Date.UTC(this.today.getFullYear(), this.today.getMonth(), this.today.getDate()); let utc2 = Date.UTC(event.jsdate.getFullYear(), event.jsdate.getMonth(), event.jsdate.getDate()); let diff = Math.floor((utc2 - utc1) / _MS_PER_DAY); let i = diff; this.day = diff; this.today_show.setDate(this.today.getDate()+i); for(var j = 0; j < 24; j++) { // console.log(this.tutorSchedule[i][j]); if(this.tutorSchedule[i][j]==1){ this.colorsSched[i][j]='green'; }else{ this.colorsSched[i][j]='red'; }
console.log('payment form valid') Stripe.card.createToken({ number: this.payment_form.value.cardNumber, cvc: this.payment_form.value.cvc, exp_month: this.payment_form.value.expiryMonth, exp_year: this.payment_form.value.expiryYear }, function(status, response) { console.log(status); console.log(response.error.message); stripeToken = response.id; Meteor.call('chargeCard', stripeToken, amount); }); } } step1(): void{ this.checkout = false; } GoToCheckOut(): void{ // if(!this.user_skype_email){ // alert('Please enter your skype username so the teatch can contact you :)'); // }else{ // alert('you are now registered in this class :)'); // } console.log(this.slot); console.log(this.tutorSchedule) this.tutorSchedule[this.day][this.slot]=2; // Tutors.update(this.tutorId, { // $set:{times: this.tutorSchedule } // }); this.checkout=true; } onSubmit() { let success= false; let free= true; if (this.payment_form_2.valid) { this.submitted = true; let m = this.payment_form_2.value.expDate[0]+this.payment_form_2.value.expDate[1]; let y = this.payment_form_2.value.expDate[5]+
} } CheckoutFn():void{ if (this.payment_form.valid) {
random_line_split
lib.go
pongWait = 60 * time.Second // Send pings to client with this period. Must be less than pongWait. pingPeriod = (pongWait * 9) / 10 // Time allowed to write the file to the client. writeWait = 10 * time.Second ) // Event types const ( EVENT_CONNECT = iota EVENT_ROAMING EVENT_DISCONNECT EVENT_LEVEL ) type LeaseEntry struct { IP string MAC string Server string Hostname string Comment string } type ReportEntry struct { IP string Name string Interface string SSID string MAC string Signal string Hostname string Comment string } type ReportEvent struct { EventType int Old ReportEntry New ReportEntry } var WS = websocket.Upgrader{ ReadBufferSize: 1024, WriteBufferSize: 1024, } type BroadcastData struct { Report []ReportEntry ReportMap map[string]ReportEntry Data string LastUpdate time.Time sync.RWMutex ReportChan chan ReportEvent } type LeaseList struct { List []LeaseEntry sync.RWMutex } type ConfMikrotik struct { Address string `yaml:"address"` Username string `yaml:"username"` Password string `yaml:"password"` Interval time.Duration `yaml:"interval"` Mode string `yaml:"mode"` } type ConfDevice struct { Name string `yaml:"name"` MAC string `yaml:"mac"` OnConnect ConfigEvent `yaml:"on.connect"` OnDisconnect ConfigEvent `yaml:"on.disconnect"` OnRoaming ConfigEvent `yaml:"on.roaming"` OnLevel ConfigEvent `yaml:"on.level"` } type ConfigEvent struct { HttpPost string `yaml:"http.post"` HttpGet string `yaml:"http.get"` HttpPostContent string `yaml:"http.post.content"` HttpHeader map[string]string `yaml:"http.header"` } type LogInfo struct { Level log.Level `yaml:"level"` } type Config struct { Log LogInfo `yaml:"log"` Router ConfMikrotik `yaml:"router"` DHCP ConfMikrotik `yaml:"dhcp"` Devices []ConfDevice `yaml:"devices"` } // Init BroadcastData entry func (b *BroadcastData) Init() { b.ReportMap = map[string]ReportEntry{} b.ReportChan = make(chan ReportEvent) } var broadcastData BroadcastData var leaseList LeaseList var config Config var configMTX sync.RWMutex var devList map[string]ConfDevice var devListMTX sync.RWMutex func GetDHCPLeases(address, username, password string) (list []LeaseEntry, err error) { cl, err := routeros.Dial(address, username, password) if err != nil { return } defer cl.Close() reply, err := cl.Run("/ip/dhcp-server/lease/print") if err != nil { return } for _, re := range reply.Re { list = append(list, LeaseEntry{ IP: re.Map["address"], MAC: re.Map["mac-address"], Server: re.Map["server"], Hostname: re.Map["host-name"], Comment: re.Map["comment"], }) } return } func reloadDHCP() { ticker := time.NewTicker(config.DHCP.Interval) for { // nolint:gosimple select { case <-ticker.C: l, err := GetDHCPLeases(config.DHCP.Address, config.DHCP.Username, config.DHCP.Password) if err != nil { log.WithFields(log.Fields{"dhcp-addr": config.DHCP.Address}).Error("Error reloading DHCP Leases: ", err) return } else { leaseList.RLock() leaseList.List = l leaseList.RUnlock() log.WithFields(log.Fields{"count": len(l)}).Debug("Reloaded DHCP Leases") } } } } func FindLeaseByMAC(list []LeaseEntry, mac string) (e LeaseEntry, ok bool) { for _, e := range list { if e.MAC == mac { return e, true } } return } func RTLoop(c *routeros.Client, conf *Config) { for { cmd := "/caps-man/registration-table/print" if strings.ToLower(config.Router.Mode) == "wifi" { cmd = "/interface/wireless/registration-table/print" } reply, err := c.Run(cmd) if err != nil { log.WithFields(log.Fields{"address": config.Router.Address, "username": config.Router.Username}).Error("Error during request to CapsMan server: ", err) // Try to close connection c.Close() // Reconnect loop for { // Sleep for 5 sec time.Sleep(5 * time.Second) cNew, err := routeros.Dial(config.Router.Address, config.Router.Username, config.Router.Password) if err != nil { log.WithFields(log.Fields{"address": config.Router.Address, "username": config.Router.Username}).Error("Reconnect error to CapsMan server: ", err) continue } c = cNew log.WithFields(log.Fields{"address": config.Router.Address, "username": config.Router.Username}).Warn("Reconnected to CapsMan server") break } continue } var report []ReportEntry leaseList.RLock() for _, re := range reply.Re { var n, c, ip string if le, ok := FindLeaseByMAC(leaseList.List, re.Map["mac-address"]); ok { n = le.Hostname c = le.Comment ip = le.IP } devListMTX.RLock() rec := ReportEntry{ IP: ip, Name: devList[re.Map["mac-address"]].Name, Interface: re.Map["interface"], SSID: re.Map["ssid"], MAC: re.Map["mac-address"], Signal: re.Map["rx-signal"], Hostname: n, Comment: c, } if strings.ToLower(config.Router.Mode) == "wifi" { rec.Signal = re.Map["signal-strength"] if i := strings.Index(rec.Signal, "@"); i > 0 { rec.Signal = rec.Signal[0:i] } } devListMTX.RUnlock() report = append(report, rec) // fmt.Printf("%-20s\t%-20s\t%-20s\t%-10s\t%-30s\t%-30s\n", re.Map["interface"], re.Map["ssid"], re.Map["mac-address"], re.Map["rx-signal"], n, c) } log.WithFields(log.Fields{"count": len(report)}).Debug("Reloaded CapsMan entries") leaseList.RUnlock() if err = broadcastData.reportUpdate(report); err != nil { log.WithFields(log.Fields{}).Warn("Error during reportUpdate: ", err) } time.Sleep(*interval) } } func loadConfig(configFileName string) (config Config, err error) { devListMTX.RLock() defer devListMTX.RUnlock() config = Config{} devList = make(map[string]ConfDevice) source, err := ioutil.ReadFile(configFileName) if err != nil { err = fmt.Errorf("cannot read config file [%s]", configFileName) return } if err = yaml.Unmarshal(source, &config); err != nil { err = fmt.Errorf("error parsing config file [%s]: %v", configFileName, err) return } for _, v := range config.Devices { devList[strings.ToUpper(v.MAC)] = v } return } func usage() { } // Handle report update request func (b *BroadcastData) reportUpdate(report []ReportEntry) error { output, err := json.Marshal(report) if err != nil { return err } // Lock mutex b.RLock() defer b.RUnlock() // Prepare new list of entries rm := map[string]ReportEntry{} for _, v := range report { rm[v.MAC] = v } // Scan for new entries for k := range rm { if _, ok := b.ReportMap[k]; !ok { // New entry b.ReportChan <- ReportEvent{ EventType: EVENT_CONNECT, New: rm[k], } } else { // Check for roaming if rm[k].Interface != b.ReportMap[k].Interface { b.ReportChan <- ReportEvent{ EventType: EVENT_ROAMING, Old: b.ReportMap[k], New: rm[k], } } // Check for signal level change if rm[k].Signal != b.ReportMap[k].Signal { b.ReportChan <- ReportEvent{ EventType: EVENT_LEVEL, Old: b.ReportMap[k], New: rm[k],
const ( // Time allowed to read the next pong message from the client.
random_line_split
lib.go
.Time sync.RWMutex ReportChan chan ReportEvent } type LeaseList struct { List []LeaseEntry sync.RWMutex } type ConfMikrotik struct { Address string `yaml:"address"` Username string `yaml:"username"` Password string `yaml:"password"` Interval time.Duration `yaml:"interval"` Mode string `yaml:"mode"` } type ConfDevice struct { Name string `yaml:"name"` MAC string `yaml:"mac"` OnConnect ConfigEvent `yaml:"on.connect"` OnDisconnect ConfigEvent `yaml:"on.disconnect"` OnRoaming ConfigEvent `yaml:"on.roaming"` OnLevel ConfigEvent `yaml:"on.level"` } type ConfigEvent struct { HttpPost string `yaml:"http.post"` HttpGet string `yaml:"http.get"` HttpPostContent string `yaml:"http.post.content"` HttpHeader map[string]string `yaml:"http.header"` } type LogInfo struct { Level log.Level `yaml:"level"` } type Config struct { Log LogInfo `yaml:"log"` Router ConfMikrotik `yaml:"router"` DHCP ConfMikrotik `yaml:"dhcp"` Devices []ConfDevice `yaml:"devices"` } // Init BroadcastData entry func (b *BroadcastData) Init() { b.ReportMap = map[string]ReportEntry{} b.ReportChan = make(chan ReportEvent) } var broadcastData BroadcastData var leaseList LeaseList var config Config var configMTX sync.RWMutex var devList map[string]ConfDevice var devListMTX sync.RWMutex func GetDHCPLeases(address, username, password string) (list []LeaseEntry, err error) { cl, err := routeros.Dial(address, username, password) if err != nil { return } defer cl.Close() reply, err := cl.Run("/ip/dhcp-server/lease/print") if err != nil { return } for _, re := range reply.Re { list = append(list, LeaseEntry{ IP: re.Map["address"], MAC: re.Map["mac-address"], Server: re.Map["server"], Hostname: re.Map["host-name"], Comment: re.Map["comment"], }) } return } func reloadDHCP()
func FindLeaseByMAC(list []LeaseEntry, mac string) (e LeaseEntry, ok bool) { for _, e := range list { if e.MAC == mac { return e, true } } return } func RTLoop(c *routeros.Client, conf *Config) { for { cmd := "/caps-man/registration-table/print" if strings.ToLower(config.Router.Mode) == "wifi" { cmd = "/interface/wireless/registration-table/print" } reply, err := c.Run(cmd) if err != nil { log.WithFields(log.Fields{"address": config.Router.Address, "username": config.Router.Username}).Error("Error during request to CapsMan server: ", err) // Try to close connection c.Close() // Reconnect loop for { // Sleep for 5 sec time.Sleep(5 * time.Second) cNew, err := routeros.Dial(config.Router.Address, config.Router.Username, config.Router.Password) if err != nil { log.WithFields(log.Fields{"address": config.Router.Address, "username": config.Router.Username}).Error("Reconnect error to CapsMan server: ", err) continue } c = cNew log.WithFields(log.Fields{"address": config.Router.Address, "username": config.Router.Username}).Warn("Reconnected to CapsMan server") break } continue } var report []ReportEntry leaseList.RLock() for _, re := range reply.Re { var n, c, ip string if le, ok := FindLeaseByMAC(leaseList.List, re.Map["mac-address"]); ok { n = le.Hostname c = le.Comment ip = le.IP } devListMTX.RLock() rec := ReportEntry{ IP: ip, Name: devList[re.Map["mac-address"]].Name, Interface: re.Map["interface"], SSID: re.Map["ssid"], MAC: re.Map["mac-address"], Signal: re.Map["rx-signal"], Hostname: n, Comment: c, } if strings.ToLower(config.Router.Mode) == "wifi" { rec.Signal = re.Map["signal-strength"] if i := strings.Index(rec.Signal, "@"); i > 0 { rec.Signal = rec.Signal[0:i] } } devListMTX.RUnlock() report = append(report, rec) // fmt.Printf("%-20s\t%-20s\t%-20s\t%-10s\t%-30s\t%-30s\n", re.Map["interface"], re.Map["ssid"], re.Map["mac-address"], re.Map["rx-signal"], n, c) } log.WithFields(log.Fields{"count": len(report)}).Debug("Reloaded CapsMan entries") leaseList.RUnlock() if err = broadcastData.reportUpdate(report); err != nil { log.WithFields(log.Fields{}).Warn("Error during reportUpdate: ", err) } time.Sleep(*interval) } } func loadConfig(configFileName string) (config Config, err error) { devListMTX.RLock() defer devListMTX.RUnlock() config = Config{} devList = make(map[string]ConfDevice) source, err := ioutil.ReadFile(configFileName) if err != nil { err = fmt.Errorf("cannot read config file [%s]", configFileName) return } if err = yaml.Unmarshal(source, &config); err != nil { err = fmt.Errorf("error parsing config file [%s]: %v", configFileName, err) return } for _, v := range config.Devices { devList[strings.ToUpper(v.MAC)] = v } return } func usage() { } // Handle report update request func (b *BroadcastData) reportUpdate(report []ReportEntry) error { output, err := json.Marshal(report) if err != nil { return err } // Lock mutex b.RLock() defer b.RUnlock() // Prepare new list of entries rm := map[string]ReportEntry{} for _, v := range report { rm[v.MAC] = v } // Scan for new entries for k := range rm { if _, ok := b.ReportMap[k]; !ok { // New entry b.ReportChan <- ReportEvent{ EventType: EVENT_CONNECT, New: rm[k], } } else { // Check for roaming if rm[k].Interface != b.ReportMap[k].Interface { b.ReportChan <- ReportEvent{ EventType: EVENT_ROAMING, Old: b.ReportMap[k], New: rm[k], } } // Check for signal level change if rm[k].Signal != b.ReportMap[k].Signal { b.ReportChan <- ReportEvent{ EventType: EVENT_LEVEL, Old: b.ReportMap[k], New: rm[k], } } } } // Scan for deleted entries for k := range b.ReportMap { if _, ok := rm[k]; !ok { b.ReportChan <- ReportEvent{ EventType: EVENT_DISCONNECT, Old: b.ReportMap[k], } } } b.ReportMap = rm b.Report = report b.Data = string(output) b.LastUpdate = time.Now() return nil } func (b *BroadcastData) EventHandler() { for { // nolint:gosimple select { case data := <-b.ReportChan: // fmt.Printf("New event received: %v\n", data) switch data.EventType { case EVENT_CONNECT: log.WithFields(log.Fields{"action": "register", "mac": data.New.MAC, "name": data.New.Name, "interface": data.New.Interface, "ssid": data.New.SSID, "hostname": data.New.Hostname, "comment": data.New.Comment, "level-to": data.New.Signal}).Info("New connection registered") // Get device info devListMTX.RLock() dev, ok := devList[data.New.MAC] devListMTX.RUnlock()
{ ticker := time.NewTicker(config.DHCP.Interval) for { // nolint:gosimple select { case <-ticker.C: l, err := GetDHCPLeases(config.DHCP.Address, config.DHCP.Username, config.DHCP.Password) if err != nil { log.WithFields(log.Fields{"dhcp-addr": config.DHCP.Address}).Error("Error reloading DHCP Leases: ", err) return } else { leaseList.RLock() leaseList.List = l leaseList.RUnlock() log.WithFields(log.Fields{"count": len(l)}).Debug("Reloaded DHCP Leases") } } } }
identifier_body
lib.go
time.Time sync.RWMutex ReportChan chan ReportEvent } type LeaseList struct { List []LeaseEntry sync.RWMutex } type ConfMikrotik struct { Address string `yaml:"address"` Username string `yaml:"username"` Password string `yaml:"password"` Interval time.Duration `yaml:"interval"` Mode string `yaml:"mode"` } type ConfDevice struct { Name string `yaml:"name"` MAC string `yaml:"mac"` OnConnect ConfigEvent `yaml:"on.connect"` OnDisconnect ConfigEvent `yaml:"on.disconnect"` OnRoaming ConfigEvent `yaml:"on.roaming"` OnLevel ConfigEvent `yaml:"on.level"` } type ConfigEvent struct { HttpPost string `yaml:"http.post"` HttpGet string `yaml:"http.get"` HttpPostContent string `yaml:"http.post.content"` HttpHeader map[string]string `yaml:"http.header"` } type LogInfo struct { Level log.Level `yaml:"level"` } type Config struct { Log LogInfo `yaml:"log"` Router ConfMikrotik `yaml:"router"` DHCP ConfMikrotik `yaml:"dhcp"` Devices []ConfDevice `yaml:"devices"` } // Init BroadcastData entry func (b *BroadcastData) Init() { b.ReportMap = map[string]ReportEntry{} b.ReportChan = make(chan ReportEvent) } var broadcastData BroadcastData var leaseList LeaseList var config Config var configMTX sync.RWMutex var devList map[string]ConfDevice var devListMTX sync.RWMutex func GetDHCPLeases(address, username, password string) (list []LeaseEntry, err error) { cl, err := routeros.Dial(address, username, password) if err != nil { return } defer cl.Close() reply, err := cl.Run("/ip/dhcp-server/lease/print") if err != nil { return } for _, re := range reply.Re { list = append(list, LeaseEntry{ IP: re.Map["address"], MAC: re.Map["mac-address"], Server: re.Map["server"], Hostname: re.Map["host-name"], Comment: re.Map["comment"], }) } return } func reloadDHCP() { ticker := time.NewTicker(config.DHCP.Interval) for { // nolint:gosimple select { case <-ticker.C: l, err := GetDHCPLeases(config.DHCP.Address, config.DHCP.Username, config.DHCP.Password) if err != nil { log.WithFields(log.Fields{"dhcp-addr": config.DHCP.Address}).Error("Error reloading DHCP Leases: ", err) return } else { leaseList.RLock() leaseList.List = l leaseList.RUnlock() log.WithFields(log.Fields{"count": len(l)}).Debug("Reloaded DHCP Leases") } } } } func FindLeaseByMAC(list []LeaseEntry, mac string) (e LeaseEntry, ok bool) { for _, e := range list { if e.MAC == mac { return e, true } } return } func RTLoop(c *routeros.Client, conf *Config) { for { cmd := "/caps-man/registration-table/print" if strings.ToLower(config.Router.Mode) == "wifi" { cmd = "/interface/wireless/registration-table/print" } reply, err := c.Run(cmd) if err != nil { log.WithFields(log.Fields{"address": config.Router.Address, "username": config.Router.Username}).Error("Error during request to CapsMan server: ", err) // Try to close connection c.Close() // Reconnect loop for { // Sleep for 5 sec time.Sleep(5 * time.Second) cNew, err := routeros.Dial(config.Router.Address, config.Router.Username, config.Router.Password) if err != nil { log.WithFields(log.Fields{"address": config.Router.Address, "username": config.Router.Username}).Error("Reconnect error to CapsMan server: ", err) continue } c = cNew log.WithFields(log.Fields{"address": config.Router.Address, "username": config.Router.Username}).Warn("Reconnected to CapsMan server") break } continue } var report []ReportEntry leaseList.RLock() for _, re := range reply.Re { var n, c, ip string if le, ok := FindLeaseByMAC(leaseList.List, re.Map["mac-address"]); ok { n = le.Hostname c = le.Comment ip = le.IP } devListMTX.RLock() rec := ReportEntry{ IP: ip, Name: devList[re.Map["mac-address"]].Name, Interface: re.Map["interface"], SSID: re.Map["ssid"], MAC: re.Map["mac-address"], Signal: re.Map["rx-signal"], Hostname: n, Comment: c, } if strings.ToLower(config.Router.Mode) == "wifi" { rec.Signal = re.Map["signal-strength"] if i := strings.Index(rec.Signal, "@"); i > 0 { rec.Signal = rec.Signal[0:i] } } devListMTX.RUnlock() report = append(report, rec) // fmt.Printf("%-20s\t%-20s\t%-20s\t%-10s\t%-30s\t%-30s\n", re.Map["interface"], re.Map["ssid"], re.Map["mac-address"], re.Map["rx-signal"], n, c) } log.WithFields(log.Fields{"count": len(report)}).Debug("Reloaded CapsMan entries") leaseList.RUnlock() if err = broadcastData.reportUpdate(report); err != nil { log.WithFields(log.Fields{}).Warn("Error during reportUpdate: ", err) } time.Sleep(*interval) } } func loadConfig(configFileName string) (config Config, err error) { devListMTX.RLock() defer devListMTX.RUnlock() config = Config{} devList = make(map[string]ConfDevice) source, err := ioutil.ReadFile(configFileName) if err != nil { err = fmt.Errorf("cannot read config file [%s]", configFileName) return } if err = yaml.Unmarshal(source, &config); err != nil { err = fmt.Errorf("error parsing config file [%s]: %v", configFileName, err) return } for _, v := range config.Devices { devList[strings.ToUpper(v.MAC)] = v } return } func
() { } // Handle report update request func (b *BroadcastData) reportUpdate(report []ReportEntry) error { output, err := json.Marshal(report) if err != nil { return err } // Lock mutex b.RLock() defer b.RUnlock() // Prepare new list of entries rm := map[string]ReportEntry{} for _, v := range report { rm[v.MAC] = v } // Scan for new entries for k := range rm { if _, ok := b.ReportMap[k]; !ok { // New entry b.ReportChan <- ReportEvent{ EventType: EVENT_CONNECT, New: rm[k], } } else { // Check for roaming if rm[k].Interface != b.ReportMap[k].Interface { b.ReportChan <- ReportEvent{ EventType: EVENT_ROAMING, Old: b.ReportMap[k], New: rm[k], } } // Check for signal level change if rm[k].Signal != b.ReportMap[k].Signal { b.ReportChan <- ReportEvent{ EventType: EVENT_LEVEL, Old: b.ReportMap[k], New: rm[k], } } } } // Scan for deleted entries for k := range b.ReportMap { if _, ok := rm[k]; !ok { b.ReportChan <- ReportEvent{ EventType: EVENT_DISCONNECT, Old: b.ReportMap[k], } } } b.ReportMap = rm b.Report = report b.Data = string(output) b.LastUpdate = time.Now() return nil } func (b *BroadcastData) EventHandler() { for { // nolint:gosimple select { case data := <-b.ReportChan: // fmt.Printf("New event received: %v\n", data) switch data.EventType { case EVENT_CONNECT: log.WithFields(log.Fields{"action": "register", "mac": data.New.MAC, "name": data.New.Name, "interface": data.New.Interface, "ssid": data.New.SSID, "hostname": data.New.Hostname, "comment": data.New.Comment, "level-to": data.New.Signal}).Info("New connection registered") // Get device info devListMTX.RLock() dev, ok := devList[data.New.MAC] devListMTX.RUnlock()
usage
identifier_name
lib.go
"wifi" { cmd = "/interface/wireless/registration-table/print" } reply, err := c.Run(cmd) if err != nil { log.WithFields(log.Fields{"address": config.Router.Address, "username": config.Router.Username}).Error("Error during request to CapsMan server: ", err) // Try to close connection c.Close() // Reconnect loop for { // Sleep for 5 sec time.Sleep(5 * time.Second) cNew, err := routeros.Dial(config.Router.Address, config.Router.Username, config.Router.Password) if err != nil { log.WithFields(log.Fields{"address": config.Router.Address, "username": config.Router.Username}).Error("Reconnect error to CapsMan server: ", err) continue } c = cNew log.WithFields(log.Fields{"address": config.Router.Address, "username": config.Router.Username}).Warn("Reconnected to CapsMan server") break } continue } var report []ReportEntry leaseList.RLock() for _, re := range reply.Re { var n, c, ip string if le, ok := FindLeaseByMAC(leaseList.List, re.Map["mac-address"]); ok { n = le.Hostname c = le.Comment ip = le.IP } devListMTX.RLock() rec := ReportEntry{ IP: ip, Name: devList[re.Map["mac-address"]].Name, Interface: re.Map["interface"], SSID: re.Map["ssid"], MAC: re.Map["mac-address"], Signal: re.Map["rx-signal"], Hostname: n, Comment: c, } if strings.ToLower(config.Router.Mode) == "wifi" { rec.Signal = re.Map["signal-strength"] if i := strings.Index(rec.Signal, "@"); i > 0 { rec.Signal = rec.Signal[0:i] } } devListMTX.RUnlock() report = append(report, rec) // fmt.Printf("%-20s\t%-20s\t%-20s\t%-10s\t%-30s\t%-30s\n", re.Map["interface"], re.Map["ssid"], re.Map["mac-address"], re.Map["rx-signal"], n, c) } log.WithFields(log.Fields{"count": len(report)}).Debug("Reloaded CapsMan entries") leaseList.RUnlock() if err = broadcastData.reportUpdate(report); err != nil { log.WithFields(log.Fields{}).Warn("Error during reportUpdate: ", err) } time.Sleep(*interval) } } func loadConfig(configFileName string) (config Config, err error) { devListMTX.RLock() defer devListMTX.RUnlock() config = Config{} devList = make(map[string]ConfDevice) source, err := ioutil.ReadFile(configFileName) if err != nil { err = fmt.Errorf("cannot read config file [%s]", configFileName) return } if err = yaml.Unmarshal(source, &config); err != nil { err = fmt.Errorf("error parsing config file [%s]: %v", configFileName, err) return } for _, v := range config.Devices { devList[strings.ToUpper(v.MAC)] = v } return } func usage() { } // Handle report update request func (b *BroadcastData) reportUpdate(report []ReportEntry) error { output, err := json.Marshal(report) if err != nil { return err } // Lock mutex b.RLock() defer b.RUnlock() // Prepare new list of entries rm := map[string]ReportEntry{} for _, v := range report { rm[v.MAC] = v } // Scan for new entries for k := range rm { if _, ok := b.ReportMap[k]; !ok { // New entry b.ReportChan <- ReportEvent{ EventType: EVENT_CONNECT, New: rm[k], } } else { // Check for roaming if rm[k].Interface != b.ReportMap[k].Interface { b.ReportChan <- ReportEvent{ EventType: EVENT_ROAMING, Old: b.ReportMap[k], New: rm[k], } } // Check for signal level change if rm[k].Signal != b.ReportMap[k].Signal { b.ReportChan <- ReportEvent{ EventType: EVENT_LEVEL, Old: b.ReportMap[k], New: rm[k], } } } } // Scan for deleted entries for k := range b.ReportMap { if _, ok := rm[k]; !ok { b.ReportChan <- ReportEvent{ EventType: EVENT_DISCONNECT, Old: b.ReportMap[k], } } } b.ReportMap = rm b.Report = report b.Data = string(output) b.LastUpdate = time.Now() return nil } func (b *BroadcastData) EventHandler() { for { // nolint:gosimple select { case data := <-b.ReportChan: // fmt.Printf("New event received: %v\n", data) switch data.EventType { case EVENT_CONNECT: log.WithFields(log.Fields{"action": "register", "mac": data.New.MAC, "name": data.New.Name, "interface": data.New.Interface, "ssid": data.New.SSID, "hostname": data.New.Hostname, "comment": data.New.Comment, "level-to": data.New.Signal}).Info("New connection registered") // Get device info devListMTX.RLock() dev, ok := devList[data.New.MAC] devListMTX.RUnlock() if ok { if (len(dev.OnConnect.HttpPost) > 0) || (len(dev.OnConnect.HttpGet) > 0) { go makeRequest(dev.OnConnect, map[string]string{ "name": dev.Name, "mac": data.New.MAC, "roaming.to": "", "roaming.from": "", "level.to": data.New.Signal, "level.from": "", }) } } case EVENT_DISCONNECT: log.WithFields(log.Fields{"action": "disconnect", "mac": data.Old.MAC, "name": data.Old.Name, "interface": data.Old.Interface, "hostname": data.Old.Hostname, "comment": data.Old.Comment}).Info("Client disconnect") // Get device info devListMTX.RLock() dev, ok := devList[data.New.MAC] devListMTX.RUnlock() if ok { if (len(dev.OnDisconnect.HttpPost) > 0) || (len(dev.OnDisconnect.HttpGet) > 0) { go makeRequest(dev.OnDisconnect, map[string]string{ "name": dev.Name, "mac": data.Old.MAC, "roaming.to": "", "roaming.from": "", "level.to": "", "level.from": data.Old.Signal, }) } } case EVENT_ROAMING: log.WithFields(log.Fields{"action": "roaming", "mac": data.New.MAC, "name": data.New.Name, "interface-from": data.Old.Interface, "interface-to": data.New.Interface, "level-from": data.Old.Signal, "level-to": data.New.Signal}).Info("Client roaming") // Get device info devListMTX.RLock() dev, ok := devList[data.New.MAC] devListMTX.RUnlock() if ok { if (len(dev.OnRoaming.HttpPost) > 0) || (len(dev.OnRoaming.HttpGet) > 0) { go makeRequest(dev.OnRoaming, map[string]string{ "name": dev.Name, "mac": data.New.MAC, "roaming.to": data.New.Interface, "roaming.from": data.Old.Interface, "level.from": data.Old.Signal, "level.to": data.New.Signal, }) } } case EVENT_LEVEL: log.WithFields(log.Fields{"action": "level", "mac": data.New.MAC, "name": data.New.Name, "interface": data.New.Interface, "level-from": data.Old.Signal, "level-to": data.New.Signal}).Debug("Signal level change") // Get device info devListMTX.RLock() dev, ok := devList[data.New.MAC] devListMTX.RUnlock() if ok
{ if (len(dev.OnLevel.HttpPost) > 0) || (len(dev.OnLevel.HttpGet) > 0) { go makeRequest(dev.OnLevel, map[string]string{ "name": dev.Name, "mac": data.Old.MAC, "roaming.to": "", "roaming.from": "", "level.from": data.Old.Signal, "level.to": data.New.Signal, }) } }
conditional_block
uiTools.js
allPossibleKesmNames', function() { return KesmNames.find().map(function(doc) { return doc.kesmName; }); }); }); getAllDataForKesmName = function(kesmName) { var properties = {}; if (propertyCollections && _.has(propertyCollections, kesmName)) { _.each(propertyCollections[kesmName].find().fetch(), function(doc) { properties[doc.property] = doc.value; }); } var sample = Samples.findOne(properties['currentSampleID']) || null; var machine = Machines.findOne({ kesmName: kesmName }); var process; if (!! machine) { process = Processes.findOne({ className: { $in: ['KESMMain', 'KESM'] }, machineId: machine._id }); } else { machine = null; process = null; } var uri = "http://" + Meteor.settings.public.serverURIs[kesmName]; return { kesmName: kesmName, machine: machine, process: process, properties: properties, sample: sample, uri: uri }; }; Template.registerHelper('allDataForKesmName', getAllDataForKesmName); Template.registerHelper('odd', function(value) { return (value % 2 !== 0); }); Template.registerHelper('classForDot', function(active) { return active? 'green' : 'red'; }); toTitleCase = function(str) { return str.replace(/\w\S*/g, function(txt) { return txt.charAt(0).toUpperCase() + txt.substr(1).toLowerCase(); }); }; // Small Collection of utility functions to make the gui-widgits dance etc. AxisList = [{ axis: 'x', AXIS: 'X', index: 0 }, { axis: 'y', AXIS: 'Y', index: 1 }, { axis: 'z', AXIS: 'Z', index: 2 }]; Template.registerHelper('axes', function() { return AxisList; }); axisIndex = function(axisLetter) { return _.findWhere(AxisList, { axis: axisLetter.toLowerCase() }).index; }; var axisIsInFault = function(axis) { return getPropValue('stage_'+axis+'_fault', false); }; Template.registerHelper('colorClassForAxis', function(axisLetter) { var colorForAxis = { x: 'error', y: 'success', z: 'info', }; if (axisIsInFault(axisLetter)) { return 'fault'; } else { return colorForAxis[axisLetter]; } }); Template.registerHelper('isInFault', axisIsInFault); var scopeModes = [{ mode: 'maintenance' }, { mode: 'focus' }, { mode: 'slicing' }]; Template.registerHelper('scopeModes', function() { return scopeModes; }); Template.registerHelper('scopeIsInMode', function(mode) { var scopeMode = getPropValue('scopeMode', 'maintenance'); return scopeMode.value === mode; }); Template.registerHelper('changingModes', function() { var scopeMode = getPropValue('scopeMode', 'maintenance'); return scopeMode.value !== scopeMode.requestedValue; }); firmlyInSlicingMode = function() { var scopeMode = getProperty('scopeMode', 'maintenance'); return (scopeMode.value === 'slicing') && (scopeMode.value === scopeMode.requestedValue); }; slicingOrChanging = function() { var scopeMode = getProperty('scopeMode', 'maintenance'); return (scopeMode.value === 'slicing') || scopeMode.value !== scopeMode.requestedValue; }; Template.registerHelper('slicingOrChanging', slicingOrChanging); // Returns 'disabled' if the scope is in 'slicing' mode, or if it's changing modes. // This is used to disable buttons and inputs. disabledIfSlicingOrChanging = function() { return slicingOrChanging()? 'disabled' : ''; }; Template.registerHelper('disabledIfSlicingOrChanging', disabledIfSlicingOrChanging); Template.registerHelper('disabledUnlessSlicing', function() { return firmlyInSlicingMode()? '' : 'disabled'; }); // The default number of digits to display for floating point numbers. // If this is 0, then num.toFixed(0) should round to the nearest integer. var NUM_DECIMALS = 1; roundIfNumber = function(value, numDecimals) { if (_.isUndefined(numDecimals)) { numDecimals = NUM_DECIMALS; } return _.isNumber(value)? value.toFixed(numDecimals) : value; }; getRoundedProp = function(propName, defaultValue, numDecimals) { var value = getPropValue(propName, defaultValue); // Don't try to round the default value, since it could already be a string. if (value === defaultValue) { return value; } return roundIfNumber(value, numDecimals); }; Template.registerHelper('getRoundedProp', getRoundedProp); // This is a bit of a hack to make it easier to use a template // like a controller object. Use it like: // {{#with data a=true b=1}}
Template.registerHelper('data', function(data) { return data.hash; }); Meteor.cb = function(e,r) { if (e) console.error(e); if(r) { if (_.isArray(r) && console.table) { console.table(r); } else { console.log(r); } } }; guiUtils = {}; guiUtils.makeMeteorReactiveTextBox = function(textInputQuery, meteorPropertyName, options) { options = options || {}; var cleanseFunction = options.cleanseFunction || parseFloat; var updateFunction = options.updateFunction || 'updateProperty'; var interacting = false; var propertySelector = { property: meteorPropertyName }; // Meteor Property selector // Meteor -> jQuery-ui var updateTextBoxFromMeteor = function() { // Query the relevent stage data source var tbQuery = Properties.findOne(propertySelector); // Do not update if user interaction is going on // Update the text box if the mongo property has changed if (interacting) { return; } if (tbQuery) { $(textInputQuery).val(tbQuery.value); } }; // Call this first when the slider is initialized, and when changes happen Tracker.autorun(updateTextBoxFromMeteor); // jQuery-ui -> Meteor var pushTextChangeToMeteor = function() { var currentValue = $(textInputQuery).val(); if (cleanseFunction) { currentValue = cleanseFunction(currentValue); } Meteor.call(updateFunction, meteorPropertyName, currentValue, options, Meteor.cb); // Also change it back to the parsed version $(textInputQuery).val(currentValue); }; // Now we register change handlers via the various jQuery whatnot // Register doing an update on a enter key $(textInputQuery).bind('keypress', function(event) { var code = (event.keyCode ? event.keyCode : event.which); //Enter keycode if (code == 13) { // Force focus-out $(textInputQuery).blur(); } }); // Make a change commit if the field loses focus $(textInputQuery).focusout(function(event) { pushTextChangeToMeteor(); interacting = false; }); // Disable updating when user is editing values $(textInputQuery).focusin(function(event) { interacting = true; }); }; // Functions to init sliders on gui, also registers Meteor Tracker guiUtils.makeMeteorPropSlider = function(searchId, meteorPropertyName, cleanFunc, updateMethodName) { // Flag for is the slider is being used var interacting = false; // Select and create the slider element // Also register the start and stip functions $(searchId).slider({ range: "min", step: 0.000001, min: 0.0, max: 60.0, start: function(event, ui) { interacting = true; }, stop: function(event, ui) { var sliderLocation = $(searchId).slider("value"); // Update the metor when the slider is dropped somewhere new Meteor.call(updateMethodName, meteorPropertyName, cleanFunc(sliderLocation), "web"); interacting = false; } }); // Closured function to update the slider defined based on meteor values Tracker.autorun(function() { // Query the relevent stage data source var valueQuery = Properties.findOne({ property: meteorPropertyName }); if (valueQuery && !interacting) { // Update the fucking slider $(searchId).slider("value", valueQuery.value); } }); }; // Initialize the image-extents slider (diff b/c it has two sliders) guiUtils.initImageExtentSlider = function(axisName) { var interacting = false; // name for the jQuery selector var sliderName = "#image" + axisName + "AreaSlider"; // Meteor property names var propertyNameMin = "slice_" + axisName + "_min"; var propertyNameMax = "slice_" + axisName + "_max"; // Init the slider with default values . . . // The values will later be updated by autorun function
// {{> controllerTemplate}} // {{/with}}
random_line_split
uiTools.js
PossibleKesmNames', function() { return KesmNames.find().map(function(doc) { return doc.kesmName; }); }); }); getAllDataForKesmName = function(kesmName) { var properties = {}; if (propertyCollections && _.has(propertyCollections, kesmName)) { _.each(propertyCollections[kesmName].find().fetch(), function(doc) { properties[doc.property] = doc.value; }); } var sample = Samples.findOne(properties['currentSampleID']) || null; var machine = Machines.findOne({ kesmName: kesmName }); var process; if (!! machine) { process = Processes.findOne({ className: { $in: ['KESMMain', 'KESM'] }, machineId: machine._id }); } else { machine = null; process = null; } var uri = "http://" + Meteor.settings.public.serverURIs[kesmName]; return { kesmName: kesmName, machine: machine, process: process, properties: properties, sample: sample, uri: uri }; }; Template.registerHelper('allDataForKesmName', getAllDataForKesmName); Template.registerHelper('odd', function(value) { return (value % 2 !== 0); }); Template.registerHelper('classForDot', function(active) { return active? 'green' : 'red'; }); toTitleCase = function(str) { return str.replace(/\w\S*/g, function(txt) { return txt.charAt(0).toUpperCase() + txt.substr(1).toLowerCase(); }); }; // Small Collection of utility functions to make the gui-widgits dance etc. AxisList = [{ axis: 'x', AXIS: 'X', index: 0 }, { axis: 'y', AXIS: 'Y', index: 1 }, { axis: 'z', AXIS: 'Z', index: 2 }]; Template.registerHelper('axes', function() { return AxisList; }); axisIndex = function(axisLetter) { return _.findWhere(AxisList, { axis: axisLetter.toLowerCase() }).index; }; var axisIsInFault = function(axis) { return getPropValue('stage_'+axis+'_fault', false); }; Template.registerHelper('colorClassForAxis', function(axisLetter) { var colorForAxis = { x: 'error', y: 'success', z: 'info', }; if (axisIsInFault(axisLetter)) { return 'fault'; } else { return colorForAxis[axisLetter]; } }); Template.registerHelper('isInFault', axisIsInFault); var scopeModes = [{ mode: 'maintenance' }, { mode: 'focus' }, { mode: 'slicing' }]; Template.registerHelper('scopeModes', function() { return scopeModes; }); Template.registerHelper('scopeIsInMode', function(mode) { var scopeMode = getPropValue('scopeMode', 'maintenance'); return scopeMode.value === mode; }); Template.registerHelper('changingModes', function() { var scopeMode = getPropValue('scopeMode', 'maintenance'); return scopeMode.value !== scopeMode.requestedValue; }); firmlyInSlicingMode = function() { var scopeMode = getProperty('scopeMode', 'maintenance'); return (scopeMode.value === 'slicing') && (scopeMode.value === scopeMode.requestedValue); }; slicingOrChanging = function() { var scopeMode = getProperty('scopeMode', 'maintenance'); return (scopeMode.value === 'slicing') || scopeMode.value !== scopeMode.requestedValue; }; Template.registerHelper('slicingOrChanging', slicingOrChanging); // Returns 'disabled' if the scope is in 'slicing' mode, or if it's changing modes. // This is used to disable buttons and inputs. disabledIfSlicingOrChanging = function() { return slicingOrChanging()? 'disabled' : ''; }; Template.registerHelper('disabledIfSlicingOrChanging', disabledIfSlicingOrChanging); Template.registerHelper('disabledUnlessSlicing', function() { return firmlyInSlicingMode()? '' : 'disabled'; }); // The default number of digits to display for floating point numbers. // If this is 0, then num.toFixed(0) should round to the nearest integer. var NUM_DECIMALS = 1; roundIfNumber = function(value, numDecimals) { if (_.isUndefined(numDecimals)) { numDecimals = NUM_DECIMALS; } return _.isNumber(value)? value.toFixed(numDecimals) : value; }; getRoundedProp = function(propName, defaultValue, numDecimals) { var value = getPropValue(propName, defaultValue); // Don't try to round the default value, since it could already be a string. if (value === defaultValue) { return value; } return roundIfNumber(value, numDecimals); }; Template.registerHelper('getRoundedProp', getRoundedProp); // This is a bit of a hack to make it easier to use a template // like a controller object. Use it like: // {{#with data a=true b=1}} // {{> controllerTemplate}} // {{/with}} Template.registerHelper('data', function(data) { return data.hash; }); Meteor.cb = function(e,r) { if (e) console.error(e); if(r) { if (_.isArray(r) && console.table) { console.table(r); } else { console.log(r); } } }; guiUtils = {}; guiUtils.makeMeteorReactiveTextBox = function(textInputQuery, meteorPropertyName, options) { options = options || {}; var cleanseFunction = options.cleanseFunction || parseFloat; var updateFunction = options.updateFunction || 'updateProperty'; var interacting = false; var propertySelector = { property: meteorPropertyName }; // Meteor Property selector // Meteor -> jQuery-ui var updateTextBoxFromMeteor = function() { // Query the relevent stage data source var tbQuery = Properties.findOne(propertySelector); // Do not update if user interaction is going on // Update the text box if the mongo property has changed if (interacting)
if (tbQuery) { $(textInputQuery).val(tbQuery.value); } }; // Call this first when the slider is initialized, and when changes happen Tracker.autorun(updateTextBoxFromMeteor); // jQuery-ui -> Meteor var pushTextChangeToMeteor = function() { var currentValue = $(textInputQuery).val(); if (cleanseFunction) { currentValue = cleanseFunction(currentValue); } Meteor.call(updateFunction, meteorPropertyName, currentValue, options, Meteor.cb); // Also change it back to the parsed version $(textInputQuery).val(currentValue); }; // Now we register change handlers via the various jQuery whatnot // Register doing an update on a enter key $(textInputQuery).bind('keypress', function(event) { var code = (event.keyCode ? event.keyCode : event.which); //Enter keycode if (code == 13) { // Force focus-out $(textInputQuery).blur(); } }); // Make a change commit if the field loses focus $(textInputQuery).focusout(function(event) { pushTextChangeToMeteor(); interacting = false; }); // Disable updating when user is editing values $(textInputQuery).focusin(function(event) { interacting = true; }); }; // Functions to init sliders on gui, also registers Meteor Tracker guiUtils.makeMeteorPropSlider = function(searchId, meteorPropertyName, cleanFunc, updateMethodName) { // Flag for is the slider is being used var interacting = false; // Select and create the slider element // Also register the start and stip functions $(searchId).slider({ range: "min", step: 0.000001, min: 0.0, max: 60.0, start: function(event, ui) { interacting = true; }, stop: function(event, ui) { var sliderLocation = $(searchId).slider("value"); // Update the metor when the slider is dropped somewhere new Meteor.call(updateMethodName, meteorPropertyName, cleanFunc(sliderLocation), "web"); interacting = false; } }); // Closured function to update the slider defined based on meteor values Tracker.autorun(function() { // Query the relevent stage data source var valueQuery = Properties.findOne({ property: meteorPropertyName }); if (valueQuery && !interacting) { // Update the fucking slider $(searchId).slider("value", valueQuery.value); } }); }; // Initialize the image-extents slider (diff b/c it has two sliders) guiUtils.initImageExtentSlider = function(axisName) { var interacting = false; // name for the jQuery selector var sliderName = "#image" + axisName + "AreaSlider"; // Meteor property names var propertyNameMin = "slice_" + axisName + "_min"; var propertyNameMax = "slice_" + axisName + "_max"; // Init the slider with default values . . . // The values will later be updated by autorun
{ return; }
conditional_block
BatteryMonitor6813Tester.go
TC6813.New(spiConnection, 2).Initialise() // // } ltc_lock.Lock() defer ltc_lock.Unlock() for
chainLength-- if chainLength > 0 { ltc = LTC6813.New(spiConnection, chainLength) if err := ltc.Initialise(); err != nil { fmt.Print(err) log.Fatal(err) } _, err := ltc.MeasureVoltages() if err != nil { fmt.Println("MeasureVoltages - ", err) } _, err = ltc.MeasureTemperatures() if err != nil { fmt.Println("MeasureTemperatures - ", err) } } else { ltc = nil } return chainLength, nil } func performMeasurements() { var total float32 var err error var banks int fmt.Println("Measuring") if nDevices == 0 { nDevices, err = getLTC6813() if err != nil { fmt.Print(err) nErrors++ return } } if nDevices == 0 { fmt.Printf("\033cNo devices found on %s - %s", *spiDevice, time.Now().Format("15:04:05.99")) return } banks, err = ltc.MeasureVoltagesSC() if err != nil { // Retry if it failed and ignore the failure if the retry was successful banks, err = ltc.MeasureVoltagesSC() } if err != nil { fmt.Print(" Error measuring voltages - ", err) time.Sleep(time.Second * 2) nDevices = 0 nErrors++ } else { fmt.Print("\033c") fmt.Printf("%d LTC6813 found on %s - %s - %d errors.\n", nDevices, *spiDevice, time.Now().Format("15:04:05.99"), nErrors) for bank := 0; bank < banks; bank++ { fmt.Printf("Bank%2d", bank) total = 0.0 for cell := 0; cell < 18; cell++ { fmt.Printf(" : %1.4f", ltc.GetVolts(bank, cell)) total = total + ltc.GetVolts(bank, cell) } fmt.Printf(" Sum = %2.3f\n", total) } banks, err = ltc.MeasureTemperatures() if err != nil { banks, err = ltc.MeasureTemperatures() } if err != nil { fmt.Print(" Error measuring temperatures - ", err) time.Sleep(time.Second * 2) nDevices = 0 nErrors++ } fmt.Println("Temperatures") for bank := 0; bank < banks; bank++ { fmt.Printf("Bank %d", bank) for sensor := 0; sensor < 18; sensor++ { temperature, err := ltc.GetTemperature(bank, sensor) if err != nil { fmt.Printf(" : %v ", err) } else { fmt.Printf(" : %2.1f℃", temperature) } } fmt.Printf(" - Reference Volts = %1.4f - Sum of Cells = %2.3f\n", ltc.GetRefVolts(bank), ltc.GetSumOfCellsVolts(bank)) } } } func mainImpl() error { if !*verbose { log.SetOutput(ioutil.Discard) } log.SetFlags(log.Lmicroseconds) if flag.NArg() != 0 { return errors.New("unexpected argument, try -help") } for { nDevices, err := getLTC6813() if err == nil && nDevices > 0 { break } fmt.Println("Looking for a device") } done := make(chan bool) ticker := time.NewTicker(time.Second) go func() { for { select { case <-done: return case <-ticker.C: performMeasurements() } } }() // Configure and start the WEB server router := mux.NewRouter().StrictSlash(true) router.HandleFunc("/", getValues).Methods("GET") router.HandleFunc("/version", getVersion).Methods("GET") router.HandleFunc("/i2cread", getI2Cread).Methods("GET") router.HandleFunc("/i2cwrite", getI2Cwrite).Methods("GET") router.HandleFunc("/i2creadByte", getI2CreadByte).Methods("GET") router.HandleFunc("/i2cVoltage", getI2CVoltage).Methods("GET") router.HandleFunc("/i2cCharge", getI2CCharge).Methods("GET") router.HandleFunc("/i2cCurrent", getI2CCurrent).Methods("GET") router.HandleFunc("/i2cTemp", getI2CTemp).Methods("GET") http.ListenAndServe(":8080", router) // Listen on port 8080 return nil } func connectToDatabase() (*sql.DB, error) { if pDB != nil { _ = pDB.Close() pDB = nil } var sConnectionString = *pDatabaseLogin + ":" + *pDatabasePassword + "@tcp(" + *pDatabaseServer + ":" + *pDatabasePort + ")/" + *pDatabaseName fmt.Println("Connecting to [", sConnectionString, "]") db, err := sql.Open("mysql", sConnectionString) if err != nil { return nil, err } err = db.Ping() if err != nil { _ = db.Close() return nil, err } return db, err } func init() { verbose = flag.Bool("v", false, "verbose mode") spiDevice = flag.String("c", "/dev/spidev0.1", "SPI device from /dev") pDatabaseLogin = flag.String("l", "logger", "Database Login ID") pDatabasePassword = flag.String("p", "logger", "Database password") pDatabaseServer = flag.String("s", "localhost", "Database server") pDatabasePort = flag.String("o", "3306", "Database port") pDatabaseName = flag.String("d", "battery", "Name of the database") flag.Parse() logwriter, e := syslog.New(syslog.LOG_NOTICE, "Hello") if e == nil { log.SetOutput(logwriter) } // Initialise the SPI subsystem if _, err := host.Init(); err != nil { log.Fatal(err) } p, err := spireg.Open(*spiDevice) if err != nil { log.Fatal(err) } spiConnection, err = p.Connect(SPI_BAUD_RATE, spi.Mode0, SPI_BITS_PER_WORD) if err != nil { log.Fatal(err) } nErrors = 0 // Set up the database connection pDB, err = connectToDatabase() if err != nil { log.Fatalf("Failed to connect to to the database - %s - Sorry, I am giving up.", err) } } /* WEB Service to return the version information */ func getVersion(w http.ResponseWriter, r *http.Request) { w.Header().Set("Access-Control-Allow-Origin", "*") fmt.Fprint(w, `<html> <head> <Cedar Technology Battery Manager> </head> <body> <h1>Cedar Technology Battery Manager</h1> <h2>Version 1.0 - January 10th 2020</h2> </body> </html>`) } /* WEB service to return current process values */ func getValues(w http.ResponseWriter, _ *http.Request) { ltc_lock.Lock() defer ltc_lock.Unlock() // This header allows the output to be used in a WEB page from another server as a data source for some controls w.Header().Set("Access-Control-Allow-Origin", "*") if ltc != nil { _, _ = fmt.Fprintf(w, `{%s,%s}`, ltc.GetVoltagesAsJSON(), ltc.GetTemperaturesAsJSON()) } else { _, _ = fmt.Fprint(w, `{"error":"No Devices"}`) } } /** WEB service to read the I2C port */ func getI2Cread(w http.ResponseWriter, r *http.Request) { var reg int64 sReg := r.URL.Query().Get("reg") if sReg != "" { reg, _ = strconv.ParseInt(sReg, 0, 8) } else { reg = 0x1a } sensor, _ := strconv.ParseInt(r.URL.Query().Get("sensor"), 0, 8) s, err := ltc.ReadI2CWord(int(sensor), LTC6813.LTC2944Address, uint8(reg)) w.Header().Set("Access-Control-Allow-Origin", "*") fmt.Fprint(w, "Request = ", r.URL.Query().Get("reg"), "\n") if err !=
{ testLtc := LTC6813.New(spiConnection, chainLength) if _, err := testLtc.Test(); err != nil { fmt.Println(err) break } testLtc = nil chainLength++ }
conditional_block
BatteryMonitor6813Tester.go
{ nDevices, err = getLTC6813() if err != nil { fmt.Print(err) nErrors++ return } } if nDevices == 0 { fmt.Printf("\033cNo devices found on %s - %s", *spiDevice, time.Now().Format("15:04:05.99")) return } banks, err = ltc.MeasureVoltagesSC() if err != nil { // Retry if it failed and ignore the failure if the retry was successful banks, err = ltc.MeasureVoltagesSC() } if err != nil { fmt.Print(" Error measuring voltages - ", err) time.Sleep(time.Second * 2) nDevices = 0 nErrors++ } else { fmt.Print("\033c") fmt.Printf("%d LTC6813 found on %s - %s - %d errors.\n", nDevices, *spiDevice, time.Now().Format("15:04:05.99"), nErrors) for bank := 0; bank < banks; bank++ { fmt.Printf("Bank%2d", bank) total = 0.0 for cell := 0; cell < 18; cell++ { fmt.Printf(" : %1.4f", ltc.GetVolts(bank, cell)) total = total + ltc.GetVolts(bank, cell) } fmt.Printf(" Sum = %2.3f\n", total) } banks, err = ltc.MeasureTemperatures() if err != nil { banks, err = ltc.MeasureTemperatures() } if err != nil { fmt.Print(" Error measuring temperatures - ", err) time.Sleep(time.Second * 2) nDevices = 0 nErrors++ } fmt.Println("Temperatures") for bank := 0; bank < banks; bank++ { fmt.Printf("Bank %d", bank) for sensor := 0; sensor < 18; sensor++ { temperature, err := ltc.GetTemperature(bank, sensor) if err != nil { fmt.Printf(" : %v ", err) } else { fmt.Printf(" : %2.1f℃", temperature) } } fmt.Printf(" - Reference Volts = %1.4f - Sum of Cells = %2.3f\n", ltc.GetRefVolts(bank), ltc.GetSumOfCellsVolts(bank)) } } } func mainImpl() error { if !*verbose { log.SetOutput(ioutil.Discard) } log.SetFlags(log.Lmicroseconds) if flag.NArg() != 0 { return errors.New("unexpected argument, try -help") } for { nDevices, err := getLTC6813() if err == nil && nDevices > 0 { break } fmt.Println("Looking for a device") } done := make(chan bool) ticker := time.NewTicker(time.Second) go func() { for { select { case <-done: return case <-ticker.C: performMeasurements() } } }() // Configure and start the WEB server router := mux.NewRouter().StrictSlash(true) router.HandleFunc("/", getValues).Methods("GET") router.HandleFunc("/version", getVersion).Methods("GET") router.HandleFunc("/i2cread", getI2Cread).Methods("GET") router.HandleFunc("/i2cwrite", getI2Cwrite).Methods("GET") router.HandleFunc("/i2creadByte", getI2CreadByte).Methods("GET") router.HandleFunc("/i2cVoltage", getI2CVoltage).Methods("GET") router.HandleFunc("/i2cCharge", getI2CCharge).Methods("GET") router.HandleFunc("/i2cCurrent", getI2CCurrent).Methods("GET") router.HandleFunc("/i2cTemp", getI2CTemp).Methods("GET") http.ListenAndServe(":8080", router) // Listen on port 8080 return nil } func connectToDatabase() (*sql.DB, error) { if pDB != nil { _ = pDB.Close() pDB = nil } var sConnectionString = *pDatabaseLogin + ":" + *pDatabasePassword + "@tcp(" + *pDatabaseServer + ":" + *pDatabasePort + ")/" + *pDatabaseName fmt.Println("Connecting to [", sConnectionString, "]") db, err := sql.Open("mysql", sConnectionString) if err != nil { return nil, err } err = db.Ping() if err != nil { _ = db.Close() return nil, err } return db, err } func init() { verbose = flag.Bool("v", false, "verbose mode") spiDevice = flag.String("c", "/dev/spidev0.1", "SPI device from /dev") pDatabaseLogin = flag.String("l", "logger", "Database Login ID") pDatabasePassword = flag.String("p", "logger", "Database password") pDatabaseServer = flag.String("s", "localhost", "Database server") pDatabasePort = flag.String("o", "3306", "Database port") pDatabaseName = flag.String("d", "battery", "Name of the database") flag.Parse() logwriter, e := syslog.New(syslog.LOG_NOTICE, "Hello") if e == nil { log.SetOutput(logwriter) } // Initialise the SPI subsystem if _, err := host.Init(); err != nil { log.Fatal(err) } p, err := spireg.Open(*spiDevice) if err != nil { log.Fatal(err) } spiConnection, err = p.Connect(SPI_BAUD_RATE, spi.Mode0, SPI_BITS_PER_WORD) if err != nil { log.Fatal(err) } nErrors = 0 // Set up the database connection pDB, err = connectToDatabase() if err != nil { log.Fatalf("Failed to connect to to the database - %s - Sorry, I am giving up.", err) } } /* WEB Service to return the version information */ func getVersion(w http.ResponseWriter, r *http.Request) { w.Header().Set("Access-Control-Allow-Origin", "*") fmt.Fprint(w, `<html> <head> <Cedar Technology Battery Manager> </head> <body> <h1>Cedar Technology Battery Manager</h1> <h2>Version 1.0 - January 10th 2020</h2> </body> </html>`) } /* WEB service to return current process values */ func getValues(w http.ResponseWriter, _ *http.Request) { ltc_lock.Lock() defer ltc_lock.Unlock() // This header allows the output to be used in a WEB page from another server as a data source for some controls w.Header().Set("Access-Control-Allow-Origin", "*") if ltc != nil { _, _ = fmt.Fprintf(w, `{%s,%s}`, ltc.GetVoltagesAsJSON(), ltc.GetTemperaturesAsJSON()) } else { _, _ = fmt.Fprint(w, `{"error":"No Devices"}`) } } /** WEB service to read the I2C port */ func getI2Cread(w http.ResponseWriter, r *http.Request) { var reg int64 sReg := r.URL.Query().Get("reg") if sReg != "" { reg, _ = strconv.ParseInt(sReg, 0, 8) } else { reg = 0x1a } sensor, _ := strconv.ParseInt(r.URL.Query().Get("sensor"), 0, 8) s, err := ltc.ReadI2CWord(int(sensor), LTC6813.LTC2944Address, uint8(reg)) w.Header().Set("Access-Control-Allow-Origin", "*") fmt.Fprint(w, "Request = ", r.URL.Query().Get("reg"), "\n") if err != nil { fmt.Fprint(w, "Error - ", err) } else { fmt.Fprintf(w, s) } } /** WEB service to read one 8 bit register from the I2C port */ func getI2CreadByte(w http.ResponseWriter, r *http.Request) { sensor, _ := strconv.ParseInt(r.URL.Query().Get("sensor"), 0, 8) var reg int64 sReg := r.URL.Query().Get("reg") if sReg != "" { reg, _ = strconv.ParseInt(sReg, 0, 8) } else { reg = 0x1a } s, err := ltc.ReadI2CByte(int(sensor), LTC6813.LTC2944Address, uint8(reg)) w.Header().Set("Access-Control-Allow-Origin", "*") fmt.Fprint(w, "Request = ", r.URL.Query().Get("reg"), "\n") if err != nil { fmt.Fprint(w, "Error - ", err) } else { fmt.Fprintf(w, s) } }
/** WEB service to read the current from the I2C port */ func getI2CCurrent(w http.ResponseWriter, r *http.Request) {
random_line_split
BatteryMonitor6813Tester.go
TC6813.New(spiConnection, 2).Initialise() // // } ltc_lock.Lock() defer ltc_lock.Unlock() for { testLtc := LTC6813.New(spiConnection, chainLength) if _, err := testLtc.Test(); err != nil { fmt.Println(err) break } testLtc = nil chainLength++ } chainLength-- if chainLength > 0 { ltc = LTC6813.New(spiConnection, chainLength) if err := ltc.Initialise(); err != nil { fmt.Print(err) log.Fatal(err) } _, err := ltc.MeasureVoltages() if err != nil { fmt.Println("MeasureVoltages - ", err) } _, err = ltc.MeasureTemperatures() if err != nil { fmt.Println("MeasureTemperatures - ", err) } } else { ltc = nil } return chainLength, nil } func performMeasurements() { var total float32 var err error var banks int fmt.Println("Measuring") if nDevices == 0 { nDevices, err = getLTC6813() if err != nil { fmt.Print(err) nErrors++ return } } if nDevices == 0 { fmt.Printf("\033cNo devices found on %s - %s", *spiDevice, time.Now().Format("15:04:05.99")) return } banks, err = ltc.MeasureVoltagesSC() if err != nil { // Retry if it failed and ignore the failure if the retry was successful banks, err = ltc.MeasureVoltagesSC() } if err != nil { fmt.Print(" Error measuring voltages - ", err) time.Sleep(time.Second * 2) nDevices = 0 nErrors++ } else { fmt.Print("\033c") fmt.Printf("%d LTC6813 found on %s - %s - %d errors.\n", nDevices, *spiDevice, time.Now().Format("15:04:05.99"), nErrors) for bank := 0; bank < banks; bank++ { fmt.Printf("Bank%2d", bank) total = 0.0 for cell := 0; cell < 18; cell++ { fmt.Printf(" : %1.4f", ltc.GetVolts(bank, cell)) total = total + ltc.GetVolts(bank, cell) } fmt.Printf(" Sum = %2.3f\n", total) } banks, err = ltc.MeasureTemperatures() if err != nil { banks, err = ltc.MeasureTemperatures() } if err != nil { fmt.Print(" Error measuring temperatures - ", err) time.Sleep(time.Second * 2) nDevices = 0 nErrors++ } fmt.Println("Temperatures") for bank := 0; bank < banks; bank++ { fmt.Printf("Bank %d", bank) for sensor := 0; sensor < 18; sensor++ { temperature, err := ltc.GetTemperature(bank, sensor) if err != nil { fmt.Printf(" : %v ", err) } else { fmt.Printf(" : %2.1f℃", temperature) } } fmt.Printf(" - Reference Volts = %1.4f - Sum of Cells = %2.3f\n", ltc.GetRefVolts(bank), ltc.GetSumOfCellsVolts(bank)) } } } func mainImpl() error { if !*verbose { log.SetOutput(ioutil.Discard) } log.SetFlags(log.Lmicroseconds) if flag.NArg() != 0 { return errors.New("unexpected argument, try -help") } for { nDevices, err := getLTC6813() if err == nil && nDevices > 0 { break } fmt.Println("Looking for a device") } done := make(chan bool) ticker := time.NewTicker(time.Second) go func() { for { select { case <-done: return case <-ticker.C: performMeasurements() } } }() // Configure and start the WEB server router := mux.NewRouter().StrictSlash(true) router.HandleFunc("/", getValues).Methods("GET") router.HandleFunc("/version", getVersion).Methods("GET") router.HandleFunc("/i2cread", getI2Cread).Methods("GET") router.HandleFunc("/i2cwrite", getI2Cwrite).Methods("GET") router.HandleFunc("/i2creadByte", getI2CreadByte).Methods("GET") router.HandleFunc("/i2cVoltage", getI2CVoltage).Methods("GET") router.HandleFunc("/i2cCharge", getI2CCharge).Methods("GET") router.HandleFunc("/i2cCurrent", getI2CCurrent).Methods("GET") router.HandleFunc("/i2cTemp", getI2CTemp).Methods("GET") http.ListenAndServe(":8080", router) // Listen on port 8080 return nil } func co
(*sql.DB, error) { if pDB != nil { _ = pDB.Close() pDB = nil } var sConnectionString = *pDatabaseLogin + ":" + *pDatabasePassword + "@tcp(" + *pDatabaseServer + ":" + *pDatabasePort + ")/" + *pDatabaseName fmt.Println("Connecting to [", sConnectionString, "]") db, err := sql.Open("mysql", sConnectionString) if err != nil { return nil, err } err = db.Ping() if err != nil { _ = db.Close() return nil, err } return db, err } func init() { verbose = flag.Bool("v", false, "verbose mode") spiDevice = flag.String("c", "/dev/spidev0.1", "SPI device from /dev") pDatabaseLogin = flag.String("l", "logger", "Database Login ID") pDatabasePassword = flag.String("p", "logger", "Database password") pDatabaseServer = flag.String("s", "localhost", "Database server") pDatabasePort = flag.String("o", "3306", "Database port") pDatabaseName = flag.String("d", "battery", "Name of the database") flag.Parse() logwriter, e := syslog.New(syslog.LOG_NOTICE, "Hello") if e == nil { log.SetOutput(logwriter) } // Initialise the SPI subsystem if _, err := host.Init(); err != nil { log.Fatal(err) } p, err := spireg.Open(*spiDevice) if err != nil { log.Fatal(err) } spiConnection, err = p.Connect(SPI_BAUD_RATE, spi.Mode0, SPI_BITS_PER_WORD) if err != nil { log.Fatal(err) } nErrors = 0 // Set up the database connection pDB, err = connectToDatabase() if err != nil { log.Fatalf("Failed to connect to to the database - %s - Sorry, I am giving up.", err) } } /* WEB Service to return the version information */ func getVersion(w http.ResponseWriter, r *http.Request) { w.Header().Set("Access-Control-Allow-Origin", "*") fmt.Fprint(w, `<html> <head> <Cedar Technology Battery Manager> </head> <body> <h1>Cedar Technology Battery Manager</h1> <h2>Version 1.0 - January 10th 2020</h2> </body> </html>`) } /* WEB service to return current process values */ func getValues(w http.ResponseWriter, _ *http.Request) { ltc_lock.Lock() defer ltc_lock.Unlock() // This header allows the output to be used in a WEB page from another server as a data source for some controls w.Header().Set("Access-Control-Allow-Origin", "*") if ltc != nil { _, _ = fmt.Fprintf(w, `{%s,%s}`, ltc.GetVoltagesAsJSON(), ltc.GetTemperaturesAsJSON()) } else { _, _ = fmt.Fprint(w, `{"error":"No Devices"}`) } } /** WEB service to read the I2C port */ func getI2Cread(w http.ResponseWriter, r *http.Request) { var reg int64 sReg := r.URL.Query().Get("reg") if sReg != "" { reg, _ = strconv.ParseInt(sReg, 0, 8) } else { reg = 0x1a } sensor, _ := strconv.ParseInt(r.URL.Query().Get("sensor"), 0, 8) s, err := ltc.ReadI2CWord(int(sensor), LTC6813.LTC2944Address, uint8(reg)) w.Header().Set("Access-Control-Allow-Origin", "*") fmt.Fprint(w, "Request = ", r.URL.Query().Get("reg"), "\n") if err !=
nnectToDatabase()
identifier_name
BatteryMonitor6813Tester.go
LTC6813.New(spiConnection, 2).Initialise() // // } ltc_lock.Lock() defer ltc_lock.Unlock() for { testLtc := LTC6813.New(spiConnection, chainLength) if _, err := testLtc.Test(); err != nil { fmt.Println(err) break } testLtc = nil chainLength++ } chainLength-- if chainLength > 0 { ltc = LTC6813.New(spiConnection, chainLength) if err := ltc.Initialise(); err != nil { fmt.Print(err) log.Fatal(err) } _, err := ltc.MeasureVoltages() if err != nil { fmt.Println("MeasureVoltages - ", err) } _, err = ltc.MeasureTemperatures() if err != nil { fmt.Println("MeasureTemperatures - ", err) } } else { ltc = nil } return chainLength, nil } func performMeasurements() { var total float32 var err error var banks int fmt.Println("Measuring") if nDevices == 0 { nDevices, err = getLTC6813() if err != nil { fmt.Print(err) nErrors++ return } } if nDevices == 0 { fmt.Printf("\033cNo devices found on %s - %s", *spiDevice, time.Now().Format("15:04:05.99")) return } banks, err = ltc.MeasureVoltagesSC() if err != nil { // Retry if it failed and ignore the failure if the retry was successful banks, err = ltc.MeasureVoltagesSC() } if err != nil { fmt.Print(" Error measuring voltages - ", err) time.Sleep(time.Second * 2) nDevices = 0 nErrors++ } else { fmt.Print("\033c") fmt.Printf("%d LTC6813 found on %s - %s - %d errors.\n", nDevices, *spiDevice, time.Now().Format("15:04:05.99"), nErrors) for bank := 0; bank < banks; bank++ { fmt.Printf("Bank%2d", bank) total = 0.0 for cell := 0; cell < 18; cell++ { fmt.Printf(" : %1.4f", ltc.GetVolts(bank, cell)) total = total + ltc.GetVolts(bank, cell) } fmt.Printf(" Sum = %2.3f\n", total) } banks, err = ltc.MeasureTemperatures() if err != nil { banks, err = ltc.MeasureTemperatures() } if err != nil { fmt.Print(" Error measuring temperatures - ", err) time.Sleep(time.Second * 2) nDevices = 0 nErrors++ } fmt.Println("Temperatures") for bank := 0; bank < banks; bank++ { fmt.Printf("Bank %d", bank) for sensor := 0; sensor < 18; sensor++ { temperature, err := ltc.GetTemperature(bank, sensor) if err != nil { fmt.Printf(" : %v ", err) } else { fmt.Printf(" : %2.1f℃", temperature) } } fmt.Printf(" - Reference Volts = %1.4f - Sum of Cells = %2.3f\n", ltc.GetRefVolts(bank), ltc.GetSumOfCellsVolts(bank)) } } } func mainImpl() error {
select { case <-done: return case <-ticker.C: performMeasurements() } } }() // Configure and start the WEB server router := mux.NewRouter().StrictSlash(true) router.HandleFunc("/", getValues).Methods("GET") router.HandleFunc("/version", getVersion).Methods("GET") router.HandleFunc("/i2cread", getI2Cread).Methods("GET") router.HandleFunc("/i2cwrite", getI2Cwrite).Methods("GET") router.HandleFunc("/i2creadByte", getI2CreadByte).Methods("GET") router.HandleFunc("/i2cVoltage", getI2CVoltage).Methods("GET") router.HandleFunc("/i2cCharge", getI2CCharge).Methods("GET") router.HandleFunc("/i2cCurrent", getI2CCurrent).Methods("GET") router.HandleFunc("/i2cTemp", getI2CTemp).Methods("GET") http.ListenAndServe(":8080", router) // Listen on port 8080 return nil } func connectToDatabase() (*sql.DB, error) { if pDB != nil { _ = pDB.Close() pDB = nil } var sConnectionString = *pDatabaseLogin + ":" + *pDatabasePassword + "@tcp(" + *pDatabaseServer + ":" + *pDatabasePort + ")/" + *pDatabaseName fmt.Println("Connecting to [", sConnectionString, "]") db, err := sql.Open("mysql", sConnectionString) if err != nil { return nil, err } err = db.Ping() if err != nil { _ = db.Close() return nil, err } return db, err } func init() { verbose = flag.Bool("v", false, "verbose mode") spiDevice = flag.String("c", "/dev/spidev0.1", "SPI device from /dev") pDatabaseLogin = flag.String("l", "logger", "Database Login ID") pDatabasePassword = flag.String("p", "logger", "Database password") pDatabaseServer = flag.String("s", "localhost", "Database server") pDatabasePort = flag.String("o", "3306", "Database port") pDatabaseName = flag.String("d", "battery", "Name of the database") flag.Parse() logwriter, e := syslog.New(syslog.LOG_NOTICE, "Hello") if e == nil { log.SetOutput(logwriter) } // Initialise the SPI subsystem if _, err := host.Init(); err != nil { log.Fatal(err) } p, err := spireg.Open(*spiDevice) if err != nil { log.Fatal(err) } spiConnection, err = p.Connect(SPI_BAUD_RATE, spi.Mode0, SPI_BITS_PER_WORD) if err != nil { log.Fatal(err) } nErrors = 0 // Set up the database connection pDB, err = connectToDatabase() if err != nil { log.Fatalf("Failed to connect to to the database - %s - Sorry, I am giving up.", err) } } /* WEB Service to return the version information */ func getVersion(w http.ResponseWriter, r *http.Request) { w.Header().Set("Access-Control-Allow-Origin", "*") fmt.Fprint(w, `<html> <head> <Cedar Technology Battery Manager> </head> <body> <h1>Cedar Technology Battery Manager</h1> <h2>Version 1.0 - January 10th 2020</h2> </body> </html>`) } /* WEB service to return current process values */ func getValues(w http.ResponseWriter, _ *http.Request) { ltc_lock.Lock() defer ltc_lock.Unlock() // This header allows the output to be used in a WEB page from another server as a data source for some controls w.Header().Set("Access-Control-Allow-Origin", "*") if ltc != nil { _, _ = fmt.Fprintf(w, `{%s,%s}`, ltc.GetVoltagesAsJSON(), ltc.GetTemperaturesAsJSON()) } else { _, _ = fmt.Fprint(w, `{"error":"No Devices"}`) } } /** WEB service to read the I2C port */ func getI2Cread(w http.ResponseWriter, r *http.Request) { var reg int64 sReg := r.URL.Query().Get("reg") if sReg != "" { reg, _ = strconv.ParseInt(sReg, 0, 8) } else { reg = 0x1a } sensor, _ := strconv.ParseInt(r.URL.Query().Get("sensor"), 0, 8) s, err := ltc.ReadI2CWord(int(sensor), LTC6813.LTC2944Address, uint8(reg)) w.Header().Set("Access-Control-Allow-Origin", "*") fmt.Fprint(w, "Request = ", r.URL.Query().Get("reg"), "\n") if err != nil
if !*verbose { log.SetOutput(ioutil.Discard) } log.SetFlags(log.Lmicroseconds) if flag.NArg() != 0 { return errors.New("unexpected argument, try -help") } for { nDevices, err := getLTC6813() if err == nil && nDevices > 0 { break } fmt.Println("Looking for a device") } done := make(chan bool) ticker := time.NewTicker(time.Second) go func() { for {
identifier_body
image_caption.py
# Loading cap as values and images as key in dictionary tok = {} for item in range(len(cap)-1): tem = cap[item].split("#") #tem[0]= imgname.jpg ..... tem[1]=0 captionn. if tem[0] in tok: tok[tem[0]].append(tem[1][2:]) else: tok[tem[0]] = [tem[1][2:]] #tem[n]= imgName ... #tok[tem[n]] = list of caption # Making 3 files with 2 colmns as 'image_id' and 'captions' training_dataset = open('flickr_8k_train_dataset.txt','wb') training_dataset.write(b"image_id\tcap\n") valid_dataset = open('flickr_8k_val_dataset.txt','wb') valid_dataset.write(b"image_id\tcap\n") testing_dataset = open('flickr_8k_test_dataset.txt','wb') testing_dataset.write(b"image_id\tcap\n") # Loading image ids and captions for each of these images in the above 3 files for img in x_training: if img == '': continue for capt in tok[img]: caption = "<start> "+ capt + " <end>" training_dataset.write((img+"\t"+caption+"\n").encode()) training_dataset.flush() training_dataset.close() for img in x_testing: if img == '': continue for capt in tok[img]: caption = "<start> "+ capt + " <end>" testing_dataset.write((img+"\t"+caption+"\n").encode()) testing_dataset.flush() testing_dataset.close() for img in x_valid: if img == '': continue for capt in tok[img]: caption = "<start> "+ capt + " <end>" valid_dataset.write((img+"\t"+caption+"\n").encode()) valid_dataset.flush() valid_dataset.close() # Here, we're using ResNet50 Model from IPython.core.display import display, HTML model = ResNet50(include_top=False,weights='imagenet',input_shape=(224,224,3),pooling='avg') model.summary() # process images to target size def preprocess(img_path): im = image.load_img(img_path, target_size=(224,224,3)) im = image.img_to_array(im) # (x, y, z) im = np.expand_dims(im, axis=0) # (0, x, y, z) return im training_data = {} counter=0 for item in x_training: if item == "": continue if counter >= 3000: break counter+=1 if counter%1000==0: print(counter) path = img_path + item img = preprocess(path) #to change the dimensions of the image for using ResNet model pred = model.predict(img).reshape(2048) # shape of each image is (2048, 0) training_data[item] = pred # opening train_enc_img.p file and dumping content of training_data to this file with open( "train_enc_img.p", "wb" ) as pickle_f: #obj hierarchy is converted into byte stream pickle.dump(training_data, pickle_f ) # Storing image and its corresponding caption into a dataframe pd_dataset = pd.read_csv("flickr_8k_train_dataset.txt", delimiter='\t') dframe = pd_dataset.values print(dframe.shape) pd_dataset.head() # Storing all the captions from dframe into a list senten = [] for item in range(dframe.shape[0]): senten.append(dframe[item, 1]) #senten will have 30000 length # First 5 captions stored in senten senten[:5] # Splitting each captions stored in 'senten' and storing them in 'wor' as list of list wor = [i.split() for i in senten] # Creating a list of all unique wor uniq = [] for i in wor: uniq.extend(i) uniq = list(set(uniq)) print(len(uniq)) vocabulary_size = len(uniq) # making 2 lists to index each unique word and vice-versa w_to_i = {val:index for index, val in enumerate(uniq)} i_to_w = {index:val for index, val in enumerate(uniq)} w_to_i['UNK'] = 0 w_to_i['raining'] = 8253 i_to_w[0] = 'UNK' i_to_w[8253] = 'raining' vocabulary_size = len(w_to_i.keys()) print(vocabulary_size) max_len = 0 for i in senten: i = i.split() if len(i) > max_len: max_len = len(i) print(max_len) #finding longest caption pad_seq, subsequent_wor = [], [] for item in range(dframe.shape[0]): #30000 items part_seq = [] next_wor = [] text = dframe[item, 1].split() #diving each caption for every image into words text = [w_to_i[i] for i in text] #finding index for each word for i in range(1, len(text)): part_seq.append(text[:i]) #start, 1st word, ... , last word next_wor.append(text[i]) #1st word, ... , last word, end pad_part_seq = sequence.pad_sequences(part_seq, max_len, padding='post') next_wor_1hot = np.zeros([len(next_wor), vocabulary_size], dtype=np.bool) for i,next_word in enumerate(next_wor): next_wor_1hot[i, next_word] = 1 pad_seq.append(pad_part_seq ) subsequent_wor.append(next_wor_1hot) pad_seq = np.asarray(pad_seq) subsequent_wor = np.asarray(subsequent_wor) print(pad_seq.shape) print(subsequent_wor.shape) print(pad_seq[0]) for item in range(len(pad_seq[0])): for y in range(max_len): print(i_to_w[pad_seq[0][item][y]],) print("\n") print(len(pad_seq[0])) num_imgs = 2000 cap = np.zeros([0, max_len]) next_wor = np.zeros([0, vocabulary_size]) for item in range(num_imgs): #img_to_padded_seqs.shape[0]): cap = np.concatenate([cap, pad_seq[item]]) next_wor = np.concatenate([next_wor, subsequent_wor[item]]) np.save("cap.npy", cap) np.save("next_wor.npy", next_wor) print(cap.shape) print(next_wor.shape) with open('train_enc_img.p', 'rb') as f: enc_img = pickle.load(f, encoding="bytes") imgs = [] for item in range(dframe.shape[0]): #30000 if dframe[item, 0] in enc_img.keys(): #dframe[0,0], [1,0], ... , [4,0] match with 0th key of enc_img imgs.append(list(enc_img[dframe[item, 0]])) imgs = np.asarray(imgs) print(imgs.shape) images = [] img_names = [] for item in range(num_imgs): #2000 for y in range(pad_seq[item].shape[0]): #14 images.append(imgs[item]) #1st iteration: 14 times name of image in byte form img_names.append(dframe[item, 0]) # normal form images = np.asarray(images) #images contains image_name in byte form np.save("images.npy", images) img_names = np.asarray(img_names) #img_names contains image_name normally np.save("img_names.npy", img_names) print(images.shape) print(len(img_names)) cap = np.load("cap.npy") next_wor = np.load("next_wor.npy") print(cap.shape) print(next_wor.shape) images = np.load("images.npy") print(images.shape) imag = np.load("img_names.npy") print(imag.shape) embed_size = 128 max_len = 40 img_model = Sequential() img_model.add(Dense(embed_size, input_shape=(2048,), activation='relu')) img_model.add(RepeatVector(max_len)) img_model.summary() lang_model = Sequential() lang_model.add(Embedding(input_dim=vocabulary_size, output_dim=embed_size, input_length=max_len)) lang_model.add(LSTM(256, return_sequences=True)) lang_model.add(TimeDistributed(Dense(embed_size))) lang_model.summary() concat = Concatenate()([img_model.output, lang_model.output]) x = LSTM(128, return_sequences=True)(concat) x = LSTM(512, return_sequences=False)(x) x = Dense(vocabulary_size)(x) out = Activation('softmax')(x) model = Model(inputs=[img_model.input, lang_model.input], outputs = out) model.compile(loss='categorical_crossentropy', optimizer='RMS
x_training = open(training_path, 'r').read().split("\n") x_valid = open(valid_path, 'r').read().split("\n") x_testing = open(testing_path , 'r').read().split("\n")
random_line_split
image_caption.py
flickr_8k_val_dataset.txt','wb') valid_dataset.write(b"image_id\tcap\n") testing_dataset = open('flickr_8k_test_dataset.txt','wb') testing_dataset.write(b"image_id\tcap\n") # Loading image ids and captions for each of these images in the above 3 files for img in x_training: if img == '': continue for capt in tok[img]: caption = "<start> "+ capt + " <end>" training_dataset.write((img+"\t"+caption+"\n").encode()) training_dataset.flush() training_dataset.close() for img in x_testing: if img == '': continue for capt in tok[img]: caption = "<start> "+ capt + " <end>" testing_dataset.write((img+"\t"+caption+"\n").encode()) testing_dataset.flush() testing_dataset.close() for img in x_valid: if img == '': continue for capt in tok[img]: caption = "<start> "+ capt + " <end>" valid_dataset.write((img+"\t"+caption+"\n").encode()) valid_dataset.flush() valid_dataset.close() # Here, we're using ResNet50 Model from IPython.core.display import display, HTML model = ResNet50(include_top=False,weights='imagenet',input_shape=(224,224,3),pooling='avg') model.summary() # process images to target size def preprocess(img_path): im = image.load_img(img_path, target_size=(224,224,3)) im = image.img_to_array(im) # (x, y, z) im = np.expand_dims(im, axis=0) # (0, x, y, z) return im training_data = {} counter=0 for item in x_training: if item == "": continue if counter >= 3000: break counter+=1 if counter%1000==0: print(counter) path = img_path + item img = preprocess(path) #to change the dimensions of the image for using ResNet model pred = model.predict(img).reshape(2048) # shape of each image is (2048, 0) training_data[item] = pred # opening train_enc_img.p file and dumping content of training_data to this file with open( "train_enc_img.p", "wb" ) as pickle_f: #obj hierarchy is converted into byte stream pickle.dump(training_data, pickle_f ) # Storing image and its corresponding caption into a dataframe pd_dataset = pd.read_csv("flickr_8k_train_dataset.txt", delimiter='\t') dframe = pd_dataset.values print(dframe.shape) pd_dataset.head() # Storing all the captions from dframe into a list senten = [] for item in range(dframe.shape[0]): senten.append(dframe[item, 1]) #senten will have 30000 length # First 5 captions stored in senten senten[:5] # Splitting each captions stored in 'senten' and storing them in 'wor' as list of list wor = [i.split() for i in senten] # Creating a list of all unique wor uniq = [] for i in wor: uniq.extend(i) uniq = list(set(uniq)) print(len(uniq)) vocabulary_size = len(uniq) # making 2 lists to index each unique word and vice-versa w_to_i = {val:index for index, val in enumerate(uniq)} i_to_w = {index:val for index, val in enumerate(uniq)} w_to_i['UNK'] = 0 w_to_i['raining'] = 8253 i_to_w[0] = 'UNK' i_to_w[8253] = 'raining' vocabulary_size = len(w_to_i.keys()) print(vocabulary_size) max_len = 0 for i in senten: i = i.split() if len(i) > max_len: max_len = len(i) print(max_len) #finding longest caption pad_seq, subsequent_wor = [], [] for item in range(dframe.shape[0]): #30000 items part_seq = [] next_wor = [] text = dframe[item, 1].split() #diving each caption for every image into words text = [w_to_i[i] for i in text] #finding index for each word for i in range(1, len(text)): part_seq.append(text[:i]) #start, 1st word, ... , last word next_wor.append(text[i]) #1st word, ... , last word, end pad_part_seq = sequence.pad_sequences(part_seq, max_len, padding='post') next_wor_1hot = np.zeros([len(next_wor), vocabulary_size], dtype=np.bool) for i,next_word in enumerate(next_wor): next_wor_1hot[i, next_word] = 1 pad_seq.append(pad_part_seq ) subsequent_wor.append(next_wor_1hot) pad_seq = np.asarray(pad_seq) subsequent_wor = np.asarray(subsequent_wor) print(pad_seq.shape) print(subsequent_wor.shape) print(pad_seq[0]) for item in range(len(pad_seq[0])): for y in range(max_len): print(i_to_w[pad_seq[0][item][y]],) print("\n") print(len(pad_seq[0])) num_imgs = 2000 cap = np.zeros([0, max_len]) next_wor = np.zeros([0, vocabulary_size]) for item in range(num_imgs): #img_to_padded_seqs.shape[0]): cap = np.concatenate([cap, pad_seq[item]]) next_wor = np.concatenate([next_wor, subsequent_wor[item]]) np.save("cap.npy", cap) np.save("next_wor.npy", next_wor) print(cap.shape) print(next_wor.shape) with open('train_enc_img.p', 'rb') as f: enc_img = pickle.load(f, encoding="bytes") imgs = [] for item in range(dframe.shape[0]): #30000 if dframe[item, 0] in enc_img.keys(): #dframe[0,0], [1,0], ... , [4,0] match with 0th key of enc_img imgs.append(list(enc_img[dframe[item, 0]])) imgs = np.asarray(imgs) print(imgs.shape) images = [] img_names = [] for item in range(num_imgs): #2000 for y in range(pad_seq[item].shape[0]): #14 images.append(imgs[item]) #1st iteration: 14 times name of image in byte form img_names.append(dframe[item, 0]) # normal form images = np.asarray(images) #images contains image_name in byte form np.save("images.npy", images) img_names = np.asarray(img_names) #img_names contains image_name normally np.save("img_names.npy", img_names) print(images.shape) print(len(img_names)) cap = np.load("cap.npy") next_wor = np.load("next_wor.npy") print(cap.shape) print(next_wor.shape) images = np.load("images.npy") print(images.shape) imag = np.load("img_names.npy") print(imag.shape) embed_size = 128 max_len = 40 img_model = Sequential() img_model.add(Dense(embed_size, input_shape=(2048,), activation='relu')) img_model.add(RepeatVector(max_len)) img_model.summary() lang_model = Sequential() lang_model.add(Embedding(input_dim=vocabulary_size, output_dim=embed_size, input_length=max_len)) lang_model.add(LSTM(256, return_sequences=True)) lang_model.add(TimeDistributed(Dense(embed_size))) lang_model.summary() concat = Concatenate()([img_model.output, lang_model.output]) x = LSTM(128, return_sequences=True)(concat) x = LSTM(512, return_sequences=False)(x) x = Dense(vocabulary_size)(x) out = Activation('softmax')(x) model = Model(inputs=[img_model.input, lang_model.input], outputs = out) model.compile(loss='categorical_crossentropy', optimizer='RMSprop', metrics=['accuracy']) model.summary() hist = model.fit([images, cap], next_wor, batch_size=512, epochs=210) for label in ["loss"]: plt.plot(hist.history[label],label=label) plt.legend() plt.xlabel("epochs") plt.ylabel("loss") plt.show() for label in ["accuracy"]: plt.plot(hist.history[label],label=label) plt.legend() plt.xlabel("epochs") plt.ylabel("accuracy") plt.show() model.save_weights("model_weights.h5") def preprocess(img_path): im = image.load_img(img_path, target_size=(224,224,3)) im = image.img_to_array(im) #(224,224,3) im = np.expand_dims(im, axis=0) #(1,224,224,3) return im def get_encode(model, img):
image = preprocess(img) pred = model.predict(image).reshape(2048) return pred
identifier_body
image_caption.py
") x_valid = open(valid_path, 'r').read().split("\n") x_testing = open(testing_path , 'r').read().split("\n") # Loading cap as values and images as key in dictionary tok = {} for item in range(len(cap)-1): tem = cap[item].split("#") #tem[0]= imgname.jpg ..... tem[1]=0 captionn. if tem[0] in tok: tok[tem[0]].append(tem[1][2:]) else: tok[tem[0]] = [tem[1][2:]] #tem[n]= imgName ... #tok[tem[n]] = list of caption # Making 3 files with 2 colmns as 'image_id' and 'captions' training_dataset = open('flickr_8k_train_dataset.txt','wb') training_dataset.write(b"image_id\tcap\n") valid_dataset = open('flickr_8k_val_dataset.txt','wb') valid_dataset.write(b"image_id\tcap\n") testing_dataset = open('flickr_8k_test_dataset.txt','wb') testing_dataset.write(b"image_id\tcap\n") # Loading image ids and captions for each of these images in the above 3 files for img in x_training: if img == '': continue for capt in tok[img]: caption = "<start> "+ capt + " <end>" training_dataset.write((img+"\t"+caption+"\n").encode()) training_dataset.flush() training_dataset.close() for img in x_testing:
testing_dataset.close() for img in x_valid: if img == '': continue for capt in tok[img]: caption = "<start> "+ capt + " <end>" valid_dataset.write((img+"\t"+caption+"\n").encode()) valid_dataset.flush() valid_dataset.close() # Here, we're using ResNet50 Model from IPython.core.display import display, HTML model = ResNet50(include_top=False,weights='imagenet',input_shape=(224,224,3),pooling='avg') model.summary() # process images to target size def preprocess(img_path): im = image.load_img(img_path, target_size=(224,224,3)) im = image.img_to_array(im) # (x, y, z) im = np.expand_dims(im, axis=0) # (0, x, y, z) return im training_data = {} counter=0 for item in x_training: if item == "": continue if counter >= 3000: break counter+=1 if counter%1000==0: print(counter) path = img_path + item img = preprocess(path) #to change the dimensions of the image for using ResNet model pred = model.predict(img).reshape(2048) # shape of each image is (2048, 0) training_data[item] = pred # opening train_enc_img.p file and dumping content of training_data to this file with open( "train_enc_img.p", "wb" ) as pickle_f: #obj hierarchy is converted into byte stream pickle.dump(training_data, pickle_f ) # Storing image and its corresponding caption into a dataframe pd_dataset = pd.read_csv("flickr_8k_train_dataset.txt", delimiter='\t') dframe = pd_dataset.values print(dframe.shape) pd_dataset.head() # Storing all the captions from dframe into a list senten = [] for item in range(dframe.shape[0]): senten.append(dframe[item, 1]) #senten will have 30000 length # First 5 captions stored in senten senten[:5] # Splitting each captions stored in 'senten' and storing them in 'wor' as list of list wor = [i.split() for i in senten] # Creating a list of all unique wor uniq = [] for i in wor: uniq.extend(i) uniq = list(set(uniq)) print(len(uniq)) vocabulary_size = len(uniq) # making 2 lists to index each unique word and vice-versa w_to_i = {val:index for index, val in enumerate(uniq)} i_to_w = {index:val for index, val in enumerate(uniq)} w_to_i['UNK'] = 0 w_to_i['raining'] = 8253 i_to_w[0] = 'UNK' i_to_w[8253] = 'raining' vocabulary_size = len(w_to_i.keys()) print(vocabulary_size) max_len = 0 for i in senten: i = i.split() if len(i) > max_len: max_len = len(i) print(max_len) #finding longest caption pad_seq, subsequent_wor = [], [] for item in range(dframe.shape[0]): #30000 items part_seq = [] next_wor = [] text = dframe[item, 1].split() #diving each caption for every image into words text = [w_to_i[i] for i in text] #finding index for each word for i in range(1, len(text)): part_seq.append(text[:i]) #start, 1st word, ... , last word next_wor.append(text[i]) #1st word, ... , last word, end pad_part_seq = sequence.pad_sequences(part_seq, max_len, padding='post') next_wor_1hot = np.zeros([len(next_wor), vocabulary_size], dtype=np.bool) for i,next_word in enumerate(next_wor): next_wor_1hot[i, next_word] = 1 pad_seq.append(pad_part_seq ) subsequent_wor.append(next_wor_1hot) pad_seq = np.asarray(pad_seq) subsequent_wor = np.asarray(subsequent_wor) print(pad_seq.shape) print(subsequent_wor.shape) print(pad_seq[0]) for item in range(len(pad_seq[0])): for y in range(max_len): print(i_to_w[pad_seq[0][item][y]],) print("\n") print(len(pad_seq[0])) num_imgs = 2000 cap = np.zeros([0, max_len]) next_wor = np.zeros([0, vocabulary_size]) for item in range(num_imgs): #img_to_padded_seqs.shape[0]): cap = np.concatenate([cap, pad_seq[item]]) next_wor = np.concatenate([next_wor, subsequent_wor[item]]) np.save("cap.npy", cap) np.save("next_wor.npy", next_wor) print(cap.shape) print(next_wor.shape) with open('train_enc_img.p', 'rb') as f: enc_img = pickle.load(f, encoding="bytes") imgs = [] for item in range(dframe.shape[0]): #30000 if dframe[item, 0] in enc_img.keys(): #dframe[0,0], [1,0], ... , [4,0] match with 0th key of enc_img imgs.append(list(enc_img[dframe[item, 0]])) imgs = np.asarray(imgs) print(imgs.shape) images = [] img_names = [] for item in range(num_imgs): #2000 for y in range(pad_seq[item].shape[0]): #14 images.append(imgs[item]) #1st iteration: 14 times name of image in byte form img_names.append(dframe[item, 0]) # normal form images = np.asarray(images) #images contains image_name in byte form np.save("images.npy", images) img_names = np.asarray(img_names) #img_names contains image_name normally np.save("img_names.npy", img_names) print(images.shape) print(len(img_names)) cap = np.load("cap.npy") next_wor = np.load("next_wor.npy") print(cap.shape) print(next_wor.shape) images = np.load("images.npy") print(images.shape) imag = np.load("img_names.npy") print(imag.shape) embed_size = 128 max_len = 40 img_model = Sequential() img_model.add(Dense(embed_size, input_shape=(2048,), activation='relu')) img_model.add(RepeatVector(max_len)) img_model.summary() lang_model = Sequential() lang_model.add(Embedding(input_dim=vocabulary_size, output_dim=embed_size, input_length=max_len)) lang_model.add(LSTM(256, return_sequences=True)) lang_model.add(TimeDistributed(Dense(embed_size))) lang_model.summary() concat = Concatenate()([img_model.output, lang_model.output]) x = LSTM(128, return_sequences=True)(concat) x = LSTM(512, return_sequences=False)(x) x = Dense(vocabulary_size)(x) out = Activation('softmax')(x) model = Model(inputs=[img_model.input, lang_model.input], outputs = out) model.compile(loss='categorical_crossentropy', optimizer='RMSprop', metrics=['accuracy']) model.summary() hist = model.fit([images
if img == '': continue for capt in tok[img]: caption = "<start> "+ capt + " <end>" testing_dataset.write((img+"\t"+caption+"\n").encode()) testing_dataset.flush()
conditional_block
image_caption.py
lickr_8k_train_dataset.txt','wb') training_dataset.write(b"image_id\tcap\n") valid_dataset = open('flickr_8k_val_dataset.txt','wb') valid_dataset.write(b"image_id\tcap\n") testing_dataset = open('flickr_8k_test_dataset.txt','wb') testing_dataset.write(b"image_id\tcap\n") # Loading image ids and captions for each of these images in the above 3 files for img in x_training: if img == '': continue for capt in tok[img]: caption = "<start> "+ capt + " <end>" training_dataset.write((img+"\t"+caption+"\n").encode()) training_dataset.flush() training_dataset.close() for img in x_testing: if img == '': continue for capt in tok[img]: caption = "<start> "+ capt + " <end>" testing_dataset.write((img+"\t"+caption+"\n").encode()) testing_dataset.flush() testing_dataset.close() for img in x_valid: if img == '': continue for capt in tok[img]: caption = "<start> "+ capt + " <end>" valid_dataset.write((img+"\t"+caption+"\n").encode()) valid_dataset.flush() valid_dataset.close() # Here, we're using ResNet50 Model from IPython.core.display import display, HTML model = ResNet50(include_top=False,weights='imagenet',input_shape=(224,224,3),pooling='avg') model.summary() # process images to target size def preprocess(img_path): im = image.load_img(img_path, target_size=(224,224,3)) im = image.img_to_array(im) # (x, y, z) im = np.expand_dims(im, axis=0) # (0, x, y, z) return im training_data = {} counter=0 for item in x_training: if item == "": continue if counter >= 3000: break counter+=1 if counter%1000==0: print(counter) path = img_path + item img = preprocess(path) #to change the dimensions of the image for using ResNet model pred = model.predict(img).reshape(2048) # shape of each image is (2048, 0) training_data[item] = pred # opening train_enc_img.p file and dumping content of training_data to this file with open( "train_enc_img.p", "wb" ) as pickle_f: #obj hierarchy is converted into byte stream pickle.dump(training_data, pickle_f ) # Storing image and its corresponding caption into a dataframe pd_dataset = pd.read_csv("flickr_8k_train_dataset.txt", delimiter='\t') dframe = pd_dataset.values print(dframe.shape) pd_dataset.head() # Storing all the captions from dframe into a list senten = [] for item in range(dframe.shape[0]): senten.append(dframe[item, 1]) #senten will have 30000 length # First 5 captions stored in senten senten[:5] # Splitting each captions stored in 'senten' and storing them in 'wor' as list of list wor = [i.split() for i in senten] # Creating a list of all unique wor uniq = [] for i in wor: uniq.extend(i) uniq = list(set(uniq)) print(len(uniq)) vocabulary_size = len(uniq) # making 2 lists to index each unique word and vice-versa w_to_i = {val:index for index, val in enumerate(uniq)} i_to_w = {index:val for index, val in enumerate(uniq)} w_to_i['UNK'] = 0 w_to_i['raining'] = 8253 i_to_w[0] = 'UNK' i_to_w[8253] = 'raining' vocabulary_size = len(w_to_i.keys()) print(vocabulary_size) max_len = 0 for i in senten: i = i.split() if len(i) > max_len: max_len = len(i) print(max_len) #finding longest caption pad_seq, subsequent_wor = [], [] for item in range(dframe.shape[0]): #30000 items part_seq = [] next_wor = [] text = dframe[item, 1].split() #diving each caption for every image into words text = [w_to_i[i] for i in text] #finding index for each word for i in range(1, len(text)): part_seq.append(text[:i]) #start, 1st word, ... , last word next_wor.append(text[i]) #1st word, ... , last word, end pad_part_seq = sequence.pad_sequences(part_seq, max_len, padding='post') next_wor_1hot = np.zeros([len(next_wor), vocabulary_size], dtype=np.bool) for i,next_word in enumerate(next_wor): next_wor_1hot[i, next_word] = 1 pad_seq.append(pad_part_seq ) subsequent_wor.append(next_wor_1hot) pad_seq = np.asarray(pad_seq) subsequent_wor = np.asarray(subsequent_wor) print(pad_seq.shape) print(subsequent_wor.shape) print(pad_seq[0]) for item in range(len(pad_seq[0])): for y in range(max_len): print(i_to_w[pad_seq[0][item][y]],) print("\n") print(len(pad_seq[0])) num_imgs = 2000 cap = np.zeros([0, max_len]) next_wor = np.zeros([0, vocabulary_size]) for item in range(num_imgs): #img_to_padded_seqs.shape[0]): cap = np.concatenate([cap, pad_seq[item]]) next_wor = np.concatenate([next_wor, subsequent_wor[item]]) np.save("cap.npy", cap) np.save("next_wor.npy", next_wor) print(cap.shape) print(next_wor.shape) with open('train_enc_img.p', 'rb') as f: enc_img = pickle.load(f, encoding="bytes") imgs = [] for item in range(dframe.shape[0]): #30000 if dframe[item, 0] in enc_img.keys(): #dframe[0,0], [1,0], ... , [4,0] match with 0th key of enc_img imgs.append(list(enc_img[dframe[item, 0]])) imgs = np.asarray(imgs) print(imgs.shape) images = [] img_names = [] for item in range(num_imgs): #2000 for y in range(pad_seq[item].shape[0]): #14 images.append(imgs[item]) #1st iteration: 14 times name of image in byte form img_names.append(dframe[item, 0]) # normal form images = np.asarray(images) #images contains image_name in byte form np.save("images.npy", images) img_names = np.asarray(img_names) #img_names contains image_name normally np.save("img_names.npy", img_names) print(images.shape) print(len(img_names)) cap = np.load("cap.npy") next_wor = np.load("next_wor.npy") print(cap.shape) print(next_wor.shape) images = np.load("images.npy") print(images.shape) imag = np.load("img_names.npy") print(imag.shape) embed_size = 128 max_len = 40 img_model = Sequential() img_model.add(Dense(embed_size, input_shape=(2048,), activation='relu')) img_model.add(RepeatVector(max_len)) img_model.summary() lang_model = Sequential() lang_model.add(Embedding(input_dim=vocabulary_size, output_dim=embed_size, input_length=max_len)) lang_model.add(LSTM(256, return_sequences=True)) lang_model.add(TimeDistributed(Dense(embed_size))) lang_model.summary() concat = Concatenate()([img_model.output, lang_model.output]) x = LSTM(128, return_sequences=True)(concat) x = LSTM(512, return_sequences=False)(x) x = Dense(vocabulary_size)(x) out = Activation('softmax')(x) model = Model(inputs=[img_model.input, lang_model.input], outputs = out) model.compile(loss='categorical_crossentropy', optimizer='RMSprop', metrics=['accuracy']) model.summary() hist = model.fit([images, cap], next_wor, batch_size=512, epochs=210) for label in ["loss"]: plt.plot(hist.history[label],label=label) plt.legend() plt.xlabel("epochs") plt.ylabel("loss") plt.show() for label in ["accuracy"]: plt.plot(hist.history[label],label=label) plt.legend() plt.xlabel("epochs") plt.ylabel("accuracy") plt.show() model.save_weights("model_weights.h5") def preprocess(img_path): im = image.load_img(img_path, target_size=(224,224,3)) im = image.img_to_array(im) #(224,224,3) im = np.expand_dims(im, axis=0) #(1,224,224,3) return im def
get_encode
identifier_name
idempotent.rs
// we always receive `[x, n] as a` // .. // (v-- chut' ne udalil :lol:, podumal wo hn9, A NE-NE-NE) //> ... SO, if "fn" mutate own args, and computation which "generate new value, // to update old value in args (in case of [] is some "next-empty-cell" in arr, and `len`) // is based ON THIS ARG", .. /// we (most likely) will consider it as NOT IDEMPONENT /// .. SO .. WHAT IS IMPORTANT ?? // despite to be NOT IDEMPOTENT (from "caller" perspective) .. since we have // 2 refs (caller and callee) and "our"(callee) is `mut` .. WE HAVE ADVANTAGE // of EXPLICIT PASSING THIS REF(to callee) , BUT unfortunatly in JS we can't write // `&mut reff` .. // IT'S (1) PREVENT "ORDER"-PROBLEM described above (2) makes MORE OBVIOUS for // "reviewer" that "a" can be mutated (he at least SEE this "a" is passsed) .. // ..but `some_fn(num)` is MUCH WORSE // /// TRY TO MAKE ALL EFFECTFULL-FN TO BE IDEMPOTENT | 1 // ALSO: (p.s. JS here, ONLY SINGLE-threaded programs and (in this post) SYNC) { fn calculateAverage(list) { // << obviously IDEMPOTENT sum = list.reduce((sum, x) => x + sum, 0); return sum / list.length; } var sum, nums = [1,2,4,7,11,16,22]; var avg = calculateAverage( nums ); } // Getify said that (if `sum` not mutably-aliased some another closure) // mutation of `sum` here it's "UNOBSERVED SIDE-EFF" .. // (.. although, such "UNOBSERVED MUTATIONS" easy can be eliminated, and we can use // pure-style easily here (but actually here it's just MATTER OF TASTE) // ) // ... SO .. what reasons we have for IDEMPOTENT MUTATIONS aka "unobserved side-eff"? // ... either NONE or "OWNED DATA"(P.Graham) // .. and further author cite Hickey: /// "If a tree falls in the forest, but no one is around to hear it, does it still make a sound?" // (^-- NO SHARED MUTABLE ALIASING .. or .. // .. we INCAPSULATE some data to MAINTAIN invariant [like in // exmps/js_state.rs:state_shraing] ) // // .. so it "like" pure-fn, or in term of P.Graham fn what used "owned data" fn get_cached_fn() { /// AS-OPPOSED to "`a` in `some_fn`"(above) `cache` is not "mut aliased"(only one ref which /// mutate and read) var cache = {}; // << "owned data", "unheard tree in the forest" return fn XX(n) { // "unobserved mutation", "like-pure", "right incapsultaion" if (cache[n] !== undefined) return cache[n] // .... NAZYVAI KAK HOCHEW' ..... /* do computation */ cache[n] = res_of_comutation(); } } // author: "The point is that if XX(..) is the only part of the program that accesses // and updates the cache side cause/effect, so it's "like"-pure" | 2.0 // ALSO: ORDER/FREQUENCY DEPENDENCY { var users = {}, userOrders = {}; fn fetchUserData(userId) { ajax( "http://some.api/user/" + userId, fn onUserData(userData) {
} fn fetchOrders(userId) { ajax( "http://some.api/orders/" + userId, fn onOrders(orders) { for (let i = 0; i < orders.length; i++) { // keep a reference to latest order for each user users[userId].latestOrder = orders[i]; userOrders[orders[i].orderId] = orders[i]; } }); } } // `fetchUserData` should be called before `fetchOrders` ... AVOID SUCH "LOCAL STATE MACHINES" /// [GIST] there is should NOT BE ANY DEPENDENCY between ORDER/FREQUENCY OF CALL to /// "getters/setters" ..(that is (kind of) SHOULD BE ONLY IDEMPOTENT-EFFECTFUL-FN).. /// of some "local state machine" aka "class instance" aka "object"(with incapsulation) // .. // ...So.. in this exmp, (also Miwko: initializing DB, which depends on "glob-state" - that is // "aliasing") .. `fetchOrders` depends on ///| to be called.. IN SOME PARTICULAR STATE .. and this state is reached through ///| "PARAMS-REF"(aliasing) ... SO.. in such case, ///| -=! SUCH STATE SHOULD BE PASSED DIRECTLY !=-.. /// (as in Miwko db-inital we refactor too) /// ..to AVOID "ORDER/FREQUENCY-OF-CALL(getter/setter) DEPENDENCY"(to set "needed" state) // ... any Getter/Setter should be designed in such way to BE ABLE to be called in ANY STATE! //>> OR.. THIS "ORDER_DEPENDENCY/COUPLING" SHOULD BE INCAPSULATED IN fn /// (BUT EXPLICIT PASSING OF MUT_ALISASED DEPENDENCY AS ARG /// IF IT POSSIBLE /// IS BETTER) // (in Miwko db-exmp: if we don't pass explicitly "mut-aliased"-state , then we should // incapsulate all COUPLED fns in single fn .. that is prepare NEEDED STATE .. // BUT.. IF THIS CHAIN OF COUPLED-fns is too long .. // |> WE RETURN TO THE PROBLEM .. // since we refactoring of fn-that-incapsulate-coupling is ERROR PRONE .. SO /// THIS CHAIN OF COUPLED_FNs SHOULD BE SHORT .. USUALLY === 2 ) // .. // Of course to achieve some "desired side-effs"(without them programm is useless) we // OFTEN need call "getters/setters" in some particular order ..BUT.. that DOESN'T mean that // this order should become SPECIAL(HARDCODED) aka ORDER DEPENDENCY .. we just peek that order, // but it's should NOT BE SPECIAL for our "get/set-ters". // | 2.1 // So.. why I looked at ::exmps/js_state.rs::state_sharing fn .. and though: "why I need // "immutable-world", clojure, etc .. "ved'" this fn pretty good, and do its job well" // .. // .... BECAUSE: this "local-state-machine" don't rely on PARTICULAR STATE and ORDER/FREQUENCY.. // so IN ANY STATE ... we can call "get/set-ters" in ANY ORDER .. AND IT'S OK, IT DO EXACTLY // WHAT WE WANT, EXECTLY SIDE-EFF WHAT WE WANT. /// After createing an object we can start do ANY "STATE_TRANSITION" which this /// object has, BUT, if only some STATE_TRANSITIONS allowed, IT'S ERRORPRONE. // ...> whether "produce" or "consume" be called first .. WE DON'T CARE, // ...> what the state we have in the moment of call to any fn(cons/prod) .. WE DON'T CARE... // ..(some particular-state doesn't have special meaning for us) // ..... ///|"consume"(as a getter[but it aslo a setter]) is NOT IDEMPOTENT .. because.. // for this I need pass the list-for-consuming EXPLICITLY.. // (the main condition for IDEMPOTENCY - absence of MUT_ALIASING[here `list`]) /// .. BUT: /// It INCAPSULATE "INVARIANT" of "prefiex is chenged only when we "splavili" old values in a /// list and clean it up" ///>> .. INCUPSULATEs IN ONE STATE_TRANSITION // .. SO .. it's let's both (`consume` and `produce`) "fns" works without FEAR about shared // (between them ) STATE(and INCAPSULATED) on which the relie(mut_aliased) since it doesn't // aliased elsewhere("tree in the wood") .. and PROPERLY changed (INVARIANT) in `consumer`, and // this change INCAPSULATED in right place /// that_is: WE DEFINE ATOMIC_STATE_TRANSITIONs, why atomic, since state mutation/transaction /// can't be "interrupted" (here exmp: "prefix" changed BUT we not yet "splavili" values from /// `list` -- IT'S INTERRUPTION of STATE_TRANSITION) /// .. // .. also ..
users[userId] = userData; });
random_line_split
idempotent.rs
// not IDEMPOTENT because: *WRONG: we update, and/but not completly rewrite `x`, // *RIGHT: mut_aliasing (our computed value for `x` rely on mut_aliased `x`) // (BUT) // BTW(p.s.)--v: `some_fn` break rule: /// "mutate only TREE_in_the_WOOD , emit "own_copied/immutabl" value" // (TREE_in_the_WOOD is accesed or through "self/this", or "closure"-capturing) // fn some_fn(a = [], n) { a.push(n) } // ^-- here was: "ADD value, not just change." // BUT actually .. it's also can be considered as MUT_alising .. since result of `push` // depends of `a`, .. and if we call 2 times `some_fn` of course it's gives different sideeff, // since its computation based on MUTated value, ("EU ZHE") // .. // From the other hand .. `some_fn` can be considered as IDEMPONENT, since it's always gives // the same side-eff for the same input-VALUE .. that is if we alway will give `[x] as a` .. // we always receive `[x, n] as a` // .. // (v-- chut' ne udalil :lol:, podumal wo hn9, A NE-NE-NE) //> ... SO, if "fn" mutate own args, and computation which "generate new value, // to update old value in args (in case of [] is some "next-empty-cell" in arr, and `len`) // is based ON THIS ARG", .. /// we (most likely) will consider it as NOT IDEMPONENT /// .. SO .. WHAT IS IMPORTANT ?? // despite to be NOT IDEMPOTENT (from "caller" perspective) .. since we have // 2 refs (caller and callee) and "our"(callee) is `mut` .. WE HAVE ADVANTAGE // of EXPLICIT PASSING THIS REF(to callee) , BUT unfortunatly in JS we can't write // `&mut reff` .. // IT'S (1) PREVENT "ORDER"-PROBLEM described above (2) makes MORE OBVIOUS for // "reviewer" that "a" can be mutated (he at least SEE this "a" is passsed) .. // ..but `some_fn(num)` is MUCH WORSE // /// TRY TO MAKE ALL EFFECTFULL-FN TO BE IDEMPOTENT | 1 // ALSO: (p.s. JS here, ONLY SINGLE-threaded programs and (in this post) SYNC) { fn calculateAverage(list) { // << obviously IDEMPOTENT sum = list.reduce((sum, x) => x + sum, 0); return sum / list.length; } var sum, nums = [1,2,4,7,11,16,22]; var avg = calculateAverage( nums ); } // Getify said that (if `sum` not mutably-aliased some another closure) // mutation of `sum` here it's "UNOBSERVED SIDE-EFF" .. // (.. although, such "UNOBSERVED MUTATIONS" easy can be eliminated, and we can use // pure-style easily here (but actually here it's just MATTER OF TASTE) // ) // ... SO .. what reasons we have for IDEMPOTENT MUTATIONS aka "unobserved side-eff"? // ... either NONE or "OWNED DATA"(P.Graham) // .. and further author cite Hickey: /// "If a tree falls in the forest, but no one is around to hear it, does it still make a sound?" // (^-- NO SHARED MUTABLE ALIASING .. or .. // .. we INCAPSULATE some data to MAINTAIN invariant [like in // exmps/js_state.rs:state_shraing] ) // // .. so it "like" pure-fn, or in term of P.Graham fn what used "owned data" fn get_cached_fn() { /// AS-OPPOSED to "`a` in `some_fn`"(above) `cache` is not "mut aliased"(only one ref which /// mutate and read) var cache = {}; // << "owned data", "unheard tree in the forest" return fn XX(n) { // "unobserved mutation", "like-pure", "right incapsultaion" if (cache[n] !== undefined) return cache[n] // .... NAZYVAI KAK HOCHEW' ..... /* do computation */ cache[n] = res_of_comutation(); } } // author: "The point is that if XX(..) is the only part of the program that accesses // and updates the cache side cause/effect, so it's "like"-pure" | 2.0 // ALSO: ORDER/FREQUENCY DEPENDENCY { var users = {}, userOrders = {}; fn fetchUserData(userId) { ajax( "http://some.api/user/" + userId, fn onUserData(userData) { users[userId] = userData; }); } fn fetchOrders(userId) { ajax( "http://some.api/orders/" + userId, fn onOrders(orders) { for (let i = 0; i < orders.length; i++) { // keep a reference to latest order for each user users[userId].latestOrder = orders[i]; userOrders[orders[i].orderId] = orders[i]; } }); } } // `fetchUserData` should be called before `fetchOrders` ... AVOID SUCH "LOCAL STATE MACHINES" /// [GIST] there is should NOT BE ANY DEPENDENCY between ORDER/FREQUENCY OF CALL to /// "getters/setters" ..(that is (kind of) SHOULD BE ONLY IDEMPOTENT-EFFECTFUL-FN).. /// of some "local state machine" aka "class instance" aka "object"(with incapsulation) // .. // ...So.. in this exmp, (also Miwko: initializing DB, which depends on "glob-state" - that is // "aliasing") .. `fetchOrders` depends on ///| to be called.. IN SOME PARTICULAR STATE .. and this state is reached through ///| "PARAMS-REF"(aliasing) ... SO.. in such case, ///| -=! SUCH STATE SHOULD BE PASSED DIRECTLY !=-.. /// (as in Miwko db-inital we refactor too) /// ..to AVOID "ORDER/FREQUENCY-OF-CALL(getter/setter) DEPENDENCY"(to set "needed" state) // ... any Getter/Setter should be designed in such way to BE ABLE to be called in ANY STATE! //>> OR.. THIS "ORDER_DEPENDENCY/COUPLING" SHOULD BE INCAPSULATED IN fn /// (BUT EXPLICIT PASSING OF MUT_ALISASED DEPENDENCY AS ARG /// IF IT POSSIBLE /// IS BETTER) // (in Miwko db-exmp: if we don't pass explicitly "mut-aliased"-state , then we should // incapsulate all COUPLED fns in single fn .. that is prepare NEEDED STATE .. // BUT.. IF THIS CHAIN OF COUPLED-fns is too long .. // |> WE RETURN TO THE PROBLEM .. // since we refactoring of fn-that-incapsulate-coupling is ERROR PRONE .. SO /// THIS CHAIN OF COUPLED_FNs SHOULD BE SHORT .. USUALLY === 2 ) // .. // Of course to achieve some "desired side-effs"(without them programm is useless) we // OFTEN need call "getters/setters" in some particular order ..BUT.. that DOESN'T mean that // this order should become SPECIAL(HARDCODED) aka ORDER DEPENDENCY .. we just peek that order, // but it's should NOT BE SPECIAL for our "get/set-ters". // | 2.1 // So.. why I looked at ::exmps/js_state.rs::state_sharing fn .. and though: "why I need // "immutable-world", clojure, etc .. "ved'" this fn pretty good, and do its job well" // .. // .... BECAUSE: this "local-state-machine" don't rely on PARTICULAR STATE and ORDER/FREQUENCY.. // so IN ANY STATE ... we can call "get/set-ters" in ANY ORDER .. AND IT'S OK, IT DO EXACTLY // WHAT WE WANT, EXECTLY SIDE-EFF WHAT WE WANT. /// After createing an object we can start do ANY "STATE_TRANSITION" which this /// object has, BUT, if only some STATE_TRANSITIONS allowed, IT'S ERRORPRONE. // ...> whether "produce" or "consume" be called first .. WE DON'T CARE, // ...> what the state we have in the moment of call to any fn(cons/prod) .. WE DON'T CARE... // ..(some particular-state doesn't have special meaning for us) // ..... ///|"consume"(as a getter[but it aslo
{ x += n*x }
identifier_body
idempotent.rs
(a = [], n) { a.push(n) } // ^-- here was: "ADD value, not just change." // BUT actually .. it's also can be considered as MUT_alising .. since result of `push` // depends of `a`, .. and if we call 2 times `some_fn` of course it's gives different sideeff, // since its computation based on MUTated value, ("EU ZHE") // .. // From the other hand .. `some_fn` can be considered as IDEMPONENT, since it's always gives // the same side-eff for the same input-VALUE .. that is if we alway will give `[x] as a` .. // we always receive `[x, n] as a` // .. // (v-- chut' ne udalil :lol:, podumal wo hn9, A NE-NE-NE) //> ... SO, if "fn" mutate own args, and computation which "generate new value, // to update old value in args (in case of [] is some "next-empty-cell" in arr, and `len`) // is based ON THIS ARG", .. /// we (most likely) will consider it as NOT IDEMPONENT /// .. SO .. WHAT IS IMPORTANT ?? // despite to be NOT IDEMPOTENT (from "caller" perspective) .. since we have // 2 refs (caller and callee) and "our"(callee) is `mut` .. WE HAVE ADVANTAGE // of EXPLICIT PASSING THIS REF(to callee) , BUT unfortunatly in JS we can't write // `&mut reff` .. // IT'S (1) PREVENT "ORDER"-PROBLEM described above (2) makes MORE OBVIOUS for // "reviewer" that "a" can be mutated (he at least SEE this "a" is passsed) .. // ..but `some_fn(num)` is MUCH WORSE // /// TRY TO MAKE ALL EFFECTFULL-FN TO BE IDEMPOTENT | 1 // ALSO: (p.s. JS here, ONLY SINGLE-threaded programs and (in this post) SYNC) { fn calculateAverage(list) { // << obviously IDEMPOTENT sum = list.reduce((sum, x) => x + sum, 0); return sum / list.length; } var sum, nums = [1,2,4,7,11,16,22]; var avg = calculateAverage( nums ); } // Getify said that (if `sum` not mutably-aliased some another closure) // mutation of `sum` here it's "UNOBSERVED SIDE-EFF" .. // (.. although, such "UNOBSERVED MUTATIONS" easy can be eliminated, and we can use // pure-style easily here (but actually here it's just MATTER OF TASTE) // ) // ... SO .. what reasons we have for IDEMPOTENT MUTATIONS aka "unobserved side-eff"? // ... either NONE or "OWNED DATA"(P.Graham) // .. and further author cite Hickey: /// "If a tree falls in the forest, but no one is around to hear it, does it still make a sound?" // (^-- NO SHARED MUTABLE ALIASING .. or .. // .. we INCAPSULATE some data to MAINTAIN invariant [like in // exmps/js_state.rs:state_shraing] ) // // .. so it "like" pure-fn, or in term of P.Graham fn what used "owned data" fn get_cached_fn() { /// AS-OPPOSED to "`a` in `some_fn`"(above) `cache` is not "mut aliased"(only one ref which /// mutate and read) var cache = {}; // << "owned data", "unheard tree in the forest" return fn XX(n) { // "unobserved mutation", "like-pure", "right incapsultaion" if (cache[n] !== undefined) return cache[n] // .... NAZYVAI KAK HOCHEW' ..... /* do computation */ cache[n] = res_of_comutation(); } } // author: "The point is that if XX(..) is the only part of the program that accesses // and updates the cache side cause/effect, so it's "like"-pure" | 2.0 // ALSO: ORDER/FREQUENCY DEPENDENCY { var users = {}, userOrders = {}; fn fetchUserData(userId) { ajax( "http://some.api/user/" + userId, fn onUserData(userData) { users[userId] = userData; }); } fn fetchOrders(userId) { ajax( "http://some.api/orders/" + userId, fn onOrders(orders) { for (let i = 0; i < orders.length; i++) { // keep a reference to latest order for each user users[userId].latestOrder = orders[i]; userOrders[orders[i].orderId] = orders[i]; } }); } } // `fetchUserData` should be called before `fetchOrders` ... AVOID SUCH "LOCAL STATE MACHINES" /// [GIST] there is should NOT BE ANY DEPENDENCY between ORDER/FREQUENCY OF CALL to /// "getters/setters" ..(that is (kind of) SHOULD BE ONLY IDEMPOTENT-EFFECTFUL-FN).. /// of some "local state machine" aka "class instance" aka "object"(with incapsulation) // .. // ...So.. in this exmp, (also Miwko: initializing DB, which depends on "glob-state" - that is // "aliasing") .. `fetchOrders` depends on ///| to be called.. IN SOME PARTICULAR STATE .. and this state is reached through ///| "PARAMS-REF"(aliasing) ... SO.. in such case, ///| -=! SUCH STATE SHOULD BE PASSED DIRECTLY !=-.. /// (as in Miwko db-inital we refactor too) /// ..to AVOID "ORDER/FREQUENCY-OF-CALL(getter/setter) DEPENDENCY"(to set "needed" state) // ... any Getter/Setter should be designed in such way to BE ABLE to be called in ANY STATE! //>> OR.. THIS "ORDER_DEPENDENCY/COUPLING" SHOULD BE INCAPSULATED IN fn /// (BUT EXPLICIT PASSING OF MUT_ALISASED DEPENDENCY AS ARG /// IF IT POSSIBLE /// IS BETTER) // (in Miwko db-exmp: if we don't pass explicitly "mut-aliased"-state , then we should // incapsulate all COUPLED fns in single fn .. that is prepare NEEDED STATE .. // BUT.. IF THIS CHAIN OF COUPLED-fns is too long .. // |> WE RETURN TO THE PROBLEM .. // since we refactoring of fn-that-incapsulate-coupling is ERROR PRONE .. SO /// THIS CHAIN OF COUPLED_FNs SHOULD BE SHORT .. USUALLY === 2 ) // .. // Of course to achieve some "desired side-effs"(without them programm is useless) we // OFTEN need call "getters/setters" in some particular order ..BUT.. that DOESN'T mean that // this order should become SPECIAL(HARDCODED) aka ORDER DEPENDENCY .. we just peek that order, // but it's should NOT BE SPECIAL for our "get/set-ters". // | 2.1 // So.. why I looked at ::exmps/js_state.rs::state_sharing fn .. and though: "why I need // "immutable-world", clojure, etc .. "ved'" this fn pretty good, and do its job well" // .. // .... BECAUSE: this "local-state-machine" don't rely on PARTICULAR STATE and ORDER/FREQUENCY.. // so IN ANY STATE ... we can call "get/set-ters" in ANY ORDER .. AND IT'S OK, IT DO EXACTLY // WHAT WE WANT, EXECTLY SIDE-EFF WHAT WE WANT. /// After createing an object we can start do ANY "STATE_TRANSITION" which this /// object has, BUT, if only some STATE_TRANSITIONS allowed, IT'S ERRORPRONE. // ...> whether "produce" or "consume" be called first .. WE DON'T CARE, // ...> what the state we have in the moment of call to any fn(cons/prod) .. WE DON'T CARE... // ..(some particular-state doesn't have special meaning for us) // ..... ///|"consume"(as a getter[but it aslo a setter]) is NOT IDEMPOTENT .. because.. // for this I need pass the list-for-consuming EXPLICITLY.. // (the main condition for IDEMPOTENCY - absence of MUT_ALIASING[here `list`]) /// .. BUT: /// It INCAPSULATE "INVARIANT" of "prefiex is chenged only when we "splavili" old values in a /// list and clean it up" ///>> .. INCUPSULATEs IN ONE STATE_TRANSITION // .. SO .. it's let's both (`consume` and `produce`) "fns"
some_fn
identifier_name
core_convert.py
# TODO: MAP THIS ?: 4, # Grekiskt # TODO: MAP THIS ?: 5, # Indien # TODO: MAP THIS ?: 6, # Nordamerika # TODO: MAP THIS ?: 7, # Latinamerika # TODO: MAP THIS ?: 8, # Orienten # TODO: MAP THIS ?: 9, # Japan # TODO: MAP THIS ?: 10, # Italienskt # TODO: MAP THIS ?: 11, # Sydostasien # TODO: MAP THIS ?: 12, # Spansk # TODO: MAP THIS ?: 13, # Tyskland # TODO: MAP THIS ?: 14, # "Ryssland och Östeuropa" # TODO: MAP THIS ?: 15, # Internationellt # TODO: MAP THIS ?: 16, # Övriga # TODO: MAP THIS ?: 73, # Sverige # TODO: MAP THIS ?: 74, # Norge # TODO: MAP THIS ?: 75, # Kanada # TODO: MAP THIS ?: 76, # Frankrike # TODO: MAP THIS ?: 77, # Grekland # TODO: MAP THIS ?: 78, # Portugal # TODO: MAP THIS ?: 79, # Danmark # TODO: MAP THIS ?: 80, # Italien # TODO: MAP THIS ?: 81, # Finland # TODO: MAP THIS ?: 82, # Kalifornien # TODO: MAP THIS ?: 83, # Thailand # TODO: MAP THIS ?: 84, # Kina # TODO: MAP THIS ?: 85, # Belgien # TODO: MAP THIS ?: 86, # Europa # TODO: MAP THIS ?: 87, # Turkiet # TODO: MAP THIS ?: 88, # Holland # TODO: MAP THIS ?: 89, # England # TODO: MAP THIS ?: 90, # Spanien # TODO: MAP THIS ?: 91, # Nederländerna # TODO: MAP THIS ?: 92, # Polen # TODO: MAP THIS ?: 93, # "Blandat: EG och icke EG" # TODO: MAP THIS ?: 94, # Ungern # TODO: MAP THIS ?: 95, # Bulgarien # TODO: MAP THIS ?: 96, # Kroatien # TODO: MAP THIS ?: 98, # India # TODO: MAP THIS ?: 99, # Uruguay # TODO: MAP THIS ?: 100, # Irland # TODO: MAP THIS ?: 101, # "Nya Zeeland" # TODO: MAP THIS ?: 102, # Sverige/England # TODO: MAP THIS ?: 103, # Sverige/Danmark # TODO: MAP THIS ?: 104, # China # TODO: MAP THIS ?: 105, # Holland/Frankrike # TODO: MAP THIS ?: 106, # "Costa Rica" # TODO: MAP THIS ?: 107, # Zaire # TODO: MAP THIS ?: 108, # Israel/USA # TODO: MAP THIS ?: 109, # Mexico # TODO: MAP THIS ?: 110, # Holland/Belgien # TODO: MAP THIS ?: 111, # Frankrike/Italien # TODO: MAP THIS ?: 112, # Sverge # TODO: MAP THIS ?: 113, # Centralamerika # TODO: MAP THIS ?: 114, # Brasilien # TODO: MAP THIS ?: 115, # Israel/Indien # TODO: MAP THIS ?: 116, # "Italien/Nya Zeeland" # TODO: MAP THIS ?: 117, # Sydafrika # TODO: MAP THIS ?: 118, # Argentina # TODO: MAP THIS ?: 119, # China/Thailand # TODO: MAP THIS ?: 120, # USA # TODO: MAP THIS ?: 121, # Kenya # TODO: MAP THIS ?: 122, # Israel # TODO: MAP THIS ?: 123, # Malaysia # TODO: MAP THIS ?: 124, # Nordostatlanten # TODO: MAP THIS ?: 125, # Vietnam # TODO: MAP THIS ?: 126, # Norden # TODO: MAP THIS ?: 127, # Litauen # TODO: MAP THIS ?: 131, # Roslagen # TODO: MAP THIS ?: 135, # U.S.A. # TODO: MAP THIS ?: 136, # DK # TODO: MAP THIS ?: 137, # Egypten # TODO: MAP THIS ?: 138, # Marocko # TODO: MAP THIS ?: 139, # Chile # TODO: MAP THIS ?: 140, # "Dominikanska Republiken" # TODO: MAP THIS ?: 141, # Iran # TODO: MAP THIS ?: 142, # Colombia # TODO: MAP THIS ?: 143, # Peru # TODO: MAP THIS ?: 144, # Zimbabwe } return table[key] if key in table else None def convert_attributes(product, detail=None): result = [] for tag in product.tags.all(): id = get_attribute_id(tag.name) if id is not None: result.append({ 'AttributeId': id }) # Special case for "Nyhet" if not detail and product.product_detail: detail = product.product_detail.filter(store=10).first() if detail is None: detail = product.product_detail.first() if detail: first_enabled = detail.first_enabled if detail.first_enabled else datetime.now() - \ timedelta(days=60) result.append({ 'AttributeId': 2173, 'FromDate': first_enabled, 'ToDate': first_enabled + timedelta(days=30), }) return result def create_dynamic_property(key, value, store=None): prop = { 'PropertyId': get_dynamic_property_id(key), 'PropertyName': key, 'PropertyValue': value, } if store is not None: prop['StoreId'] = store return prop def convert_dynamic_properties(product): result = [ create_dynamic_property('Volume', product.volume_dm3), create_dynamic_property('Weight', product.weight_g), create_dynamic_property('KfpHeightMM', product.height_mm), create_dynamic_property('KfpLengthtMM', product.length_mm), create_dynamic_property('KfpWidthMM', product.width_mm), create_dynamic_property('Season', '.'.join( convert_season_tags(product))), create_dynamic_property('LastReceiptDay', product.last_receipt_day), create_dynamic_property('LastSalesDay', product.last_sales_day), create_dynamic_property('TruckrouteOptimizationProd3', convert_order_route_from_product_type(product.product_type)), create_dynamic_property('BDMaterialNumber', product.prefered_merchantarticle.external_id if product.prefered_merchantarticle else None), create_dynamic_property('SupplierArticleNumber', product.prefered_merchantarticle.external_id if product.prefered_merchantarticle else None), ] base_unit_quantity = get_base_unit_quantity(product, product.article.gtin) if base_unit_quantity is not None: create_dynamic_property('KfpDfp', base_unit_quantity) for detail in product.product_detail.all(): result.append(create_dynamic_property( 'OrderFactor', 1 if detail.orderfactor else 0, detail.store)) result.append(create_dynamic_property( 'BDMaterialNumber', detail.prefered_merchantarticle.external_id if detail.prefered_merchantarticle else None, detail.store)) result.append(create_dynamic_property( 'SupplierArticleNumber', detail.prefered_merchantarticle.external_id if detail.prefered_merchantarticle else None, detail.store)) base_unit_quantity = get_base_unit_quantity( detail, product.article.gtin) if base_unit_quantity is not None: create_dynamic_property('KfpDfp', base_unit_quantity, detail.store) return result def get_base_unit_quantity(product, base_unit_gtin): if product.prefered_merchantarticle is not None: if product.prefered_merchantarticle.article.child_gtin == base_unit_gtin: return product.prefered_merchantarticle.article.quantity_of_lower_layer else: upper_quantity = product.prefered_merchantarticle.article.quantity_of_lower_layer next_lower_article = Article.objects.filter( gtin=product.prefered_merchantarticle.article.child_gtin).first() if next_lower_article is not None: if next_lower_article.child_gtin
== product.article.gtin: return next_lower_article.quantity_of_lower_layer * upper_quantity return None def convert_u
conditional_block
core_convert.py
'5+ dagar': 2197, '7+ dagar': 2200, '10+ dagar': 2203, '30+ dagar': 2206, 'Svenskt ursprung': 2209, 'Svensk fågel': 2212, '4+ dagar': 2215, 'Vegansk': 2218, 'MSC': 2219, 'Strategisk produkt': 2222, 'Svenskt sigill klimatcertifierad': 2224, 'ASC': 2227, 'Från Sverige': 2230, 'Kött från Sverige': 2233, 'Mjölk från Sverige': 2236, 'Faroklass brandfarligt': 2239, 'Faroklass miljöfarligt': 2242, 'Faroklass skadligt': 2245, 'Faroklass Warning': 2248, 'Energiklass A+': 2251, 'Energiklass C': 2254, 'Energiklass D': 2257, 'Energiklass E': 2260, 'Energiklass A++': 2263, 'Energiklass A': 2266, 'Energiklass B': 2269, } return table[key] if key in table else None def get_dynamic_property_id (key): table = { 'Volume': 1, 'Weight': 2, 'KfpDfp': 3, 'LastSalesDay': 4, 'LastReceiptDay': 5, 'OldPz1': 6, 'OldPz2': 7, 'OldPz3': 8, 'MaxStock': 9, 'Season': 10, 'OrderFactor': 11, 'MinStock': 12, 'DfpLengthMM': 13, 'DfpWidthMM': 14, 'DfpHeightMM': 15, 'DfpWeightG': 16, 'DfpType': 17, 'SupplierArticleNumber': 18, 'AxfoodArticleId': 19, 'TruckrouteOptimizationProd3': 20, 'KfpHeightMM': 21, 'KfpLengthtMM': 22, 'KfpWidthMM': 23, 'IsFakeStockBalance': 24, 'ExternalImageUrl': 25, 'ProductSupplier': 26, 'ValdioDFPWidthMM': 27, 'ValdioDFPHeightMM': 28, 'ValidoDFPLengthtMM': 29, 'ValdioDFPWeightG': 30, 'DFPEANCode': 31, 'SafetyStock': 33, 'KfpDfpPurchaseOrder': 36, 'NoNutritionsNeeded': 38, 'NoIngredientsNeeded': 41, 'NoAllergensNeeded': 44, 'DeliveredUnitConversionFactor': 45, 'HandlingUnitQuantity': 46, 'BDMaterialNumber': 49, 'ProductSegment': 55, 'StandardUnitKfp': 56, 'StandardUnitGtin': 59, 'LimitedOfferProduct': 61, 'QLPricing': 64, 'QLMatching': 67, 'FirstSalesDate': 70, 'CategoryManager': 73, } return table[key] if key in table else None def get_origin_id(key): table = { 752: 1, # Svensk 249: 2, # Fransk # TODO: MAP THIS ?: 3, # Afrika # TODO: MAP THIS ?: 4, # Grekiskt # TODO: MAP THIS ?: 5, # Indien # TODO: MAP THIS ?: 6, # Nordamerika # TODO: MAP THIS ?: 7, # Latinamerika # TODO: MAP THIS ?: 8, # Orienten # TODO: MAP THIS ?: 9, # Japan # TODO: MAP THIS ?: 10, # Italienskt # TODO: MAP THIS ?: 11, # Sydostasien # TODO: MAP THIS ?: 12, # Spansk # TODO: MAP THIS ?: 13, # Tyskland # TODO: MAP THIS ?: 14, # "Ryssland och Östeuropa" # TODO: MAP THIS ?: 15, # Internationellt # TODO: MAP THIS ?: 16, # Övriga # TODO: MAP THIS ?: 73, # Sverige # TODO: MAP THIS ?: 74, # Norge # TODO: MAP THIS ?: 75, # Kanada # TODO: MAP THIS ?: 76, # Frankrike # TODO: MAP THIS ?: 77, # Grekland # TODO: MAP THIS ?: 78, # Portugal # TODO: MAP THIS ?: 79, # Danmark # TODO: MAP THIS ?: 80, # Italien # TODO: MAP THIS ?: 81, # Finland # TODO: MAP THIS ?: 82, # Kalifornien # TODO: MAP THIS ?: 83, # Thailand # TODO: MAP THIS ?: 84, # Kina # TODO: MAP THIS ?: 85, # Belgien # TODO: MAP THIS ?: 86, # Europa # TODO: MAP THIS ?: 87, # Turkiet # TODO: MAP THIS ?: 88, # Holland # TODO: MAP THIS ?: 89, # England # TODO: MAP THIS ?: 90, # Spanien # TODO: MAP THIS ?: 91, # Nederländerna # TODO: MAP THIS ?: 92, # Polen # TODO: MAP THIS ?: 93, # "Blandat: EG och icke EG" # TODO: MAP THIS ?: 94, # Ungern # TODO: MAP THIS ?: 95, # Bulgarien # TODO: MAP THIS ?: 96, # Kroatien # TODO: MAP THIS ?: 98, # India # TODO: MAP THIS ?: 99, # Uruguay # TODO: MAP THIS ?: 100, # Irland # TODO: MAP THIS ?: 101, # "Nya Zeeland" # TODO: MAP THIS ?: 102, # Sverige/England # TODO: MAP THIS ?: 103, # Sverige/Danmark # TODO: MAP THIS ?: 104, # China # TODO: MAP THIS ?: 105, # Holland/Frankrike # TODO: MAP THIS ?: 106, # "Costa Rica" # TODO: MAP THIS ?: 107, # Zaire # TODO: MAP THIS ?: 108, # Israel/USA # TODO: MAP THIS ?: 109, # Mexico # TODO: MAP THIS ?: 110, # Holland/Belgien # TODO: MAP THIS ?: 111, # Frankrike/Italien # TODO: MAP THIS ?: 112, # Sver
'Ekonomipack': 1, 'Nyckelhålsmärkt': 1736, 'Ekologisk': 2167, 'Glutenfri': 2168, 'Laktosfri': 2169, 'Låglaktos': 2170, 'Premiumkvalité': 2171, 'Mjölkproteinfri': 2172, # 'Nyhet': 2173, '18Åldersgräns': 2174, 'Fairtrade': 2175, 'Svanenmärkt': 2176, 'Kravmärkt': 2177, 'Video': 2178, 'Äkta vara': 2181, 'Astma- och Allergiförbundet': 2184, 'test': 2187, 'Rosa bandet': 2190, 'Svenskt sigill': 2191, '3+ dagar': 2194,
identifier_body
core_convert.py
: MAP THIS ?: 10, # Italienskt # TODO: MAP THIS ?: 11, # Sydostasien # TODO: MAP THIS ?: 12, # Spansk # TODO: MAP THIS ?: 13, # Tyskland # TODO: MAP THIS ?: 14, # "Ryssland och Östeuropa" # TODO: MAP THIS ?: 15, # Internationellt # TODO: MAP THIS ?: 16, # Övriga # TODO: MAP THIS ?: 73, # Sverige # TODO: MAP THIS ?: 74, # Norge # TODO: MAP THIS ?: 75, # Kanada # TODO: MAP THIS ?: 76, # Frankrike # TODO: MAP THIS ?: 77, # Grekland # TODO: MAP THIS ?: 78, # Portugal # TODO: MAP THIS ?: 79, # Danmark # TODO: MAP THIS ?: 80, # Italien # TODO: MAP THIS ?: 81, # Finland # TODO: MAP THIS ?: 82, # Kalifornien # TODO: MAP THIS ?: 83, # Thailand # TODO: MAP THIS ?: 84, # Kina # TODO: MAP THIS ?: 85, # Belgien # TODO: MAP THIS ?: 86, # Europa # TODO: MAP THIS ?: 87, # Turkiet # TODO: MAP THIS ?: 88, # Holland # TODO: MAP THIS ?: 89, # England # TODO: MAP THIS ?: 90, # Spanien # TODO: MAP THIS ?: 91, # Nederländerna # TODO: MAP THIS ?: 92, # Polen # TODO: MAP THIS ?: 93, # "Blandat: EG och icke EG" # TODO: MAP THIS ?: 94, # Ungern # TODO: MAP THIS ?: 95, # Bulgarien # TODO: MAP THIS ?: 96, # Kroatien # TODO: MAP THIS ?: 98, # India # TODO: MAP THIS ?: 99, # Uruguay # TODO: MAP THIS ?: 100, # Irland # TODO: MAP THIS ?: 101, # "Nya Zeeland" # TODO: MAP THIS ?: 102, # Sverige/England # TODO: MAP THIS ?: 103, # Sverige/Danmark # TODO: MAP THIS ?: 104, # China # TODO: MAP THIS ?: 105, # Holland/Frankrike # TODO: MAP THIS ?: 106, # "Costa Rica" # TODO: MAP THIS ?: 107, # Zaire # TODO: MAP THIS ?: 108, # Israel/USA # TODO: MAP THIS ?: 109, # Mexico # TODO: MAP THIS ?: 110, # Holland/Belgien # TODO: MAP THIS ?: 111, # Frankrike/Italien # TODO: MAP THIS ?: 112, # Sverge # TODO: MAP THIS ?: 113, # Centralamerika # TODO: MAP THIS ?: 114, # Brasilien # TODO: MAP THIS ?: 115, # Israel/Indien # TODO: MAP THIS ?: 116, # "Italien/Nya Zeeland" # TODO: MAP THIS ?: 117, # Sydafrika # TODO: MAP THIS ?: 118, # Argentina # TODO: MAP THIS ?: 119, # China/Thailand # TODO: MAP THIS ?: 120, # USA # TODO: MAP THIS ?: 121, # Kenya # TODO: MAP THIS ?: 122, # Israel # TODO: MAP THIS ?: 123, # Malaysia # TODO: MAP THIS ?: 124, # Nordostatlanten # TODO: MAP THIS ?: 125, # Vietnam # TODO: MAP THIS ?: 126, # Norden # TODO: MAP THIS ?: 127, # Litauen # TODO: MAP THIS ?: 131, # Roslagen # TODO: MAP THIS ?: 135, # U.S.A. # TODO: MAP THIS ?: 136, # DK # TODO: MAP THIS ?: 137, # Egypten # TODO: MAP THIS ?: 138, # Marocko # TODO: MAP THIS ?: 139, # Chile # TODO: MAP THIS ?: 140, # "Dominikanska Republiken" # TODO: MAP THIS ?: 141, # Iran # TODO: MAP THIS ?: 142, # Colombia # TODO: MAP THIS ?: 143, # Peru # TODO: MAP THIS ?: 144, # Zimbabwe } return table[key] if key in table else None def convert_attributes(product, detail=None): result = [] for tag in product.tags.all(): id = get_attribute_id(tag.name) if id is not None: result.append({ 'AttributeId': id }) # Special case for "Nyhet" if not detail and product.product_detail: detail = product.product_detail.filter(store=10).first() if detail is None: detail = product.product_detail.first() if detail: first_enabled = detail.first_enabled if detail.first_enabled else datetime.now() - \ timedelta(days=60) result.append({ 'AttributeId': 2173, 'FromDate': first_enabled, 'ToDate': first_enabled + timedelta(days=30), }) return result def create_dynamic_property(key, value, store=None): prop = { 'PropertyId': get_dynamic_property_id(key), 'PropertyName': key, 'PropertyValue': value, } if store is not None: prop['StoreId'] = store return prop def convert_dynamic_properties(product): result = [ create_dynamic_property('Volume', product.volume_dm3), create_dynamic_property('Weight', product.weight_g), create_dynamic_property('KfpHeightMM', product.height_mm), create_dynamic_property('KfpLengthtMM', product.length_mm), create_dynamic_property('KfpWidthMM', product.width_mm), create_dynamic_property('Season', '.'.join( convert_season_tags(product))), create_dynamic_property('LastReceiptDay', product.last_receipt_day), create_dynamic_property('LastSalesDay', product.last_sales_day), create_dynamic_property('TruckrouteOptimizationProd3', convert_order_route_from_product_type(product.product_type)), create_dynamic_property('BDMaterialNumber', product.prefered_merchantarticle.external_id if product.prefered_merchantarticle else None), create_dynamic_property('SupplierArticleNumber', product.prefered_merchantarticle.external_id if product.prefered_merchantarticle else None), ] base_unit_quantity = get_base_unit_quantity(product, product.article.gtin) if base_unit_quantity is not None: create_dynamic_property('KfpDfp', base_unit_quantity) for detail in product.product_detail.all(): result.append(create_dynamic_property( 'OrderFactor', 1 if detail.orderfactor else 0, detail.store)) result.append(create_dynamic_property( 'BDMaterialNumber', detail.prefered_merchantarticle.external_id if detail.prefered_merchantarticle else None, detail.store)) result.append(create_dynamic_property( 'SupplierArticleNumber', detail.prefered_merchantarticle.external_id if detail.prefered_merchantarticle else None, detail.store)) base_unit_quantity = get_base_unit_quantity( detail, product.article.gtin) if base_unit_quantity is not None: create_dynamic_property('KfpDfp', base_unit_quantity, detail.store) return result def get_base_unit_quantity(product, base_unit_gtin): if product.prefered_merchantarticle is not None: if product.prefered_merchantarticle.article.child_gtin == base_unit_gtin: return product.prefered_merchantarticle.article.quantity_of_lower_layer else: upper_quantity = product.prefered_merchantarticle.article.quantity_of_lower_layer next_lower_article = Article.objects.filter( gtin=product.prefered_merchantarticle.article.child_gtin).first() if next_lower_article is not None: if next_lower_article.child_gtin == product.article.gtin: return next_lower_article.quantity_of_lower_layer * upper_quantity return None def convert_unit(validoo_unit): # data from prefilledautomaten.unit unit_table = {
"H87": 1, # st, PIECES "GRM": 2, # g, WEIGHT "KGM": 3, # kg, WEIGHT "DLT": 6, # dl, VOLUME "LTR": 7, # L, VOLUME
random_line_split
core_convert.py
able = { "Crossdocking": "X", "Nightorder": "A", } return table[key] if key in table else None def get_attribute_id(key): # data from prefilledautomaten.attribute table = { 'Ekonomipack': 1, 'Nyckelhålsmärkt': 1736, 'Ekologisk': 2167, 'Glutenfri': 2168, 'Laktosfri': 2169, 'Låglaktos': 2170, 'Premiumkvalité': 2171, 'Mjölkproteinfri': 2172, # 'Nyhet': 2173, '18Åldersgräns': 2174, 'Fairtrade': 2175, 'Svanenmärkt': 2176, 'Kravmärkt': 2177, 'Video': 2178, 'Äkta vara': 2181, 'Astma- och Allergiförbundet': 2184, 'test': 2187, 'Rosa bandet': 2190, 'Svenskt sigill': 2191, '3+ dagar': 2194, '5+ dagar': 2197, '7+ dagar': 2200, '10+ dagar': 2203, '30+ dagar': 2206, 'Svenskt ursprung': 2209, 'Svensk fågel': 2212, '4+ dagar': 2215, 'Vegansk': 2218, 'MSC': 2219, 'Strategisk produkt': 2222, 'Svenskt sigill klimatcertifierad': 2224, 'ASC': 2227, 'Från Sverige': 2230, 'Kött från Sverige': 2233, 'Mjölk från Sverige': 2236, 'Faroklass brandfarligt': 2239, 'Faroklass miljöfarligt': 2242, 'Faroklass skadligt': 2245, 'Faroklass Warning': 2248, 'Energiklass A+': 2251, 'Energiklass C': 2254, 'Energiklass D': 2257, 'Energiklass E': 2260, 'Energiklass A++': 2263, 'Energiklass A': 2266, 'Energiklass B': 2269, } return table[key] if key in table else None def get_dynamic_property_id(key): table = { 'Volume': 1, 'Weight': 2, 'KfpDfp': 3, 'LastSalesDay': 4, 'LastReceiptDay': 5, 'OldPz1': 6, 'OldPz2': 7, 'OldPz3': 8, 'MaxStock': 9, 'Season': 10, 'OrderFactor': 11, 'MinStock': 12, 'DfpLengthMM': 13, 'DfpWidthMM': 14, 'DfpHeightMM': 15, 'DfpWeightG': 16, 'DfpType': 17, 'SupplierArticleNumber': 18, 'AxfoodArticleId': 19, 'TruckrouteOptimizationProd3': 20, 'KfpHeightMM': 21, 'KfpLengthtMM': 22, 'KfpWidthMM': 23, 'IsFakeStockBalance': 24, 'ExternalImageUrl': 25, 'ProductSupplier': 26, 'ValdioDFPWidthMM': 27, 'ValdioDFPHeightMM': 28, 'ValidoDFPLengthtMM': 29, 'ValdioDFPWeightG': 30, 'DFPEANCode': 31, 'SafetyStock': 33, 'KfpDfpPurchaseOrder': 36, 'NoNutritionsNeeded': 38, 'NoIngredientsNeeded': 41, 'NoAllergensNeeded': 44, 'DeliveredUnitConversionFactor': 45, 'HandlingUnitQuantity': 46, 'BDMaterialNumber': 49, 'ProductSegment': 55, 'StandardUnitKfp': 56, 'StandardUnitGtin': 59, 'LimitedOfferProduct': 61, 'QLPricing': 64, 'QLMatching': 67, 'FirstSalesDate': 70, 'CategoryManager': 73, } return table[key] if key in table else None def get_origin_id(key): table = { 752: 1, # Svensk 249: 2, # Fransk # TODO: MAP THIS ?: 3, # Afrika # TODO: MAP THIS ?: 4, # Grekiskt # TODO: MAP THIS ?: 5, # Indien # TODO: MAP THIS ?: 6, # Nordamerika # TODO: MAP THIS ?: 7, # Latinamerika # TODO: MAP THIS ?: 8, # Orienten # TODO: MAP THIS ?: 9, # Japan # TODO: MAP THIS ?: 10, # Italienskt # TODO: MAP THIS ?: 11, # Sydostasien # TODO: MAP THIS ?: 12, # Spansk # TODO: MAP THIS ?: 13, # Tyskland # TODO: MAP THIS ?: 14, # "Ryssland och Östeuropa" # TODO: MAP THIS ?: 15, # Internationellt # TODO: MAP THIS ?: 16, # Övriga # TODO: MAP THIS ?: 73, # Sverige # TODO: MAP THIS ?: 74, # Norge # TODO: MAP THIS ?: 75, # Kanada # TODO: MAP THIS ?: 76, # Frankrike # TODO: MAP THIS ?: 77, # Grekland # TODO: MAP THIS ?: 78, # Portugal # TODO: MAP THIS ?: 79, # Danmark # TODO: MAP THIS ?: 80, # Italien # TODO: MAP THIS ?: 81, # Finland # TODO: MAP THIS ?: 82, # Kalifornien # TODO: MAP THIS ?: 83, # Thailand # TODO: MAP THIS ?: 84, # Kina # TODO: MAP THIS ?: 85, # Belgien # TODO: MAP THIS ?: 86, # Europa # TODO: MAP THIS ?: 87, # Turkiet # TODO: MAP THIS ?: 88, # Holland # TODO: MAP THIS ?: 89, # England # TODO: MAP THIS ?: 90, # Spanien # TODO: MAP THIS ?: 91, # Nederländerna # TODO: MAP THIS ?: 92, # Polen # TODO: MAP THIS ?: 93, # "Blandat: EG och icke EG" # TODO: MAP THIS ?: 94, # Ungern # TODO: MAP THIS ?: 95, # Bulgarien # TODO: MAP THIS ?: 96, # Kroatien # TODO: MAP THIS ?: 98, # India # TODO: MAP THIS ?: 99, # Uruguay # TODO: MAP THIS ?: 100, # Irland # TODO: MAP THIS ?: 101, # "Nya Zeeland" # TODO: MAP THIS ?: 102, # Sverige/England # TODO: MAP THIS ?: 103, # Sverige/Danmark # TODO: MAP THIS ?: 104, # China # TODO: MAP THIS ?: 105, # Holland/Frankrike # TODO: MAP THIS ?: 106, # "Costa Rica" # TODO: MAP THIS ?: 107, # Zaire # TODO: MAP THIS ?: 108, # Israel/USA # TODO: MAP
r_route_from_product_type(key): t
identifier_name
parser.go
/juniperUDP/port" "github.com/golang/protobuf/proto" _"strings" ) // JuniperUDPParser is an object for Parsing incoming metrics. type JuniperUDPParser struct { // DefaultTags will be added to every parsed metric DefaultTags map[string]string } /* func (p *JuniperUDPParser) ParseWithDefaultTimePrecision(buf []byte, t time.Time, precision string) ([]telegraf.Metric, error) { if !bytes.HasSuffix(buf, []byte("\n")) { buf = append(buf, '\n') } // parse even if the buffer begins with a newline buf = bytes.TrimPrefix(buf, []byte("\n")) metrics, err := metric.ParseWithDefaultTimePrecision(buf, t, precision) if len(p.DefaultTags) > 0 { for _, m := range metrics { for k, v := range p.DefaultTags { // only set the default tag if it doesn't already exist: if !m.HasTag(k) { m.AddTag(k, v) } } } } return metrics, err } */ func parseArray(data []interface{}, masterKey string) []interface{} { var arrData []interface{} for _,val := range data{ valType := reflect.ValueOf(val).Kind() if valType == reflect.Map{ mapData := parseMap(val.(map[string]interface{}), masterKey) for _,tmpData := range(mapData){ arrData = append(arrData, tmpData) } } else {fmt.Println("Error!!!! Leaf elements in array are not coded. Please open a issue.")} } return arrData } func parseMap(data map[string]interface{}, masterKey string) []interface{} { var leafData map[string]interface{} var arrData []interface{} var arrKey []string var finalData []interface{} var newMasterKey string leafData = make(map[string]interface{}) for key,val := range data{ if masterKey == ""{ newMasterKey = key } else { newMasterKey = masterKey + "." + key } valType := reflect.ValueOf(val).Kind() if valType == reflect.Map{ mapData := parseMap(val.(map[string]interface{}), newMasterKey) if reflect.TypeOf(mapData).Kind() == reflect.Map{ var tmpArr []interface{} tmpArr = append(tmpArr, mapData) arrData = append(arrData, tmpArr) } else if reflect.TypeOf(mapData).Kind() == reflect.Slice{ arrData = append(arrData, mapData) } arrKey = append(arrKey, newMasterKey) } else if valType == reflect.Slice{ arrData = append(arrData, parseArray(val.([]interface{}), newMasterKey)) arrKey = append(arrKey, newMasterKey) } else { leafData[newMasterKey] = val} } if len(leafData) != 0 { for i,key := range arrKey{ _ = key for _,data_aa := range arrData[i].([]interface{}){ leafTmp := leafData if data_aa != nil { if reflect.ValueOf(data_aa).Kind() == reflect.Map{ for key_aa, value_aa := range data_aa.(map[string]interface{}){ leafTmp[key_aa] = value_aa } finalData = append(finalData, make(map[string]interface{})) for k,v := range leafTmp{ finalData[len(finalData)-1].(map[string]interface{})[k] = v } } else { for _,data_ha := range data_aa.([]interface{}){ leafTmp = leafData for key_aa,value_aa := range data_ha.(map[string]interface{}){ leafTmp[key_aa] = value_aa } finalData = append(finalData, make(map[string]interface{})) for k,v := range leafTmp{ finalData[len(finalData)-1].(map[string]interface{})[k] = v } } } } } } } else {finalData = arrData} arrData = arrData[:0] if (len(finalData) == 0) && (len(leafData)!= 0) { finalData = append(finalData, leafData) } return finalData } // Parse returns a slice of Metrics from a text representation of a // metric (in line-protocol format) // with each metric separated by newlines. If any metrics fail to parse, // a non-nil error will be returned in addition to the metrics that parsed // successfully. func (p *JuniperUDPParser) Parse(buf []byte) ([]telegraf.Metric, error) { //out, _ := os.Create("telegraf_udp.log") dir, err := filepath.Abs(filepath.Dir(os.Args[0])) if err != nil { log.Fatal(err) } path := dir+"/logs" if _, err := os.Stat(path); os.IsNotExist(err) { os.Mkdir(path, 0777) } out, errFile := os.OpenFile("logs/telegraf_udp.log", os.O_APPEND|os.O_CREATE|os.O_WRONLY, 0644) if errFile != nil { log.Fatal(errFile) } prefix := "" flag := log.LstdFlags | log.Lmicroseconds | log.Lshortfile newLog := log.New(out, prefix, flag) newLog.Printf("Data byte buffer received!!\n") sensorMapping := map[string]string{"jnpr_interface_ext" : "/junos/system/linecard/interface/", "cpu_memory_util_ext" : "/junos/system/linecard/cpu/memory/", "fabricMessageExt" : "/junos/system/linecard/fabric", "jnpr_firewall_ext" : "/junos/system/linecard/firewall/", "jnprLogicalInterfaceExt" : "/junos/system/linecard/interface/logical/usage/", "npu_memory_ext" : "/junos/system/linecard/npu/memory/", "jnpr_npu_utilization_ext" : "/junos/system/linecard/npu/utilization/", "jnpr_optics_ext" : "/junos/system/linecard/optics/", "jnpr_packet_statistics_ext" : "/junos/system/linecard/packet/usage/", "jnpr_qmon_ext" : "/junos/system/linecard/qmon/", "inline_jflow_stats_ext" : "/junos/system/linecard/services/inline-jflow/", "jnpr_cmerror_data_ext" : "NA", "jnpr_cmerror_ext" : "NA", "jnpr_lsp_statistics_ext" : "NA", "jnpr_interface_exp_ext" : "NA", "jnpr_sr_stats_per_if_egress_ext" : "NA", "jnpr_sr_stats_per_if_ingress_ext" : "NA", "jnpr_sr_stats_per_sid_ext" : "NA", } _ = sensorMapping go func() { for { var m runtime.MemStats runtime.ReadMemStats(&m) //log.Printf("\nAlloc = %v\nTotalAlloc = %v\nSys = %v\nNumGC = %v\n\n", m.Alloc / 1024, m.TotalAlloc / 1024, m.Sys / 1024, m.NumGC) time.Sleep(5 * time.Second) } }() //s := string(buf[:len(buf)]) //fmt.Println("#######################################################################") //fmt.Printf("%v",s) //fmt.Println("#######################################################################") ts := &telemetry_top.TelemetryStream{} if err := proto.Unmarshal(buf, ts); err != nil { fmt.Println("Error!! Unable to parse data: ", err) return nil,err } //fmt.Printf("%v",ts) host,errHost := os.Hostname() _ = host if errHost != nil { fmt.Println("Error!! Host name not found: ", errHost) return nil, errHost } deviceName := ts.GetSystemId() newLog.Printf("Device : %v", deviceName) newLog.Printf("Host : %v", host) gpbTime := ts.GetTimestamp() measurementPrefix := "enterprise.juniperNetworks" jnprSensorName := ts.GetSensorName() sensorName := jnprSensorName _ = gpbTime _ = measurementPrefix _ = jnprSensorName m := &jsonpb.Marshaler{} tsJSON,err := m.MarshalToString(ts) if err!= nil{ fmt.Println("Error!! ", err) } var data map[string]interface{} errU := json.Unmarshal([]byte(tsJSON), &data) if errU != nil { fmt.Println("Error!! Unable to unmarshal: ",errU) return nil, nil //panic(errU) } //newLog.Printf("Data received : %v", data) enterpriseSensorData := data["enterprise"] sensorData, ok := enterpriseSensorData.(map[string]interface{}) jnprSensorData := sensorData["[juniperNetworks]"] if !ok { return nil, nil panic("inner map is not a map!") } metrics := make([]telegraf.Metric, 0) sensorNum := 0
random_line_split
parser.go
uniperUDP/npu_utilization" _"github.com/influxdata/telegraf/plugins/parsers/juniperUDP/optics" _"github.com/influxdata/telegraf/plugins/parsers/juniperUDP/packet_stats" _"github.com/influxdata/telegraf/plugins/parsers/juniperUDP/port_exp" _"github.com/influxdata/telegraf/plugins/parsers/juniperUDP/port" "github.com/golang/protobuf/proto" _"strings" ) // JuniperUDPParser is an object for Parsing incoming metrics. type JuniperUDPParser struct { // DefaultTags will be added to every parsed metric DefaultTags map[string]string } /* func (p *JuniperUDPParser) ParseWithDefaultTimePrecision(buf []byte, t time.Time, precision string) ([]telegraf.Metric, error) { if !bytes.HasSuffix(buf, []byte("\n")) { buf = append(buf, '\n') } // parse even if the buffer begins with a newline buf = bytes.TrimPrefix(buf, []byte("\n")) metrics, err := metric.ParseWithDefaultTimePrecision(buf, t, precision) if len(p.DefaultTags) > 0 { for _, m := range metrics { for k, v := range p.DefaultTags { // only set the default tag if it doesn't already exist: if !m.HasTag(k) { m.AddTag(k, v) } } } } return metrics, err } */ func parseArray(data []interface{}, masterKey string) []interface{} { var arrData []interface{} for _,val := range data{ valType := reflect.ValueOf(val).Kind() if valType == reflect.Map{ mapData := parseMap(val.(map[string]interface{}), masterKey) for _,tmpData := range(mapData){ arrData = append(arrData, tmpData) } } else {fmt.Println("Error!!!! Leaf elements in array are not coded. Please open a issue.")} } return arrData } func parseMap(data map[string]interface{}, masterKey string) []interface{} { var leafData map[string]interface{} var arrData []interface{} var arrKey []string var finalData []interface{} var newMasterKey string leafData = make(map[string]interface{}) for key,val := range data{ if masterKey == ""{ newMasterKey = key } else { newMasterKey = masterKey + "." + key } valType := reflect.ValueOf(val).Kind() if valType == reflect.Map{ mapData := parseMap(val.(map[string]interface{}), newMasterKey) if reflect.TypeOf(mapData).Kind() == reflect.Map{ var tmpArr []interface{} tmpArr = append(tmpArr, mapData) arrData = append(arrData, tmpArr) } else if reflect.TypeOf(mapData).Kind() == reflect.Slice{ arrData = append(arrData, mapData) } arrKey = append(arrKey, newMasterKey) } else if valType == reflect.Slice{ arrData = append(arrData, parseArray(val.([]interface{}), newMasterKey)) arrKey = append(arrKey, newMasterKey) } else { leafData[newMasterKey] = val} } if len(leafData) != 0 { for i,key := range arrKey{ _ = key for _,data_aa := range arrData[i].([]interface{}){ leafTmp := leafData if data_aa != nil { if reflect.ValueOf(data_aa).Kind() == reflect.Map{ for key_aa, value_aa := range data_aa.(map[string]interface{}){ leafTmp[key_aa] = value_aa } finalData = append(finalData, make(map[string]interface{})) for k,v := range leafTmp{ finalData[len(finalData)-1].(map[string]interface{})[k] = v } } else { for _,data_ha := range data_aa.([]interface{}){ leafTmp = leafData for key_aa,value_aa := range data_ha.(map[string]interface{}){ leafTmp[key_aa] = value_aa } finalData = append(finalData, make(map[string]interface{})) for k,v := range leafTmp{ finalData[len(finalData)-1].(map[string]interface{})[k] = v } } } } } } } else {finalData = arrData} arrData = arrData[:0] if (len(finalData) == 0) && (len(leafData)!= 0) { finalData = append(finalData, leafData) } return finalData } // Parse returns a slice of Metrics from a text representation of a // metric (in line-protocol format) // with each metric separated by newlines. If any metrics fail to parse, // a non-nil error will be returned in addition to the metrics that parsed // successfully. func (p *JuniperUDPParser) Parse(buf []byte) ([]telegraf.Metric, error)
"fabricMessageExt" : "/junos/system/linecard/fabric", "jnpr_firewall_ext" : "/junos/system/linecard/firewall/", "jnprLogicalInterfaceExt" : "/junos/system/linecard/interface/logical/usage/", "npu_memory_ext" : "/junos/system/linecard/npu/memory/", "jnpr_npu_utilization_ext" : "/junos/system/linecard/npu/utilization/", "jnpr_optics_ext" : "/junos/system/linecard/optics/", "jnpr_packet_statistics_ext" : "/junos/system/linecard/packet/usage/", "jnpr_qmon_ext" : "/junos/system/linecard/qmon/", "inline_jflow_stats_ext" : "/junos/system/linecard/services/inline-jflow/", "jnpr_cmerror_data_ext" : "NA", "jnpr_cmerror_ext" : "NA", "jnpr_lsp_statistics_ext" : "NA", "jnpr_interface_exp_ext" : "NA", "jnpr_sr_stats_per_if_egress_ext" : "NA", "jnpr_sr_stats_per_if_ingress_ext" : "NA", "jnpr_sr_stats_per_sid_ext" : "NA", } _ = sensorMapping go func() { for { var m runtime.MemStats runtime.ReadMemStats(&m) //log.Printf("\nAlloc = %v\nTotalAlloc = %v\nSys = %v\nNumGC = %v\n\n", m.Alloc / 1024, m.TotalAlloc / 1024, m.Sys / 1024, m.NumGC) time.Sleep(5 * time.Second) } }() //s := string(buf[:len(buf)]) //fmt.Println("#######################################################################") //fmt.Printf("%v",s) //fmt.Println("#######################################################################") ts := &telemetry_top.TelemetryStream{} if err := proto.Unmarshal(buf, ts); err != nil { fmt.Println("Error!! Unable to parse data: ", err) return nil,err } //fmt.Printf("%v",ts) host,errHost := os.Hostname() _ = host if errHost != nil { fmt.Println("Error!! Host name not found: ", errHost) return nil, errHost } deviceName := ts.GetSystemId() newLog.Printf("Device : %v", deviceName) newLog.Printf("Host : %v", host) gpbTime := ts.GetTimestamp() measurementPrefix := "enterprise.juniperNetworks" jnprSensorName := ts.GetSensorName() sensorName := jnprSensorName _ = gpbTime _ = measurementPrefix _ = jnprSensorName m := &jsonpb.Marshaler{} tsJSON,err := m.MarshalToString(ts) if err!= nil{ fmt.Println("Error!! ", err) } var data map[string]interface{} errU := json.Unmarshal([]byte(tsJSON), &data) if errU != nil { fmt.Println("Error!! Unable to unmarshal: ",errU) return nil, nil //panic(errU) } //newLog.Printf("Data received : %v
{ //out, _ := os.Create("telegraf_udp.log") dir, err := filepath.Abs(filepath.Dir(os.Args[0])) if err != nil { log.Fatal(err) } path := dir+"/logs" if _, err := os.Stat(path); os.IsNotExist(err) { os.Mkdir(path, 0777) } out, errFile := os.OpenFile("logs/telegraf_udp.log", os.O_APPEND|os.O_CREATE|os.O_WRONLY, 0644) if errFile != nil { log.Fatal(errFile) } prefix := "" flag := log.LstdFlags | log.Lmicroseconds | log.Lshortfile newLog := log.New(out, prefix, flag) newLog.Printf("Data byte buffer received!!\n") sensorMapping := map[string]string{"jnpr_interface_ext" : "/junos/system/linecard/interface/", "cpu_memory_util_ext" : "/junos/system/linecard/cpu/memory/",
identifier_body
parser.go
.HasTag(k) { m.AddTag(k, v) } } } } return metrics, err } */ func parseArray(data []interface{}, masterKey string) []interface{} { var arrData []interface{} for _,val := range data{ valType := reflect.ValueOf(val).Kind() if valType == reflect.Map{ mapData := parseMap(val.(map[string]interface{}), masterKey) for _,tmpData := range(mapData){ arrData = append(arrData, tmpData) } } else {fmt.Println("Error!!!! Leaf elements in array are not coded. Please open a issue.")} } return arrData } func parseMap(data map[string]interface{}, masterKey string) []interface{} { var leafData map[string]interface{} var arrData []interface{} var arrKey []string var finalData []interface{} var newMasterKey string leafData = make(map[string]interface{}) for key,val := range data{ if masterKey == ""{ newMasterKey = key } else { newMasterKey = masterKey + "." + key } valType := reflect.ValueOf(val).Kind() if valType == reflect.Map{ mapData := parseMap(val.(map[string]interface{}), newMasterKey) if reflect.TypeOf(mapData).Kind() == reflect.Map{ var tmpArr []interface{} tmpArr = append(tmpArr, mapData) arrData = append(arrData, tmpArr) } else if reflect.TypeOf(mapData).Kind() == reflect.Slice{ arrData = append(arrData, mapData) } arrKey = append(arrKey, newMasterKey) } else if valType == reflect.Slice{ arrData = append(arrData, parseArray(val.([]interface{}), newMasterKey)) arrKey = append(arrKey, newMasterKey) } else { leafData[newMasterKey] = val} } if len(leafData) != 0 { for i,key := range arrKey{ _ = key for _,data_aa := range arrData[i].([]interface{}){ leafTmp := leafData if data_aa != nil { if reflect.ValueOf(data_aa).Kind() == reflect.Map{ for key_aa, value_aa := range data_aa.(map[string]interface{}){ leafTmp[key_aa] = value_aa } finalData = append(finalData, make(map[string]interface{})) for k,v := range leafTmp{ finalData[len(finalData)-1].(map[string]interface{})[k] = v } } else { for _,data_ha := range data_aa.([]interface{}){ leafTmp = leafData for key_aa,value_aa := range data_ha.(map[string]interface{}){ leafTmp[key_aa] = value_aa } finalData = append(finalData, make(map[string]interface{})) for k,v := range leafTmp{ finalData[len(finalData)-1].(map[string]interface{})[k] = v } } } } } } } else {finalData = arrData} arrData = arrData[:0] if (len(finalData) == 0) && (len(leafData)!= 0) { finalData = append(finalData, leafData) } return finalData } // Parse returns a slice of Metrics from a text representation of a // metric (in line-protocol format) // with each metric separated by newlines. If any metrics fail to parse, // a non-nil error will be returned in addition to the metrics that parsed // successfully. func (p *JuniperUDPParser) Parse(buf []byte) ([]telegraf.Metric, error) { //out, _ := os.Create("telegraf_udp.log") dir, err := filepath.Abs(filepath.Dir(os.Args[0])) if err != nil { log.Fatal(err) } path := dir+"/logs" if _, err := os.Stat(path); os.IsNotExist(err) { os.Mkdir(path, 0777) } out, errFile := os.OpenFile("logs/telegraf_udp.log", os.O_APPEND|os.O_CREATE|os.O_WRONLY, 0644) if errFile != nil { log.Fatal(errFile) } prefix := "" flag := log.LstdFlags | log.Lmicroseconds | log.Lshortfile newLog := log.New(out, prefix, flag) newLog.Printf("Data byte buffer received!!\n") sensorMapping := map[string]string{"jnpr_interface_ext" : "/junos/system/linecard/interface/", "cpu_memory_util_ext" : "/junos/system/linecard/cpu/memory/", "fabricMessageExt" : "/junos/system/linecard/fabric", "jnpr_firewall_ext" : "/junos/system/linecard/firewall/", "jnprLogicalInterfaceExt" : "/junos/system/linecard/interface/logical/usage/", "npu_memory_ext" : "/junos/system/linecard/npu/memory/", "jnpr_npu_utilization_ext" : "/junos/system/linecard/npu/utilization/", "jnpr_optics_ext" : "/junos/system/linecard/optics/", "jnpr_packet_statistics_ext" : "/junos/system/linecard/packet/usage/", "jnpr_qmon_ext" : "/junos/system/linecard/qmon/", "inline_jflow_stats_ext" : "/junos/system/linecard/services/inline-jflow/", "jnpr_cmerror_data_ext" : "NA", "jnpr_cmerror_ext" : "NA", "jnpr_lsp_statistics_ext" : "NA", "jnpr_interface_exp_ext" : "NA", "jnpr_sr_stats_per_if_egress_ext" : "NA", "jnpr_sr_stats_per_if_ingress_ext" : "NA", "jnpr_sr_stats_per_sid_ext" : "NA", } _ = sensorMapping go func() { for { var m runtime.MemStats runtime.ReadMemStats(&m) //log.Printf("\nAlloc = %v\nTotalAlloc = %v\nSys = %v\nNumGC = %v\n\n", m.Alloc / 1024, m.TotalAlloc / 1024, m.Sys / 1024, m.NumGC) time.Sleep(5 * time.Second) } }() //s := string(buf[:len(buf)]) //fmt.Println("#######################################################################") //fmt.Printf("%v",s) //fmt.Println("#######################################################################") ts := &telemetry_top.TelemetryStream{} if err := proto.Unmarshal(buf, ts); err != nil { fmt.Println("Error!! Unable to parse data: ", err) return nil,err } //fmt.Printf("%v",ts) host,errHost := os.Hostname() _ = host if errHost != nil { fmt.Println("Error!! Host name not found: ", errHost) return nil, errHost } deviceName := ts.GetSystemId() newLog.Printf("Device : %v", deviceName) newLog.Printf("Host : %v", host) gpbTime := ts.GetTimestamp() measurementPrefix := "enterprise.juniperNetworks" jnprSensorName := ts.GetSensorName() sensorName := jnprSensorName _ = gpbTime _ = measurementPrefix _ = jnprSensorName m := &jsonpb.Marshaler{} tsJSON,err := m.MarshalToString(ts) if err!= nil{ fmt.Println("Error!! ", err) } var data map[string]interface{} errU := json.Unmarshal([]byte(tsJSON), &data) if errU != nil { fmt.Println("Error!! Unable to unmarshal: ",errU) return nil, nil //panic(errU) } //newLog.Printf("Data received : %v", data) enterpriseSensorData := data["enterprise"] sensorData, ok := enterpriseSensorData.(map[string]interface{}) jnprSensorData := sensorData["[juniperNetworks]"] if !ok { return nil, nil panic("inner map is not a map!") } metrics := make([]telegraf.Metric, 0) sensorNum := 0 sequenceNum := 0 for key, sensorData := range jnprSensorData.(map[string]interface{})
{ var fields map[string]interface{} if reflect.ValueOf(sensorData).Kind() == reflect.Map { _ = sensorName sensorName = key[1:len(key)-1] var measurementName string measurementName = sensorName /* if val, ok := sensorMapping[sensorName]; ok { measurementName = val } else { measurementName = sensorName } */ newLog.Printf("Sensor : %v", measurementName) measurementName = "juniperNetworks." + measurementName newLog.Printf("Measurement : %v", measurementName) //newLog.Printf("Data received : %v", data) parsedData := parseMap(sensorData.(map[string]interface{}), "") for _,finalData := range(parsedData){
conditional_block
parser.go
_aa] = value_aa } finalData = append(finalData, make(map[string]interface{})) for k,v := range leafTmp{ finalData[len(finalData)-1].(map[string]interface{})[k] = v } } else { for _,data_ha := range data_aa.([]interface{}){ leafTmp = leafData for key_aa,value_aa := range data_ha.(map[string]interface{}){ leafTmp[key_aa] = value_aa } finalData = append(finalData, make(map[string]interface{})) for k,v := range leafTmp{ finalData[len(finalData)-1].(map[string]interface{})[k] = v } } } } } } } else {finalData = arrData} arrData = arrData[:0] if (len(finalData) == 0) && (len(leafData)!= 0) { finalData = append(finalData, leafData) } return finalData } // Parse returns a slice of Metrics from a text representation of a // metric (in line-protocol format) // with each metric separated by newlines. If any metrics fail to parse, // a non-nil error will be returned in addition to the metrics that parsed // successfully. func (p *JuniperUDPParser) Parse(buf []byte) ([]telegraf.Metric, error) { //out, _ := os.Create("telegraf_udp.log") dir, err := filepath.Abs(filepath.Dir(os.Args[0])) if err != nil { log.Fatal(err) } path := dir+"/logs" if _, err := os.Stat(path); os.IsNotExist(err) { os.Mkdir(path, 0777) } out, errFile := os.OpenFile("logs/telegraf_udp.log", os.O_APPEND|os.O_CREATE|os.O_WRONLY, 0644) if errFile != nil { log.Fatal(errFile) } prefix := "" flag := log.LstdFlags | log.Lmicroseconds | log.Lshortfile newLog := log.New(out, prefix, flag) newLog.Printf("Data byte buffer received!!\n") sensorMapping := map[string]string{"jnpr_interface_ext" : "/junos/system/linecard/interface/", "cpu_memory_util_ext" : "/junos/system/linecard/cpu/memory/", "fabricMessageExt" : "/junos/system/linecard/fabric", "jnpr_firewall_ext" : "/junos/system/linecard/firewall/", "jnprLogicalInterfaceExt" : "/junos/system/linecard/interface/logical/usage/", "npu_memory_ext" : "/junos/system/linecard/npu/memory/", "jnpr_npu_utilization_ext" : "/junos/system/linecard/npu/utilization/", "jnpr_optics_ext" : "/junos/system/linecard/optics/", "jnpr_packet_statistics_ext" : "/junos/system/linecard/packet/usage/", "jnpr_qmon_ext" : "/junos/system/linecard/qmon/", "inline_jflow_stats_ext" : "/junos/system/linecard/services/inline-jflow/", "jnpr_cmerror_data_ext" : "NA", "jnpr_cmerror_ext" : "NA", "jnpr_lsp_statistics_ext" : "NA", "jnpr_interface_exp_ext" : "NA", "jnpr_sr_stats_per_if_egress_ext" : "NA", "jnpr_sr_stats_per_if_ingress_ext" : "NA", "jnpr_sr_stats_per_sid_ext" : "NA", } _ = sensorMapping go func() { for { var m runtime.MemStats runtime.ReadMemStats(&m) //log.Printf("\nAlloc = %v\nTotalAlloc = %v\nSys = %v\nNumGC = %v\n\n", m.Alloc / 1024, m.TotalAlloc / 1024, m.Sys / 1024, m.NumGC) time.Sleep(5 * time.Second) } }() //s := string(buf[:len(buf)]) //fmt.Println("#######################################################################") //fmt.Printf("%v",s) //fmt.Println("#######################################################################") ts := &telemetry_top.TelemetryStream{} if err := proto.Unmarshal(buf, ts); err != nil { fmt.Println("Error!! Unable to parse data: ", err) return nil,err } //fmt.Printf("%v",ts) host,errHost := os.Hostname() _ = host if errHost != nil { fmt.Println("Error!! Host name not found: ", errHost) return nil, errHost } deviceName := ts.GetSystemId() newLog.Printf("Device : %v", deviceName) newLog.Printf("Host : %v", host) gpbTime := ts.GetTimestamp() measurementPrefix := "enterprise.juniperNetworks" jnprSensorName := ts.GetSensorName() sensorName := jnprSensorName _ = gpbTime _ = measurementPrefix _ = jnprSensorName m := &jsonpb.Marshaler{} tsJSON,err := m.MarshalToString(ts) if err!= nil{ fmt.Println("Error!! ", err) } var data map[string]interface{} errU := json.Unmarshal([]byte(tsJSON), &data) if errU != nil { fmt.Println("Error!! Unable to unmarshal: ",errU) return nil, nil //panic(errU) } //newLog.Printf("Data received : %v", data) enterpriseSensorData := data["enterprise"] sensorData, ok := enterpriseSensorData.(map[string]interface{}) jnprSensorData := sensorData["[juniperNetworks]"] if !ok { return nil, nil panic("inner map is not a map!") } metrics := make([]telegraf.Metric, 0) sensorNum := 0 sequenceNum := 0 for key, sensorData := range jnprSensorData.(map[string]interface{}){ var fields map[string]interface{} if reflect.ValueOf(sensorData).Kind() == reflect.Map { _ = sensorName sensorName = key[1:len(key)-1] var measurementName string measurementName = sensorName /* if val, ok := sensorMapping[sensorName]; ok { measurementName = val } else { measurementName = sensorName } */ newLog.Printf("Sensor : %v", measurementName) measurementName = "juniperNetworks." + measurementName newLog.Printf("Measurement : %v", measurementName) //newLog.Printf("Data received : %v", data) parsedData := parseMap(sensorData.(map[string]interface{}), "") for _,finalData := range(parsedData){ //sequenceNum := 0 for _,fin := range(finalData.([]interface{})){ //fin = fin.(map[string] interface{}) fin.(map[string]interface{})["device"] = deviceName //fin.(map[string]interface{})["host"] = host fin.(map[string]interface{})["sensor_name"] = sensorName //fin.(map[string]interface{})["_seq"] = sequenceNum fields = fin.(map[string]interface{}) tags := make(map[string]string) for k, v := range p.DefaultTags { tags[k] = v } tags["_seq"] = strconv.Itoa(sequenceNum) timestamp := time.Unix(int64(gpbTime)/1000, int64(gpbTime)%1000*1000000) mtrc,err := metric.New(measurementName, tags, fields,timestamp) metrics = append(metrics, mtrc) if err!=nil { fmt.Println("Error!! Unable to create telegraf metrics: ", err) } sensorNum++ sequenceNum++ } } } } // fmt.Printf("\nData (JSON) = \n%s\n", data) // fmt.Println("\nJuniper Sensor Name: \n%s\n", jnprSensorName) // fmt.Println("\nDevice name: \n%s\n", deviceName) // fmt.Println("\nGPB time: \n%s\n", gpbTime) // fmt.Println(measurementPrefix) // fmt.Println("\nMetrics: \n") // fmt.Println(metrics) newLog.Printf("Parsed Data : %v\n", metrics) if errFileClose := out.Close(); err != nil { log.Fatal(errFileClose) } return metrics, err // return p.ParseWithDefaultTimePrecision(buf, time.Now(), "") } func (p *JuniperUDPParser) ParseLine(line string) (telegraf.Metric, error) { metrics, err := p.Parse([]byte(line + "\n")) if err != nil { return nil, err } if len(metrics) < 1 { return nil, fmt.Errorf( "Can not parse the line: %s, for data format: influx ", line) } return metrics[0], nil } func (p *JuniperUDPParser)
SetDefaultTags
identifier_name
alain.js
resolve(result.value) : adopt(result.value).then(fulfilled, rejected); } step((generator = generator.apply(thisArg, _arguments || [])).next()); }); }; Object.defineProperty(exports, "__esModule", { value: true }); exports.buildAlain = exports.addValueToVariable = exports.addProviderToModule = exports.addImportToModule = exports.refreshPathRoot = void 0; const core_1 = require("@angular-devkit/core"); const schematics_1 = require("@angular-devkit/schematics"); const ast_utils_1 = require("@schematics/angular/utility/ast-utils"); const change_1 = require("@schematics/angular/utility/change"); const find_module_1 = require("@schematics/angular/utility/find-module"); const parse_name_1 = require("@schematics/angular/utility/parse-name"); const validation_1 = require("@schematics/angular/utility/validation"); const fs = require("fs"); const path = require("path"); const ts = require("typescript"); const ast_1 = require("./ast"); const workspace_1 = require("./workspace"); const TEMPLATE_FILENAME_RE = /\.template$/; function buildSelector(schema, projectPrefix) { const ret = []; if (!schema.withoutPrefix) { if (schema.prefix) { ret.push(schema.prefix); } else if (schema.prefix === undefined && projectPrefix) { ret.push(projectPrefix); } } // module name if (schema.module) { ret.push(schema.module); } // target name if (schema.target) { ret.push(...schema.target.split('/')); } // name ret.push(core_1.strings.dasherize(schema.name)); return ret.join('-'); } function buildName(schema, prefix) { const ret = schema.withoutModulePrefixInComponentName === true ? [] : [schema.module]; if (schema.target && schema.target.length > 0) { ret.push(...schema.target.split('/')); } ret.push(schema.name); // 服务类自动过滤 list, empty 两个页面的后缀 if (prefix === 'Service' && ['list', 'empty'].includes(schema.name)) { ret.pop(); } ret.push(prefix); return core_1.strings.classify(ret.join('-')); } function refreshPathRoot(project, sch
ct) { var _a; if (schema.path === undefined) { schema.path = `/${path.join(project.sourceRoot, (_a = alainProject === null || alainProject === void 0 ? void 0 : alainProject.routesRoot) !== null && _a !== void 0 ? _a : 'app/routes')}`; } } exports.refreshPathRoot = refreshPathRoot; function resolveSchema(tree, project, schema, alainProject) { // module name if (!schema.module) { throw new schematics_1.SchematicsException(`Must specify module name. (e.g: ng g ng-alain:list <list name> -m=<module name>)`); } // path refreshPathRoot(project, schema, alainProject); schema.path += `/${schema.module}`; const parsedPath = (0, parse_name_1.parseName)(schema.path, schema.name); schema.name = parsedPath.name; schema.path = parsedPath.path; const rootPath = path.resolve(__dirname, '../../..'); const fullPath = path.join(rootPath, schema.path, schema.name); if (fs.existsSync(fullPath) && fs.readdirSync(fullPath).length > 0) { throw new schematics_1.SchematicsException(`The directory (${fullPath}) already exists`); } schema.importModulePath = (0, find_module_1.findModuleFromOptions)(tree, schema); if (!schema._filesPath) { // 若基础页尝试从 `_cli-tpl/_${schema.schematicName!}` 下查找该目录,若存在则优先使用 if (['list', 'edit', 'view', 'empty'].includes(schema.schematicName)) { const overrideDir = `/${[project.root, `_cli-tpl/_${schema.schematicName}`].filter(i => !!i).join('/')}`; const overridePath = `${overrideDir}/__path__/__name@dasherize@if-flat__/__name@dasherize__.component.ts`; if (tree.exists(overridePath) || tree.exists(`${overridePath}.template`)) { // 所在目录与命令目录同属一个目录结构,因此无须特殊处理 schema._filesPath = path.relative(__dirname, rootPath) + overrideDir; } } schema._filesPath = schema._filesPath || './files'; } // fill target if (schema.target) { schema.path += core_1.strings.dasherize(`/${schema.target}`); } schema.routerModulePath = schema.importModulePath.replace('.module.ts', '-routing.module.ts'); // html selector schema.selector = schema.selector || buildSelector(schema, project.prefix); (0, validation_1.validateHtmlSelector)(schema.selector); } function addImportToModule(tree, filePath, symbolName, fileName) { const source = (0, ast_1.getSourceFile)(tree, filePath); const change = (0, ast_utils_1.insertImport)(source, filePath, symbolName, fileName); if (change.path == null) return; const declarationRecorder = tree.beginUpdate(filePath); declarationRecorder.insertLeft(change.pos, change.toAdd); tree.commitUpdate(declarationRecorder); } exports.addImportToModule = addImportToModule; function addProviderToModule(tree, filePath, serviceName, importPath) { const source = (0, ast_1.getSourceFile)(tree, filePath); const changes = (0, ast_utils_1.addProviderToModule)(source, filePath, serviceName, importPath); const declarationRecorder = tree.beginUpdate(filePath); changes.forEach(change => { if (change.path == null) return; if (change instanceof change_1.InsertChange) { declarationRecorder.insertLeft(change.pos, change.toAdd); } }); tree.commitUpdate(declarationRecorder); } exports.addProviderToModule = addProviderToModule; function addValueToVariable(tree, filePath, variableName, text, needWrap = true) { const source = (0, ast_1.getSourceFile)(tree, filePath); const node = (0, ast_utils_1.findNode)(source, ts.SyntaxKind.Identifier, variableName); if (!node) { throw new schematics_1.SchematicsException(`Could not find any [${variableName}] variable in path '${filePath}'.`); } // eslint-disable-next-line @typescript-eslint/no-explicit-any const arr = node.parent.initializer; const change = new change_1.InsertChange(filePath, arr.end - 1, `${arr.elements && arr.elements.length > 0 ? ',' : ''}${needWrap ? '\n ' : ''}${text}`); const declarationRecorder = tree.beginUpdate(filePath); declarationRecorder.insertLeft(change.pos, change.toAdd); tree.commitUpdate(declarationRecorder); } exports.addValueToVariable = addValueToVariable; function getRelativePath(filePath, schema, prefix) { const importPath = `/${schema.path}/${schema.flat ? '' : `${core_1.strings.dasherize(schema.name)}/`}${core_1.strings.dasherize(schema.name)}.${prefix}`; return (0, find_module_1.buildRelativePath)(filePath, importPath); } function addDeclaration(schema) { return (tree) => { if (schema.skipImport || !schema.module) { return tree; } // imports addImportToModule(tree, schema.importModulePath, schema.componentName, getRelativePath(schema.importModulePath, schema, 'component')); addValueToVariable(tree, schema.importModulePath, 'COMPONENTS', schema.componentName); // component if (schema.modal !== true) { // routing addImportToModule(tree, schema.routerModulePath, schema.componentName, getRelativePath(schema.routerModulePath, schema, 'component')); addValueToVariable(tree, schema.routerModulePath, 'routes', `{ path: '${schema.name}', component: ${schema.componentName} }`); } // service if (schema.service === 'none') { addProviderToModule(tree, schema.importModulePath, schema.serviceName, getRelativePath(schema.importModulePath, schema, 'service')); } return tree; }; } function buildAlain(schema) { return (tree) => __awaiter(this, void 0, void 0, function* () { const res = yield (0, workspace_1.getProject)(tree, schema.project); if (schema.project && res.name !== schema.project) { throw new schematics_1.SchematicsException(`The specified project does not match '${schema.project}', current: ${res.name}`); } const project = res.project; resolveSchema(tree, project, schema, res.alainProject); schema.componentName = buildName(schema, 'Component'); schema.serviceName = buildName(schema, 'Service'); // Don't support inline schema.inlineTemplate = false; const templateSource = (0, schematics_1.apply)((0, schematics_1.url)(schema._filesPath), [ (0, schematics_1.filter)(filePath => !filePath.endsWith('.DS_Store')), schema.service === 'ignore' ? (0, schematics_1.filter)(filePath => !filePath.endsWith('.service.ts.template')) : (0,
ema, alainProje
identifier_name
alain.js
? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); } step((generator = generator.apply(thisArg, _arguments || [])).next()); }); }; Object.defineProperty(exports, "__esModule", { value: true }); exports.buildAlain = exports.addValueToVariable = exports.addProviderToModule = exports.addImportToModule = exports.refreshPathRoot = void 0; const core_1 = require("@angular-devkit/core"); const schematics_1 = require("@angular-devkit/schematics"); const ast_utils_1 = require("@schematics/angular/utility/ast-utils"); const change_1 = require("@schematics/angular/utility/change"); const find_module_1 = require("@schematics/angular/utility/find-module"); const parse_name_1 = require("@schematics/angular/utility/parse-name"); const validation_1 = require("@schematics/angular/utility/validation"); const fs = require("fs"); const path = require("path"); const ts = require("typescript"); const ast_1 = require("./ast"); const workspace_1 = require("./workspace"); const TEMPLATE_FILENAME_RE = /\.template$/; function buildSelector(schema, projectPrefix) { const ret = []; if (!schema.withoutPrefix) { if (schema.prefix) { ret.push(schema.prefix); } else if (schema.prefix === undefined && projectPrefix) { ret.push(projectPrefix); } } // module name if (schema.module) { ret.push(schema.module); } // target name if (schema.target) { ret.push(...schema.target.split('/')); } // name ret.push(core_1.strings.dasherize(schema.name)); return ret.join('-'); } function buildName(schema, prefix) { const ret = schema.withoutModulePrefixInComponentName === true ? [] : [schema.module]; if (schema.target && schema.target.length > 0) { ret.push(...schema.target.split('/')); } ret.push(schema.name); // 服务类自动过滤 list, empty 两个页面的后缀 if (prefix === 'Service' && ['list', 'empty'].includes(schema.name)) { ret.pop(); } ret.push(prefix); return core_1.strings.classify(ret.join('-')); } function refreshPathRoot(project, schema, alainProject) { var _a; if (schema.path === undefined) { schema.path = `/${path.join(project.sourceRoot, (_a = alainProject === null || alainProject === void 0 ? void 0 : alainProject.routesRoot) !== null && _a !== void 0 ? _a : 'app/routes')}`; } } exports.refreshPathRoot = refreshPathRoot; function resolveSchema(tree, project, schema, alainProject) { // module name if (!schema.module) { throw new schematics_1.SchematicsException(`Must specify module name. (e.g: ng g ng-alain:list <list name> -m=<module name>)`); } // path refreshPathRoot(project, schema, alainProject); schema.path += `/${schema.module}`; const parsedPath = (0, parse_name_1.parseName)(schema.path, schema.name); schema.name = parsedPath.name; schema.path = parsedPath.path; const rootPath = path.resolve(__dirname, '../../..'); const fullPath = path.join(rootPath, schema.path, schema.name); if (fs.existsSync(fullPath) && fs.readdirSync(fullPath).length > 0) { throw new schematics_1.SchematicsException(`The directory (${fullPath}) already exists`); } schema.importModulePath = (0, find_module_1.findModuleFromOptions)(tree, schema); if (!schema._filesPath) { // 若基础页尝试从 `_cli-tpl/_${schema.schematicName!}` 下查找该目录,若存在则优先使用 if (['list', 'edit', 'view', 'empty'].includes(schema.schematicName)) { const overrideDir = `/${[project.root, `_cli-tpl/_${schema.schematicName}`].filter(i => !!i).join('/')}`; const overridePath = `${overrideDir}/__path__/__name@dasherize@if-flat__/__name@dasherize__.component.ts`; if (tree.exists(overridePath) || tree.exists(`${overridePath}.template`)) { // 所在目录与命令目录同属一个目录结构,因此无须特殊处理 schema._filesPath = path.relative(__dirname, rootPath) + overrideDir; } } schema._filesPath = schema._filesPath || './files'; } // fill target if (schema.target) { schema.path += core_1.strings.dasherize(`/${schema.target}`); } schema.routerModulePath = schema.importModulePath.replace('.module.ts', '-routing.module.ts'); // html selector schema.selector = schema.selector || buildSelector(schema, project.prefix); (0, validation_1.validateHtmlSelector)(schema.selector); } function addImportToModule(tree, filePath, symbolName, fileName) { const source = (0, ast_1.getSourceFile)(tree, filePath); const change = (0, ast_utils_1.insertImport)(source, filePath, symbolName, fileName); if (change.path == null) return; const declarationRecorder = tree.beginUpdate(filePath); declarationRecorder.insertLeft(change.pos, change.toAdd); tree.commitUpdate(declarationRecorder); } exports.addImportToModule = addImportToModule; function addProviderToModule(tree, filePath, serviceName, importPath) { const source = (0, ast_1.getSourceFile)(tree, filePath); const changes = (0, ast_utils_1.addProviderToModule)(sour
= true) { const source = (0, ast_1.getSourceFile)(tree, filePath); const node = (0, ast_utils_1.findNode)(source, ts.SyntaxKind.Identifier, variableName); if (!node) { throw new schematics_1.SchematicsException(`Could not find any [${variableName}] variable in path '${filePath}'.`); } // eslint-disable-next-line @typescript-eslint/no-explicit-any const arr = node.parent.initializer; const change = new change_1.InsertChange(filePath, arr.end - 1, `${arr.elements && arr.elements.length > 0 ? ',' : ''}${needWrap ? '\n ' : ''}${text}`); const declarationRecorder = tree.beginUpdate(filePath); declarationRecorder.insertLeft(change.pos, change.toAdd); tree.commitUpdate(declarationRecorder); } exports.addValueToVariable = addValueToVariable; function getRelativePath(filePath, schema, prefix) { const importPath = `/${schema.path}/${schema.flat ? '' : `${core_1.strings.dasherize(schema.name)}/`}${core_1.strings.dasherize(schema.name)}.${prefix}`; return (0, find_module_1.buildRelativePath)(filePath, importPath); } function addDeclaration(schema) { return (tree) => { if (schema.skipImport || !schema.module) { return tree; } // imports addImportToModule(tree, schema.importModulePath, schema.componentName, getRelativePath(schema.importModulePath, schema, 'component')); addValueToVariable(tree, schema.importModulePath, 'COMPONENTS', schema.componentName); // component if (schema.modal !== true) { // routing addImportToModule(tree, schema.routerModulePath, schema.componentName, getRelativePath(schema.routerModulePath, schema, 'component')); addValueToVariable(tree, schema.routerModulePath, 'routes', `{ path: '${schema.name}', component: ${schema.componentName} }`); } // service if (schema.service === 'none') { addProviderToModule(tree, schema.importModulePath, schema.serviceName, getRelativePath(schema.importModulePath, schema, 'service')); } return tree; }; } function buildAlain(schema) { return (tree) => __awaiter(this, void 0, void 0, function* () { const res = yield (0, workspace_1.getProject)(tree, schema.project); if (schema.project && res.name !== schema.project) { throw new schematics_1.SchematicsException(`The specified project does not match '${schema.project}', current: ${res.name}`); } const project = res.project; resolveSchema(tree, project, schema, res.alainProject); schema.componentName = buildName(schema, 'Component'); schema.serviceName = buildName(schema, 'Service'); // Don't support inline schema.inlineTemplate = false; const templateSource = (0, schematics_1.apply)((0, schematics_1.url)(schema._filesPath), [ (0, schematics_1.filter)(filePath => !filePath.endsWith('.DS_Store')), schema.service === 'ignore' ? (0, schematics_1.filter)(filePath => !filePath.endsWith('.service.ts.template')) : (0,
ce, filePath, serviceName, importPath); const declarationRecorder = tree.beginUpdate(filePath); changes.forEach(change => { if (change.path == null) return; if (change instanceof change_1.InsertChange) { declarationRecorder.insertLeft(change.pos, change.toAdd); } }); tree.commitUpdate(declarationRecorder); } exports.addProviderToModule = addProviderToModule; function addValueToVariable(tree, filePath, variableName, text, needWrap
identifier_body
alain.js
? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); } step((generator = generator.apply(thisArg, _arguments || [])).next()); }); }; Object.defineProperty(exports, "__esModule", { value: true }); exports.buildAlain = exports.addValueToVariable = exports.addProviderToModule = exports.addImportToModule = exports.refreshPathRoot = void 0; const core_1 = require("@angular-devkit/core"); const schematics_1 = require("@angular-devkit/schematics"); const ast_utils_1 = require("@schematics/angular/utility/ast-utils"); const change_1 = require("@schematics/angular/utility/change"); const find_module_1 = require("@schematics/angular/utility/find-module"); const parse_name_1 = require("@schematics/angular/utility/parse-name"); const validation_1 = require("@schematics/angular/utility/validation"); const fs = require("fs"); const path = require("path"); const ts = require("typescript"); const ast_1 = require("./ast"); const workspace_1 = require("./workspace"); const TEMPLATE_FILENAME_RE = /\.template$/; function buildSelector(schema, projectPrefix) { const ret = []; if (!schema.withoutPrefix) { if (schema.prefix) { ret.push(schema.prefix); } else if (schema.prefix === undefined && projectPrefix)
} // module name if (schema.module) { ret.push(schema.module); } // target name if (schema.target) { ret.push(...schema.target.split('/')); } // name ret.push(core_1.strings.dasherize(schema.name)); return ret.join('-'); } function buildName(schema, prefix) { const ret = schema.withoutModulePrefixInComponentName === true ? [] : [schema.module]; if (schema.target && schema.target.length > 0) { ret.push(...schema.target.split('/')); } ret.push(schema.name); // 服务类自动过滤 list, empty 两个页面的后缀 if (prefix === 'Service' && ['list', 'empty'].includes(schema.name)) { ret.pop(); } ret.push(prefix); return core_1.strings.classify(ret.join('-')); } function refreshPathRoot(project, schema, alainProject) { var _a; if (schema.path === undefined) { schema.path = `/${path.join(project.sourceRoot, (_a = alainProject === null || alainProject === void 0 ? void 0 : alainProject.routesRoot) !== null && _a !== void 0 ? _a : 'app/routes')}`; } } exports.refreshPathRoot = refreshPathRoot; function resolveSchema(tree, project, schema, alainProject) { // module name if (!schema.module) { throw new schematics_1.SchematicsException(`Must specify module name. (e.g: ng g ng-alain:list <list name> -m=<module name>)`); } // path refreshPathRoot(project, schema, alainProject); schema.path += `/${schema.module}`; const parsedPath = (0, parse_name_1.parseName)(schema.path, schema.name); schema.name = parsedPath.name; schema.path = parsedPath.path; const rootPath = path.resolve(__dirname, '../../..'); const fullPath = path.join(rootPath, schema.path, schema.name); if (fs.existsSync(fullPath) && fs.readdirSync(fullPath).length > 0) { throw new schematics_1.SchematicsException(`The directory (${fullPath}) already exists`); } schema.importModulePath = (0, find_module_1.findModuleFromOptions)(tree, schema); if (!schema._filesPath) { // 若基础页尝试从 `_cli-tpl/_${schema.schematicName!}` 下查找该目录,若存在则优先使用 if (['list', 'edit', 'view', 'empty'].includes(schema.schematicName)) { const overrideDir = `/${[project.root, `_cli-tpl/_${schema.schematicName}`].filter(i => !!i).join('/')}`; const overridePath = `${overrideDir}/__path__/__name@dasherize@if-flat__/__name@dasherize__.component.ts`; if (tree.exists(overridePath) || tree.exists(`${overridePath}.template`)) { // 所在目录与命令目录同属一个目录结构,因此无须特殊处理 schema._filesPath = path.relative(__dirname, rootPath) + overrideDir; } } schema._filesPath = schema._filesPath || './files'; } // fill target if (schema.target) { schema.path += core_1.strings.dasherize(`/${schema.target}`); } schema.routerModulePath = schema.importModulePath.replace('.module.ts', '-routing.module.ts'); // html selector schema.selector = schema.selector || buildSelector(schema, project.prefix); (0, validation_1.validateHtmlSelector)(schema.selector); } function addImportToModule(tree, filePath, symbolName, fileName) { const source = (0, ast_1.getSourceFile)(tree, filePath); const change = (0, ast_utils_1.insertImport)(source, filePath, symbolName, fileName); if (change.path == null) return; const declarationRecorder = tree.beginUpdate(filePath); declarationRecorder.insertLeft(change.pos, change.toAdd); tree.commitUpdate(declarationRecorder); } exports.addImportToModule = addImportToModule; function addProviderToModule(tree, filePath, serviceName, importPath) { const source = (0, ast_1.getSourceFile)(tree, filePath); const changes = (0, ast_utils_1.addProviderToModule)(source, filePath, serviceName, importPath); const declarationRecorder = tree.beginUpdate(filePath); changes.forEach(change => { if (change.path == null) return; if (change instanceof change_1.InsertChange) { declarationRecorder.insertLeft(change.pos, change.toAdd); } }); tree.commitUpdate(declarationRecorder); } exports.addProviderToModule = addProviderToModule; function addValueToVariable(tree, filePath, variableName, text, needWrap = true) { const source = (0, ast_1.getSourceFile)(tree, filePath); const node = (0, ast_utils_1.findNode)(source, ts.SyntaxKind.Identifier, variableName); if (!node) { throw new schematics_1.SchematicsException(`Could not find any [${variableName}] variable in path '${filePath}'.`); } // eslint-disable-next-line @typescript-eslint/no-explicit-any const arr = node.parent.initializer; const change = new change_1.InsertChange(filePath, arr.end - 1, `${arr.elements && arr.elements.length > 0 ? ',' : ''}${needWrap ? '\n ' : ''}${text}`); const declarationRecorder = tree.beginUpdate(filePath); declarationRecorder.insertLeft(change.pos, change.toAdd); tree.commitUpdate(declarationRecorder); } exports.addValueToVariable = addValueToVariable; function getRelativePath(filePath, schema, prefix) { const importPath = `/${schema.path}/${schema.flat ? '' : `${core_1.strings.dasherize(schema.name)}/`}${core_1.strings.dasherize(schema.name)}.${prefix}`; return (0, find_module_1.buildRelativePath)(filePath, importPath); } function addDeclaration(schema) { return (tree) => { if (schema.skipImport || !schema.module) { return tree; } // imports addImportToModule(tree, schema.importModulePath, schema.componentName, getRelativePath(schema.importModulePath, schema, 'component')); addValueToVariable(tree, schema.importModulePath, 'COMPONENTS', schema.componentName); // component if (schema.modal !== true) { // routing addImportToModule(tree, schema.routerModulePath, schema.componentName, getRelativePath(schema.routerModulePath, schema, 'component')); addValueToVariable(tree, schema.routerModulePath, 'routes', `{ path: '${schema.name}', component: ${schema.componentName} }`); } // service if (schema.service === 'none') { addProviderToModule(tree, schema.importModulePath, schema.serviceName, getRelativePath(schema.importModulePath, schema, 'service')); } return tree; }; } function buildAlain(schema) { return (tree) => __awaiter(this, void 0, void 0, function* () { const res = yield (0, workspace_1.getProject)(tree, schema.project); if (schema.project && res.name !== schema.project) { throw new schematics_1.SchematicsException(`The specified project does not match '${schema.project}', current: ${res.name}`); } const project = res.project; resolveSchema(tree, project, schema, res.alainProject); schema.componentName = buildName(schema, 'Component'); schema.serviceName = buildName(schema, 'Service'); // Don't support inline schema.inlineTemplate = false; const templateSource = (0, schematics_1.apply)((0, schematics_1.url)(schema._filesPath), [ (0, schematics_1.filter)(filePath => !filePath.endsWith('.DS_Store')), schema.service === 'ignore' ? (0, schematics_1.filter)(filePath => !filePath.endsWith('.service.ts.template')) : (0,
{ ret.push(projectPrefix); }
conditional_block
alain.js
? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); } step((generator = generator.apply(thisArg, _arguments || [])).next()); }); }; Object.defineProperty(exports, "__esModule", { value: true }); exports.buildAlain = exports.addValueToVariable = exports.addProviderToModule = exports.addImportToModule = exports.refreshPathRoot = void 0; const core_1 = require("@angular-devkit/core"); const schematics_1 = require("@angular-devkit/schematics"); const ast_utils_1 = require("@schematics/angular/utility/ast-utils"); const change_1 = require("@schematics/angular/utility/change"); const find_module_1 = require("@schematics/angular/utility/find-module"); const parse_name_1 = require("@schematics/angular/utility/parse-name"); const validation_1 = require("@schematics/angular/utility/validation"); const fs = require("fs"); const path = require("path"); const ts = require("typescript"); const ast_1 = require("./ast"); const workspace_1 = require("./workspace"); const TEMPLATE_FILENAME_RE = /\.template$/; function buildSelector(schema, projectPrefix) { const ret = []; if (!schema.withoutPrefix) { if (schema.prefix) { ret.push(schema.prefix); } else if (schema.prefix === undefined && projectPrefix) { ret.push(projectPrefix); } } // module name if (schema.module) { ret.push(schema.module); } // target name if (schema.target) { ret.push(...schema.target.split('/')); } // name ret.push(core_1.strings.dasherize(schema.name)); return ret.join('-'); } function buildName(schema, prefix) { const ret = schema.withoutModulePrefixInComponentName === true ? [] : [schema.module]; if (schema.target && schema.target.length > 0) { ret.push(...schema.target.split('/')); } ret.push(schema.name); // 服务类自动过滤 list, empty 两个页面的后缀 if (prefix === 'Service' && ['list', 'empty'].includes(schema.name)) { ret.pop(); } ret.push(prefix); return core_1.strings.classify(ret.join('-')); } function refreshPathRoot(project, schema, alainProject) { var _a; if (schema.path === undefined) { schema.path = `/${path.join(project.sourceRoot, (_a = alainProject === null || alainProject === void 0 ? void 0 : alainProject.routesRoot) !== null && _a !== void 0 ? _a : 'app/routes')}`; } } exports.refreshPathRoot = refreshPathRoot; function resolveSchema(tree, project, schema, alainProject) { // module name if (!schema.module) { throw new schematics_1.SchematicsException(`Must specify module name. (e.g: ng g ng-alain:list <list name> -m=<module name>)`); } // path refreshPathRoot(project, schema, alainProject); schema.path += `/${schema.module}`; const parsedPath = (0, parse_name_1.parseName)(schema.path, schema.name); schema.name = parsedPath.name; schema.path = parsedPath.path; const rootPath = path.resolve(__dirname, '../../..'); const fullPath = path.join(rootPath, schema.path, schema.name); if (fs.existsSync(fullPath) && fs.readdirSync(fullPath).length > 0) { throw new schematics_1.SchematicsException(`The directory (${fullPath}) already exists`); } schema.importModulePath = (0, find_module_1.findModuleFromOptions)(tree, schema); if (!schema._filesPath) { // 若基础页尝试从 `_cli-tpl/_${schema.schematicName!}` 下查找该目录,若存在则优先使用 if (['list', 'edit', 'view', 'empty'].includes(schema.schematicName)) { const overrideDir = `/${[project.root, `_cli-tpl/_${schema.schematicName}`].filter(i => !!i).join('/')}`; const overridePath = `${overrideDir}/__path__/__name@dasherize@if-flat__/__name@dasherize__.component.ts`; if (tree.exists(overridePath) || tree.exists(`${overridePath}.template`)) { // 所在目录与命令目录同属一个目录结构,因此无须特殊处理 schema._filesPath = path.relative(__dirname, rootPath) + overrideDir; } } schema._filesPath = schema._filesPath || './files'; } // fill target if (schema.target) { schema.path += core_1.strings.dasherize(`/${schema.target}`); } schema.routerModulePath = schema.importModulePath.replace('.module.ts', '-routing.module.ts'); // html selector schema.selector = schema.selector || buildSelector(schema, project.prefix); (0, validation_1.validateHtmlSelector)(schema.selector); } function addImportToModule(tree, filePath, symbolName, fileName) { const source = (0, ast_1.getSourceFile)(tree, filePath); const change = (0, ast_utils_1.insertImport)(source, filePath, symbolName, fileName); if (change.path == null) return; const declarationRecorder = tree.beginUpdate(filePath); declarationRecorder.insertLeft(change.pos, change.toAdd); tree.commitUpdate(declarationRecorder); } exports.addImportToModule = addImportToModule; function addProviderToModule(tree, filePath, serviceName, importPath) { const source = (0, ast_1.getSourceFile)(tree, filePath); const changes = (0, ast_utils_1.addProviderToModule)(source, filePath, serviceName, importPath); const declarationRecorder = tree.beginUpdate(filePath); changes.forEach(change => { if (change.path == null) return; if (change instanceof change_1.InsertChange) { declarationRecorder.insertLeft(change.pos, change.toAdd); } }); tree.commitUpdate(declarationRecorder); } exports.addProviderToModule = addProviderToModule; function addValueToVariable(tree, filePath, variableName, text, needWrap = true) { const source = (0, ast_1.getSourceFile)(tree, filePath); const node = (0, ast_utils_1.findNode)(source, ts.SyntaxKind.Identifier, variableName); if (!node) { throw new schematics_1.SchematicsException(`Could not find any [${variableName}] variable in path '${filePath}'.`); } // eslint-disable-next-line @typescript-eslint/no-explicit-any const arr = node.parent.initializer; const change = new change_1.InsertChange(filePath, arr.end - 1, `${arr.elements && arr.elements.length > 0 ? ',' : ''}${needWrap ? '\n ' : ''}${text}`); const declarationRecorder = tree.beginUpdate(filePath); declarationRecorder.insertLeft(change.pos, change.toAdd); tree.commitUpdate(declarationRecorder); } exports.addValueToVariable = addValueToVariable; function getRelativePath(filePath, schema, prefix) { const importPath = `/${schema.path}/${schema.flat ? '' : `${core_1.strings.dasherize(schema.name)}/`}${core_1.strings.dasherize(schema.name)}.${prefix}`; return (0, find_module_1.buildRelativePath)(filePath, importPath); } function addDeclaration(schema) { return (tree) => { if (schema.skipImport || !schema.module) { return tree; } // imports addImportToModule(tree, schema.importModulePath, schema.componentName, getRelativePath(schema.importModulePath, schema, 'component')); addValueToVariable(tree, schema.importModulePath, 'COMPONENTS', schema.componentName); // component if (schema.modal !== true) { // routing addImportToModule(tree, schema.routerModulePath, schema.componentName, getRelativePath(schema.routerModulePath, schema, 'component')); addValueToVariable(tree, schema.routerModulePath, 'routes', `{ path: '${schema.name}', component: ${schema.componentName} }`); } // service if (schema.service === 'none') { addProviderToModule(tree, schema.importModulePath, schema.serviceName, getRelativePath(schema.importModulePath, schema, 'service')); } return tree; }; } function buildAlain(schema) { return (tree) => __awaiter(this, void 0, void 0, function* () { const res = yield (0, workspace_1.getProject)(tree, schema.project); if (schema.project && res.name !== schema.project) { throw new schematics_1.SchematicsException(`The specified project does not match '${schema.project}', current: ${res.name}`); } const project = res.project; resolveSchema(tree, project, schema, res.alainProject); schema.componentName = buildName(schema, 'Component'); schema.serviceName = buildName(schema, 'Service');
schema.service === 'ignore' ? (0, schematics_1.filter)(filePath => !filePath.endsWith('.service.ts.template')) : (0, schematics
// Don't support inline schema.inlineTemplate = false; const templateSource = (0, schematics_1.apply)((0, schematics_1.url)(schema._filesPath), [ (0, schematics_1.filter)(filePath => !filePath.endsWith('.DS_Store')),
random_line_split
plot.py
.png') # Display to screen plt.show() def
(_band, _period): """ Plots an observed band using pyplot :param _band: Array to be plotted :param _period: Period of object """ # Frequency = 1 / Period _freq = 1 / _period _xfit, _lobf = calclobf(_band, _period) # Plot the data in the array to screen, lightly coloured and z rank behind the line of best fit plt.style.use('seaborn-whitegrid') plt.errorbar((_band[:, 0] * _freq) % 1, _band[:, 1], _band[:, 2], fmt='.', color='gray', ecolor='lightgray', capsize=0, zorder=0) # Plot the graph of the line of best fit plt.plot(_xfit, _lobf, '-k', lw=2, zorder=2) # Set x-axis limits to 1 period plt.xlim(0, 1) # Set graph and axis titles plt.xlabel("Phase") plt.ylabel("Magnitude") plt.title("Folded light curve") # Invert y-axis as convention plt.gca().invert_yaxis() # Display to screen plt.show() def calclobf(_band, _period): """ Creates a line of best fit using Lomb-Scargle methods :param _inputband: Band array to be fit :param _period: Period of object :return: Returns a linearly spaced x-axis, with y-axis values for line of best fit """ # Create a model with 10 terms _ls = LombScargle(_band[:, 0], _band[:, 1], _band[:, 2], nterms=10) # Create n linearly spaced points between phase 0 and 1 _xfit = np.linspace(0, 1, 1000) # Frequency = 1 / Period _freq = 1 / _period # Plot the line of best fit generated _lobf = _ls.model(_xfit / _freq, _freq) return _xfit, _lobf def foldcurve(_band, _period): """ Folds the magnitude measurements to a light curve using provided period :param _band: Observation band to be folded :param _period: Period of object :return: Array same size as _band, but with a phase instead of Julian date """ # Set epoch to first date observed _epoch = _band[0][0] # Iterate through array, update date to phase for i in range(0, _band.shape[0]): _band[i, 0] = ((_band[i, 0] - _epoch) / _period) % 1 # Return folded array return _band def doublearrayphase(_inputarray): """ Doubles a band array from phase 0 -> 1 to 0 -> 2 as convention :param _inputarray: Array to be doubled :return: Returns an array from phase 0 -> 2, size [n * 2, 3] """ # Create a new array twice the size of the input _newarray = np.zeros((_inputarray.shape[0] * 2, _inputarray.shape[1]), dtype=float) # Iterate through the input array for i in range(0, _newarray.shape[0]): # Before phase 1 simply copy data into new array if i < _inputarray.shape[0]: _newarray[i] = _inputarray[i] # After phase 1, simply shift all phases by +1 else: _newarray[i] = _inputarray[i - _inputarray.shape[0]] _newarray[i][0] = _newarray[i][0] + 1 # Return the new doubled array return _newarray def plotblackbody(_zband, _yband, _jband, _hband, _kband, _parallax, _perr): """ Determines the black body curve and determines mass, radius and luminosities in solar units :param _zband: Observed z-band :param _yband: Observed y-band :param _jband: Observed j-band :param _hband: Observed h-band :param _kband: Observed k-band :param _parallax: Parallax angle (mas) :param _perr: Parallax angle error (mas) """ # Set pyplot style to be consistent within the program plt.style.use('seaborn-whitegrid') # Import raw data to plot Hertzsprung-Russell diagram _hrdata = inithr('hr.dat') # Determine distance in parsecs _distance = 1 / np.tan(_parallax * 10**-3) _derr = (_perr * 10**-3) / ((_parallax * 10**-3)**2) # Create single data array with all bands _bands = [_zband, _yband, _jband, _hband, _kband] _lambda = [0.9, 1.02, 1.22, 1.63, 2.2] # Set up empty arrays for each star _largestar = np.zeros((1, 2)) _smallstar = np.zeros((1, 2)) # Determine the spectral flux density from the large star i = 0 while i < 5: # Determine the maximum and minimum values of the observed band _max, _min = lightcurve.maxminvals(_bands[i]) # The large star uses the maximum flux value (smallest magnitude) _largestar = np.append(_largestar, np.array([_lambda[i], (magtoflux(_min, i))], ndmin=2), axis=0) i += 1 # Delete first empty row of the array _largestar = np.delete(_largestar, 0, axis=0) # Determine the spectral flux density from the small star i = 0 while i < 5: # Determine the maximum and minimum values of the observed band _max, _min = lightcurve.maxminvals(_bands[i]) # Smaller star flux value is combined value minus the large star _smallstar = np.append(_smallstar, np.array([_lambda[i], (magtoflux(_max, i) - magtoflux(_min, i))], ndmin=2), axis=0) i += 1 # Delete the first empty row of the array _smallstar = np.delete(_smallstar, 0, axis=0) # Determine the luminosity and effective temperature of each star _luma, _lumaerr, _wiena = getwientemp(_largestar, _distance, _derr, 1) _lumb, _lumberr, _wienb = getwientemp(_smallstar, _distance, _derr, 2) # Calculate luminosities in solar units _solluma = _luma / (3.828*10**26) _sollumb = _lumb / (3.828*10**26) _lumaerr = _lumaerr / (3.828*10**26) _lumberr = _lumberr / (3.828*10**26) # Calculate masses using the mass/luminosity relation in solar mass units # N.B. only works as an approximation for main sequence stars, giants and dwarfs are not sutiable for this # approximation _solmassa = np.power(_solluma, 1/3.5) _solmassaerr = ((_solmassa * (1/3.5) * _lumaerr) / _solluma)**2 _solmassb = np.power(_sollumb, 1/3.5) _solmassberr = ((_solmassb * (1 / 3.5) * _lumberr) / _sollumb) ** 2 # Calculate stellar radius in solar radii using the relationship between luminosity, surface area and temperature _solrada = np.sqrt(_solluma / np.power(_wiena / 5778, 4)) _solradb = np.sqrt(_sollumb / np.power(_wienb / 5778, 4)) _solradaerr = ((_solrada * 0.5 * _lumaerr) / _solluma)**2 _solradberr = ((_solradb * 0.5 * _lumberr) / _sollumb)**2 # Output determined values to the screen and write to file print('Values for the large star:') print('Effective temperature: ' + str(round
plotband
identifier_name
plot.py
first empty row of the array _smallstar = np.delete(_smallstar, 0, axis=0) # Determine the luminosity and effective temperature of each star _luma, _lumaerr, _wiena = getwientemp(_largestar, _distance, _derr, 1) _lumb, _lumberr, _wienb = getwientemp(_smallstar, _distance, _derr, 2) # Calculate luminosities in solar units _solluma = _luma / (3.828*10**26) _sollumb = _lumb / (3.828*10**26) _lumaerr = _lumaerr / (3.828*10**26) _lumberr = _lumberr / (3.828*10**26) # Calculate masses using the mass/luminosity relation in solar mass units # N.B. only works as an approximation for main sequence stars, giants and dwarfs are not sutiable for this # approximation _solmassa = np.power(_solluma, 1/3.5) _solmassaerr = ((_solmassa * (1/3.5) * _lumaerr) / _solluma)**2 _solmassb = np.power(_sollumb, 1/3.5) _solmassberr = ((_solmassb * (1 / 3.5) * _lumberr) / _sollumb) ** 2 # Calculate stellar radius in solar radii using the relationship between luminosity, surface area and temperature _solrada = np.sqrt(_solluma / np.power(_wiena / 5778, 4)) _solradb = np.sqrt(_sollumb / np.power(_wienb / 5778, 4)) _solradaerr = ((_solrada * 0.5 * _lumaerr) / _solluma)**2 _solradberr = ((_solradb * 0.5 * _lumberr) / _sollumb)**2 # Output determined values to the screen and write to file print('Values for the large star:') print('Effective temperature: ' + str(round_sig(_wiena))) print('Solar luminosities: ' + str(round_sig(_solluma)) + ', error: ' + str(round_sig(_lumaerr))) print('Solar radii: ' + str(round_sig(_solrada)) + ', error: ' + str(round_sig(_solradaerr))) print('Solar masses: ' + str(round_sig(_solmassa)) + ', error: ' + str(round_sig(_solmassaerr))) print('-----------------------------------------------------') print('Values for the small star:') print('Effective temperature: ' + str(round_sig(_wienb))) print('Solar luminosities: ' + str(round_sig(_sollumb)) + ', error: ' + str(round_sig(_lumberr))) print('Solar radii: ' + str(round_sig(_solradb)) + ', error: ' + str(round_sig(_solradberr))) print('Solar masses: ' + str(round_sig(_solmassb)) + ', error: ' + str(round_sig(_solmassberr))) # Convert from luminosity to magnitude in solar units _luma = -2.5 * np.log10(_luma / (3.0128 * 10**28)) _lumb = -2.5 * np.log10(_lumb / (3.0128 * 10**28)) # Plot Hertzsprung-Russell diagram using provided array plt.scatter(_hrdata[:, 1], _hrdata[:, 0], s=0.5) # Plot determined values for each star plt.scatter(_wiena, _luma, s=16, c='red', label='Larger Star') plt.scatter(_wienb, _lumb, s=16, c='green', label='Smaller Star') # Set the x and y axis limits to sensible values plt.legend() plt.xlim(3000, 10000) plt.ylim(-10, 20) # Invert both axes as convention plt.gca().invert_xaxis() plt.gca().invert_yaxis() # Save figure to current folder plt.savefig('hr.png') # Display to screen plt.show() def getwientemp(_inputdata, _distance, _derr, _id): """ Determines the effective temperature using Wien's law :param _inputdata: Black body curve of object :param _distance: Distance to object (parsecs) :param _id: 1 for large star, 2 for small star :return: Luminosity and effective surface temperature """ # Maxwell-Boltzmann distribution formula probability density function def curve(_x, _a, _scale): _a1 = np.sqrt(2 / np.pi) _a2 = _x**2 / (2 * _a**2) return _scale * _a1 * (_x**2 * np.exp(-_a2)) / _a**3 # Set pyplot style to be consistent through the program plt.style.use('seaborn-whitegrid') # Convert the distance in parsecs to metres _distance = 3.0857 * 10**16 * _distance _derr = 3.0857 * 10**16 * _derr # Create array for x and y axis data _xdata = _inputdata[:, 0] _ydata = _inputdata[:, 1] _ydatalum = _ydata # Iterate through each band and convert from Janskys to W/m^2/um i = 0 while i < 5: _ydata[i] = 3*10**14 * (_ydata[i] * 10**-26) / (Wavelength[i]**2) i += 1 # Calculate optimal values and covariance using scipy curve_fit function _popt, _pcov = curve_fit(curve, _xdata, _ydata) # Create x axis to plot curve against _x = np.linspace(0, 5, 100) # Determine y value for each point on the x axis _yplot = curve(_x, *_popt) # Plot the curve to the screen plt.plot(_x, _yplot) # Determine the area under the graph, integral gives total energy recieved per m^2 _area = np.trapz(_yplot, dx=5/100) # Total luminosity found by multiplying by the surface area of a sphere with diameter of the distance _lum = 4 * np.pi * _distance**2 * _area _lumerr = 4 * np.pi * _distance * _derr * _area # Peak value of Maxwell-Boltzmann distribution _mu = 2 * _popt[0] * np.sqrt(2 / np.pi) # Plot data on the graph plt.plot(_xdata, _ydata, '.') # Set axis labels plt.xlabel('Wavelength (um)') plt.ylabel('Spectral Irradiance (W m^-2 um^-1)') if _id == 1: _str = 'Large Star' else: _str = 'Small Star' # Calculate effective surface temperature using Wien's law _wien = round_sig(2898 / _mu) # Round luminosity to 2 significant figures _lum = round_sig(_lum) # Set graph title plt.suptitle('Black Body Plot for the ' + _str) # Save to current folder _filename = _str + '.png' plt.savefig(_filename) # Display to the screen plt.show() # Returns calculated values return _lum, _lumerr, _wien def inithr(_filename): """ Parses required data for plotting a Hertzsprung-Russell diagram :param _filename: File containing observed data :return: (n x 3) size array containing magnitude, effective temperature and parallax angle """ # Open file provided _file = open(_filename) # Create empty array to hold data _data = np.zeros((1, 3), dtype=float) # Iterate through the file line by line
for _line in _file: # Split each line into constituent values _x = _line.split() # Append data array with each value, converted to float, convert parallax angle to distance _data = np.append(_data, np.array([float(_x[1]), float(_x[2]), (1 / float(_x[3]))], ndmin=2), axis=0)
random_line_split
plot.py
.png') # Display to screen plt.show() def plotband(_band, _period): """ Plots an observed band using pyplot :param _band: Array to be plotted :param _period: Period of object """ # Frequency = 1 / Period _freq = 1 / _period _xfit, _lobf = calclobf(_band, _period) # Plot the data in the array to screen, lightly coloured and z rank behind the line of best fit plt.style.use('seaborn-whitegrid') plt.errorbar((_band[:, 0] * _freq) % 1, _band[:, 1], _band[:, 2], fmt='.', color='gray', ecolor='lightgray', capsize=0, zorder=0) # Plot the graph of the line of best fit plt.plot(_xfit, _lobf, '-k', lw=2, zorder=2) # Set x-axis limits to 1 period plt.xlim(0, 1) # Set graph and axis titles plt.xlabel("Phase") plt.ylabel("Magnitude") plt.title("Folded light curve") # Invert y-axis as convention plt.gca().invert_yaxis() # Display to screen plt.show() def calclobf(_band, _period): """ Creates a line of best fit using Lomb-Scargle methods :param _inputband: Band array to be fit :param _period: Period of object :return: Returns a linearly spaced x-axis, with y-axis values for line of best fit """ # Create a model with 10 terms _ls = LombScargle(_band[:, 0], _band[:, 1], _band[:, 2], nterms=10) # Create n linearly spaced points between phase 0 and 1 _xfit = np.linspace(0, 1, 1000) # Frequency = 1 / Period _freq = 1 / _period # Plot the line of best fit generated _lobf = _ls.model(_xfit / _freq, _freq) return _xfit, _lobf def foldcurve(_band, _period): """ Folds the magnitude measurements to a light curve using provided period :param _band: Observation band to be folded :param _period: Period of object :return: Array same size as _band, but with a phase instead of Julian date """ # Set epoch to first date observed _epoch = _band[0][0] # Iterate through array, update date to phase for i in range(0, _band.shape[0]): _band[i, 0] = ((_band[i, 0] - _epoch) / _period) % 1 # Return folded array return _band def doublearrayphase(_inputarray): """ Doubles a band array from phase 0 -> 1 to 0 -> 2 as convention :param _inputarray: Array to be doubled :return: Returns an array from phase 0 -> 2, size [n * 2, 3] """ # Create a new array twice the size of the input _newarray = np.zeros((_inputarray.shape[0] * 2, _inputarray.shape[1]), dtype=float) # Iterate through the input array for i in range(0, _newarray.shape[0]): # Before phase 1 simply copy data into new array if i < _inputarray.shape[0]: _newarray[i] = _inputarray[i] # After phase 1, simply shift all phases by +1 else: _newarray[i] = _inputarray[i - _inputarray.shape[0]] _newarray[i][0] = _newarray[i][0] + 1 # Return the new doubled array return _newarray def plotblackbody(_zband, _yband, _jband, _hband, _kband, _parallax, _perr): """ Determines the black body curve and determines mass, radius and luminosities in solar units :param _zband: Observed z-band :param _yband: Observed y-band :param _jband: Observed j-band :param _hband: Observed h-band :param _kband: Observed k-band :param _parallax: Parallax angle (mas) :param _perr: Parallax angle error (mas) """ # Set pyplot style to be consistent within the program plt.style.use('seaborn-whitegrid') # Import raw data to plot Hertzsprung-Russell diagram _hrdata = inithr('hr.dat') # Determine distance in parsecs _distance = 1 / np.tan(_parallax * 10**-3) _derr = (_perr * 10**-3) / ((_parallax * 10**-3)**2) # Create single data array with all bands _bands = [_zband, _yband, _jband, _hband, _kband] _lambda = [0.9, 1.02, 1.22, 1.63, 2.2] # Set up empty arrays for each star _largestar = np.zeros((1, 2)) _smallstar = np.zeros((1, 2)) # Determine the spectral flux density from the large star i = 0 while i < 5: # Determine the maximum and minimum values of the observed band
# Delete first empty row of the array _largestar = np.delete(_largestar, 0, axis=0) # Determine the spectral flux density from the small star i = 0 while i < 5: # Determine the maximum and minimum values of the observed band _max, _min = lightcurve.maxminvals(_bands[i]) # Smaller star flux value is combined value minus the large star _smallstar = np.append(_smallstar, np.array([_lambda[i], (magtoflux(_max, i) - magtoflux(_min, i))], ndmin=2), axis=0) i += 1 # Delete the first empty row of the array _smallstar = np.delete(_smallstar, 0, axis=0) # Determine the luminosity and effective temperature of each star _luma, _lumaerr, _wiena = getwientemp(_largestar, _distance, _derr, 1) _lumb, _lumberr, _wienb = getwientemp(_smallstar, _distance, _derr, 2) # Calculate luminosities in solar units _solluma = _luma / (3.828*10**26) _sollumb = _lumb / (3.828*10**26) _lumaerr = _lumaerr / (3.828*10**26) _lumberr = _lumberr / (3.828*10**26) # Calculate masses using the mass/luminosity relation in solar mass units # N.B. only works as an approximation for main sequence stars, giants and dwarfs are not sutiable for this # approximation _solmassa = np.power(_solluma, 1/3.5) _solmassaerr = ((_solmassa * (1/3.5) * _lumaerr) / _solluma)**2 _solmassb = np.power(_sollumb, 1/3.5) _solmassberr = ((_solmassb * (1 / 3.5) * _lumberr) / _sollumb) ** 2 # Calculate stellar radius in solar radii using the relationship between luminosity, surface area and temperature _solrada = np.sqrt(_solluma / np.power(_wiena / 5778, 4)) _solradb = np.sqrt(_sollumb / np.power(_wienb / 5778, 4)) _solradaerr = ((_solrada * 0.5 * _lumaerr) / _solluma)**2 _solradberr = ((_solradb * 0.5 * _lumberr) / _sollumb)**2 # Output determined values to the screen and write to file print('Values for the large star:') print('Effective temperature: ' + str(round_sig(_
_max, _min = lightcurve.maxminvals(_bands[i]) # The large star uses the maximum flux value (smallest magnitude) _largestar = np.append(_largestar, np.array([_lambda[i], (magtoflux(_min, i))], ndmin=2), axis=0) i += 1
conditional_block
plot.py
def plotallbands(_zband, _yband, _jband, _hband, _kband, _period): """ Plots all observed bands to the same graph :param _zband: Observed z-band :param _yband: Observed y-band :param _jband: Observed j-band :param _hband: Observed h-band :param _kband: Observed k-band :param _period: Period of variability """ # Set pyplot style to be consisten within the program plt.style.use('seaborn-whitegrid') # Frequency = 1 / Period _freq = 1 / _period # Create single dataset from all bands _bands = [_zband, _yband, _jband, _hband, _kband] # Iterate through each band and plot to screen i = 0 while i < 5: # Array to set colours for each band _colours = ['-b', '-g', '-r', '-c', '-m'] # Array to set strings for graph legend _legend = ['Z-band', 'Y-band', 'J-band', 'H-band', 'K-band'] # Determine the line of best fit for each band _xfit, _lobf = calclobf(_bands[i], _period) # Plot the data in the array to screen, lightly coloured and z rank behind the line of best fit plt.plot(_xfit, _lobf, _colours[i], lw=1, zorder=2, label=_legend[i]) i += 1 # Set x-axis limit to a single period plt.xlim(0, 1) # Set graph and axis titles plt.xlabel("Phase") plt.ylabel("Magnitude") plt.title("Folded light curve") # Show the legend plt.legend() # Invert y-axis as convention plt.gca().invert_yaxis() # Save to current folder plt.savefig('curve.png') # Display to screen plt.show() def plotband(_band, _period): """ Plots an observed band using pyplot :param _band: Array to be plotted :param _period: Period of object """ # Frequency = 1 / Period _freq = 1 / _period _xfit, _lobf = calclobf(_band, _period) # Plot the data in the array to screen, lightly coloured and z rank behind the line of best fit plt.style.use('seaborn-whitegrid') plt.errorbar((_band[:, 0] * _freq) % 1, _band[:, 1], _band[:, 2], fmt='.', color='gray', ecolor='lightgray', capsize=0, zorder=0) # Plot the graph of the line of best fit plt.plot(_xfit, _lobf, '-k', lw=2, zorder=2) # Set x-axis limits to 1 period plt.xlim(0, 1) # Set graph and axis titles plt.xlabel("Phase") plt.ylabel("Magnitude") plt.title("Folded light curve") # Invert y-axis as convention plt.gca().invert_yaxis() # Display to screen plt.show() def calclobf(_band, _period): """ Creates a line of best fit using Lomb-Scargle methods :param _inputband: Band array to be fit :param _period: Period of object :return: Returns a linearly spaced x-axis, with y-axis values for line of best fit """ # Create a model with 10 terms _ls = LombScargle(_band[:, 0], _band[:, 1], _band[:, 2], nterms=10) # Create n linearly spaced points between phase 0 and 1 _xfit = np.linspace(0, 1, 1000) # Frequency = 1 / Period _freq = 1 / _period # Plot the line of best fit generated _lobf = _ls.model(_xfit / _freq, _freq) return _xfit, _lobf def foldcurve(_band, _period): """ Folds the magnitude measurements to a light curve using provided period :param _band: Observation band to be folded :param _period: Period of object :return: Array same size as _band, but with a phase instead of Julian date """ # Set epoch to first date observed _epoch = _band[0][0] # Iterate through array, update date to phase for i in range(0, _band.shape[0]): _band[i, 0] = ((_band[i, 0] - _epoch) / _period) % 1 # Return folded array return _band def doublearrayphase(_inputarray): """ Doubles a band array from phase 0 -> 1 to 0 -> 2 as convention :param _inputarray: Array to be doubled :return: Returns an array from phase 0 -> 2, size [n * 2, 3] """ # Create a new array twice the size of the input _newarray = np.zeros((_inputarray.shape[0] * 2, _inputarray.shape[1]), dtype=float) # Iterate through the input array for i in range(0, _newarray.shape[0]): # Before phase 1 simply copy data into new array if i < _inputarray.shape[0]: _newarray[i] = _inputarray[i] # After phase 1, simply shift all phases by +1 else: _newarray[i] = _inputarray[i - _inputarray.shape[0]] _newarray[i][0] = _newarray[i][0] + 1 # Return the new doubled array return _newarray def plotblackbody(_zband, _yband, _jband, _hband, _kband, _parallax, _perr): """ Determines the black body curve and determines mass, radius and luminosities in solar units :param _zband: Observed z-band :param _yband: Observed y-band :param _jband: Observed j-band :param _hband: Observed h-band :param _kband: Observed k-band :param _parallax: Parallax angle (mas) :param _perr: Parallax angle error (mas) """ # Set pyplot style to be consistent within the program plt.style.use('seaborn-whitegrid') # Import raw data to plot Hertzsprung-Russell diagram _hrdata = inithr('hr.dat') # Determine distance in parsecs _distance = 1 / np.tan(_parallax * 10**-3) _derr = (_perr * 10**-3) / ((_parallax * 10**-3)**2) # Create single data array with all bands _bands = [_zband, _yband, _jband, _hband, _kband] _lambda = [0.9, 1.02, 1.22, 1.63, 2.2] # Set up empty arrays for each star _largestar = np.zeros((1, 2)) _smallstar = np.zeros((1, 2)) # Determine the spectral flux density from the large star i = 0 while i < 5: # Determine the maximum and minimum values of the observed band _max, _min = lightcurve.maxminvals(_bands[i]) # The large star uses the maximum flux value (smallest magnitude) _largestar = np.append(_largestar, np.array([_lambda[i], (magtoflux(_min, i))], ndmin=2), axis=0) i += 1 # Delete first empty row of the array _largestar = np.delete(_largestar, 0, axis=0) # Determine the spectral flux density from the small star i = 0 while i < 5: # Determine the maximum and minimum values of the observed band _max, _min = lightcurve.maxminvals(_bands[i]) # Smaller star flux value is combined value minus the large star _smallstar = np.append(_smallstar, np.array([_lambda[i], (magtoflux(_max, i) - magtoflux(_min, i))], ndmin=2), axis=0) i += 1 # Delete the first empty row of the array _smallstar = np.delete(_small
""" Rounds the provided value to 2 significant figures :param _val: Value to be rounded :return: Float, original value rounded to 2 significant figures """ return round(_val, 3 - int(floor(log10(abs(_val)))) - 1)
identifier_body
routing.rs
nodes, we know our incoming edges have already been updated to point to the egress nodes. let mut ingresses = HashMap::new(); for node in topo_list { let domain = graph[node].domain(); // First, we add egress nodes for any of our cross-domain children let children: Vec<_> = graph.neighbors_directed(node, petgraph::EdgeDirection::Outgoing) .collect(); // collect so we can mutate graph // We then need to make sure that we're acting on up-to-date information about existing // egress/ingress pairs. In particular, we want to know about egresses this node already // has (and to which domains). In the process we also populate the information about // ingress nodes in other domains that point here (those shouldn't be re-created if new // nodes are created in the corresponding domains). let mut egresses = HashMap::new(); for child in &children { if !new.contains(child) { continue; } if let node::Type::Egress { .. } = *graph[*child] { for ingress in graph.neighbors_directed(*child, petgraph::EdgeDirection::Outgoing) { // this egress already contains this node to the ingress' domain egresses.insert(graph[ingress].domain(), *child); // also keep track of the corresponding ingress node so we can re-use it ingresses.entry(graph[ingress].domain()) .or_insert_with(HashMap::new) .insert(node, ingress); } } } for child in children { let cdomain = graph[child].domain(); if domain != cdomain { // child is in a different domain if !egresses.contains_key(&cdomain) { // create an egress node to handle that // NOTE: technically, this doesn't need to mirror its parent, but meh let proxy = graph[node].mirror(node::Type::Egress { tags: Default::default(), txs: Default::default(), }); let egress = graph.add_node(proxy); graph.add_edge(node, egress, false); new.insert(egress); egresses.insert(cdomain, egress); trace!(log, "adding cross-domain egress to new node"; "node" => node.index(), "egress" => egress.index()); } else { trace!(log, "re-using cross-domain egress to new node"; "node" => node.index(), "egress" => egresses[&cdomain].index()); } // we need to hook that node in between us and this child let egress = egresses[&cdomain]; let old = graph.find_edge(node, child).unwrap(); let was_materialized = graph.remove_edge(old).unwrap(); graph.add_edge(egress, child, was_materialized); // this ends up being re-executed, but that's okay swaps.entry(cdomain).or_insert_with(HashMap::new).insert(node, egress); } } // Then, we look for any parents in the graph that // // a) are in a different domain, and // b) aren't egress nodes // // This situation arises whenever a cross-domain edge is added as the result of a // migration. We need to find or make an egress domain in that other domain, and hook that // up as the parent of this node instead of the original internal foreign domain node. // // Note that same-domain parents are never interesting to us for this purpose. let mut parents: Vec<_> = graph.neighbors_directed(node, petgraph::EdgeDirection::Incoming) .filter(|&ni| ni == source || graph[ni].domain() != domain) .collect(); // collect so we can mutate graph for parent in &mut parents { if *parent == source { // no egress needed continue; } // since we are traversing in topological order, egress nodes should have been added to // all our parents, and our incoming edges should have been updated. if that *isn't* // the case for a given parent, it must be a pre-existing parent. if let node::Type::Egress { .. } = *graph[*parent] { continue; } // let's first see if this parent already has an egress we can use let egress = graph.neighbors_directed(*parent, petgraph::EdgeDirection::Outgoing) .find(|&ni| graph[ni].is_egress()); let egress = egress.unwrap_or_else(|| { // no, okay, so we need to add an egress for that other node, let proxy = graph[*parent].mirror(node::Type::Egress{ txs: Default::default(), tags: Default::default() }); let egress = graph.add_node(proxy); trace!(log, "adding cross-domain egress to existing node"; "node" => parent.index(), "egress" => egress.index()); graph.add_edge(*parent, egress, false); new.insert(egress); egress }); // now, let's use that egress as our parent instead let old = graph.find_edge(*parent, node).unwrap(); let was_materialized = graph.remove_edge(old).unwrap(); graph.add_edge(egress, node, was_materialized); // all references to our original parent should now refer to the egress swaps.entry(domain).or_insert_with(HashMap::new).insert(*parent, egress); // and we should now just consider the egress our parent instead *parent = egress; } // Now that we know all our foreign parents are egress nodes, we can add ingress nodes. // Note that by this time (due to the topological walk), we know that `ingresses` has been // sufficiently populated to contain any relevant existing ingress nodes. for parent in parents { // is there already an ingress node we can re-use? let mut ingress = ingresses.get(&domain).and_then(|ingresses| ingresses.get(&parent)).map(|ni| *ni); if ingress.is_none() { // nope -- create our new ingress node let mut i = graph[parent].mirror(node::Type::Ingress); i.add_to(domain); // it belongs to this domain, not that of the parent let i = graph.add_node(i); graph.add_edge(parent, i, false); // we also now need to deal with this ingress node new.insert(i); if parent == source { trace!(log, "adding source ingress"; "base" => node.index(), "ingress" => i.index()); // we don't re-use source ingress nodes } else { trace!(log, "adding cross-domain ingress"; "to" => node.index(), "from" => parent.index(), "ingress" => i.index()); ingresses.entry(domain).or_insert_with(HashMap::new).insert(parent, i); } ingress = Some(i); } else { trace!(log, "re-using cross-domain ingress"; "to" => node.index(), "from" => parent.index(), "ingress" => ingress.unwrap().index()); } let ingress = ingress.unwrap(); // we need to hook the ingress node in between us and the parent let old = graph.find_edge(parent, node).unwrap(); let was_materialized = graph.remove_edge(old).unwrap(); graph.add_edge(ingress, node, was_materialized); // tracking swaps here is a bit tricky because we've already swapped the "true" parents // of `node` with the ids of the egress nodes. thus, we actually need to do swaps on // the values in `swaps`, not insert new entries (that, or we'd need to change the // resolution process to be recursive, which is painful and unnecessary). note that we // *also* need to special-case handing base nodes, because there there *won't* be a // parent egress swap if parent != source { for (_, to) in swaps.get_mut(&domain).unwrap().iter_mut() { if *to == parent { *to = ingress; } } } } } swaps } pub fn connect(log: &Logger, graph: &mut Graph, main_txs: &HashMap<domain::Index, mpsc::SyncSender<Packet>>, new: &HashSet<NodeIndex>)
{ // ensure all egress nodes contain the tx channel of the domains of their child ingress nodes for &node in new { let n = &graph[node]; if let node::Type::Ingress = **n { // check the egress connected to this ingress } else { continue; } for egress in graph.neighbors_directed(node, petgraph::EdgeDirection::Incoming) { match *graph[egress] { node::Type::Egress { ref txs, .. } => { trace!(log, "connecting"; "egress" => egress.index(), "ingress" => node.index()); txs.lock() .unwrap() .push((node.into(), n.addr(), main_txs[&n.domain()].clone())); continue; }
identifier_body
routing.rs
*graph); while let Some(node) = topo.next(&*graph) { if node == source { continue; } if !new.contains(&node) { continue; } topo_list.push(node); } // we need to keep track of all the times we change the parent of a node (by replacing it with // an egress, and then with an ingress), since this remapping must be communicated to the nodes // so they know the true identifier of their parent in the graph. let mut swaps = HashMap::new(); // we also need to keep track of the ingress nodes we've added to each domain so that we don't // end up with two ingress nodes for a given egress node. that would cause unnecessary // cross-domain communication. this is domain => source => NodeIndex (of ingress). note that // `source` here is actually the *egress* node. this is because by the time we add ingress // nodes, we know our incoming edges have already been updated to point to the egress nodes. let mut ingresses = HashMap::new(); for node in topo_list { let domain = graph[node].domain(); // First, we add egress nodes for any of our cross-domain children let children: Vec<_> = graph.neighbors_directed(node, petgraph::EdgeDirection::Outgoing) .collect(); // collect so we can mutate graph // We then need to make sure that we're acting on up-to-date information about existing // egress/ingress pairs. In particular, we want to know about egresses this node already // has (and to which domains). In the process we also populate the information about // ingress nodes in other domains that point here (those shouldn't be re-created if new // nodes are created in the corresponding domains). let mut egresses = HashMap::new(); for child in &children { if !new.contains(child) { continue; } if let node::Type::Egress { .. } = *graph[*child] { for ingress in graph.neighbors_directed(*child, petgraph::EdgeDirection::Outgoing) { // this egress already contains this node to the ingress' domain egresses.insert(graph[ingress].domain(), *child); // also keep track of the corresponding ingress node so we can re-use it ingresses.entry(graph[ingress].domain()) .or_insert_with(HashMap::new) .insert(node, ingress); } } } for child in children { let cdomain = graph[child].domain(); if domain != cdomain { // child is in a different domain if !egresses.contains_key(&cdomain) { // create an egress node to handle that // NOTE: technically, this doesn't need to mirror its parent, but meh let proxy = graph[node].mirror(node::Type::Egress { tags: Default::default(), txs: Default::default(), }); let egress = graph.add_node(proxy); graph.add_edge(node, egress, false); new.insert(egress); egresses.insert(cdomain, egress); trace!(log, "adding cross-domain egress to new node"; "node" => node.index(), "egress" => egress.index()); } else { trace!(log, "re-using cross-domain egress to new node"; "node" => node.index(), "egress" => egresses[&cdomain].index()); } // we need to hook that node in between us and this child let egress = egresses[&cdomain]; let old = graph.find_edge(node, child).unwrap(); let was_materialized = graph.remove_edge(old).unwrap(); graph.add_edge(egress, child, was_materialized); // this ends up being re-executed, but that's okay swaps.entry(cdomain).or_insert_with(HashMap::new).insert(node, egress); } } // Then, we look for any parents in the graph that // // a) are in a different domain, and // b) aren't egress nodes // // This situation arises whenever a cross-domain edge is added as the result of a // migration. We need to find or make an egress domain in that other domain, and hook that // up as the parent of this node instead of the original internal foreign domain node. // // Note that same-domain parents are never interesting to us for this purpose. let mut parents: Vec<_> = graph.neighbors_directed(node, petgraph::EdgeDirection::Incoming) .filter(|&ni| ni == source || graph[ni].domain() != domain) .collect(); // collect so we can mutate graph for parent in &mut parents { if *parent == source { // no egress needed continue; } // since we are traversing in topological order, egress nodes should have been added to // all our parents, and our incoming edges should have been updated. if that *isn't* // the case for a given parent, it must be a pre-existing parent. if let node::Type::Egress { .. } = *graph[*parent] { continue; } // let's first see if this parent already has an egress we can use let egress = graph.neighbors_directed(*parent, petgraph::EdgeDirection::Outgoing) .find(|&ni| graph[ni].is_egress()); let egress = egress.unwrap_or_else(|| { // no, okay, so we need to add an egress for that other node, let proxy = graph[*parent].mirror(node::Type::Egress{ txs: Default::default(), tags: Default::default() }); let egress = graph.add_node(proxy); trace!(log, "adding cross-domain egress to existing node"; "node" => parent.index(), "egress" => egress.index()); graph.add_edge(*parent, egress, false); new.insert(egress); egress }); // now, let's use that egress as our parent instead let old = graph.find_edge(*parent, node).unwrap(); let was_materialized = graph.remove_edge(old).unwrap(); graph.add_edge(egress, node, was_materialized); // all references to our original parent should now refer to the egress swaps.entry(domain).or_insert_with(HashMap::new).insert(*parent, egress); // and we should now just consider the egress our parent instead *parent = egress; } // Now that we know all our foreign parents are egress nodes, we can add ingress nodes. // Note that by this time (due to the topological walk), we know that `ingresses` has been // sufficiently populated to contain any relevant existing ingress nodes. for parent in parents { // is there already an ingress node we can re-use? let mut ingress = ingresses.get(&domain).and_then(|ingresses| ingresses.get(&parent)).map(|ni| *ni); if ingress.is_none() { // nope -- create our new ingress node let mut i = graph[parent].mirror(node::Type::Ingress); i.add_to(domain); // it belongs to this domain, not that of the parent let i = graph.add_node(i); graph.add_edge(parent, i, false); // we also now need to deal with this ingress node new.insert(i); if parent == source { trace!(log, "adding source ingress"; "base" => node.index(), "ingress" => i.index()); // we don't re-use source ingress nodes } else { trace!(log, "adding cross-domain ingress"; "to" => node.index(), "from" => parent.index(), "ingress" => i.index()); ingresses.entry(domain).or_insert_with(HashMap::new).insert(parent, i); } ingress = Some(i); } else { trace!(log, "re-using cross-domain ingress"; "to" => node.index(), "from" => parent.index(), "ingress" => ingress.unwrap().index()); } let ingress = ingress.unwrap(); // we need to hook the ingress node in between us and the parent let old = graph.find_edge(parent, node).unwrap(); let was_materialized = graph.remove_edge(old).unwrap(); graph.add_edge(ingress, node, was_materialized); // tracking swaps here is a bit tricky because we've already swapped the "true" parents // of `node` with the ids of the egress nodes. thus, we actually need to do swaps on // the values in `swaps`, not insert new entries (that, or we'd need to change the // resolution process to be recursive, which is painful and unnecessary). note that we // *also* need to special-case handing base nodes, because there there *won't* be a // parent egress swap if parent != source { for (_, to) in swaps.get_mut(&domain).unwrap().iter_mut() { if *to == parent { *to = ingress; } } } } } swaps } pub fn
connect
identifier_name
routing.rs
domain::Index, HashMap<NodeIndex, NodeIndex>> { // find all new nodes in topological order. we collect first since we'll be mutating the graph // below. it's convenient to have the nodes in topological order, because we then know that // we'll first add egress nodes, and then the related ingress nodes. if we're ever required to // add an ingress node, and its parent isn't an egress node, we know that we're seeing a // connection between an old node in one domain, and a new node in a different domain. let mut topo_list = Vec::with_capacity(new.len()); let mut topo = petgraph::visit::Topo::new(&*graph); while let Some(node) = topo.next(&*graph) { if node == source { continue; } if !new.contains(&node) { continue; } topo_list.push(node); } // we need to keep track of all the times we change the parent of a node (by replacing it with // an egress, and then with an ingress), since this remapping must be communicated to the nodes // so they know the true identifier of their parent in the graph. let mut swaps = HashMap::new(); // we also need to keep track of the ingress nodes we've added to each domain so that we don't // end up with two ingress nodes for a given egress node. that would cause unnecessary // cross-domain communication. this is domain => source => NodeIndex (of ingress). note that // `source` here is actually the *egress* node. this is because by the time we add ingress // nodes, we know our incoming edges have already been updated to point to the egress nodes. let mut ingresses = HashMap::new(); for node in topo_list { let domain = graph[node].domain(); // First, we add egress nodes for any of our cross-domain children let children: Vec<_> = graph.neighbors_directed(node, petgraph::EdgeDirection::Outgoing) .collect(); // collect so we can mutate graph // We then need to make sure that we're acting on up-to-date information about existing // egress/ingress pairs. In particular, we want to know about egresses this node already // has (and to which domains). In the process we also populate the information about // ingress nodes in other domains that point here (those shouldn't be re-created if new // nodes are created in the corresponding domains). let mut egresses = HashMap::new(); for child in &children { if !new.contains(child) { continue; } if let node::Type::Egress { .. } = *graph[*child] { for ingress in graph.neighbors_directed(*child, petgraph::EdgeDirection::Outgoing) { // this egress already contains this node to the ingress' domain egresses.insert(graph[ingress].domain(), *child); // also keep track of the corresponding ingress node so we can re-use it ingresses.entry(graph[ingress].domain()) .or_insert_with(HashMap::new) .insert(node, ingress); } } } for child in children { let cdomain = graph[child].domain(); if domain != cdomain { // child is in a different domain if !egresses.contains_key(&cdomain) { // create an egress node to handle that // NOTE: technically, this doesn't need to mirror its parent, but meh let proxy = graph[node].mirror(node::Type::Egress { tags: Default::default(), txs: Default::default(), }); let egress = graph.add_node(proxy); graph.add_edge(node, egress, false); new.insert(egress); egresses.insert(cdomain, egress); trace!(log, "adding cross-domain egress to new node"; "node" => node.index(), "egress" => egress.index()); } else { trace!(log, "re-using cross-domain egress to new node"; "node" => node.index(), "egress" => egresses[&cdomain].index()); } // we need to hook that node in between us and this child let egress = egresses[&cdomain]; let old = graph.find_edge(node, child).unwrap(); let was_materialized = graph.remove_edge(old).unwrap(); graph.add_edge(egress, child, was_materialized); // this ends up being re-executed, but that's okay swaps.entry(cdomain).or_insert_with(HashMap::new).insert(node, egress); } } // Then, we look for any parents in the graph that // // a) are in a different domain, and // b) aren't egress nodes // // This situation arises whenever a cross-domain edge is added as the result of a // migration. We need to find or make an egress domain in that other domain, and hook that // up as the parent of this node instead of the original internal foreign domain node. // // Note that same-domain parents are never interesting to us for this purpose. let mut parents: Vec<_> = graph.neighbors_directed(node, petgraph::EdgeDirection::Incoming) .filter(|&ni| ni == source || graph[ni].domain() != domain) .collect(); // collect so we can mutate graph for parent in &mut parents { if *parent == source { // no egress needed continue; } // since we are traversing in topological order, egress nodes should have been added to // all our parents, and our incoming edges should have been updated. if that *isn't* // the case for a given parent, it must be a pre-existing parent. if let node::Type::Egress { .. } = *graph[*parent] { continue; } // let's first see if this parent already has an egress we can use let egress = graph.neighbors_directed(*parent, petgraph::EdgeDirection::Outgoing) .find(|&ni| graph[ni].is_egress()); let egress = egress.unwrap_or_else(|| { // no, okay, so we need to add an egress for that other node, let proxy = graph[*parent].mirror(node::Type::Egress{ txs: Default::default(), tags: Default::default() }); let egress = graph.add_node(proxy); trace!(log, "adding cross-domain egress to existing node"; "node" => parent.index(), "egress" => egress.index()); graph.add_edge(*parent, egress, false); new.insert(egress); egress }); // now, let's use that egress as our parent instead let old = graph.find_edge(*parent, node).unwrap(); let was_materialized = graph.remove_edge(old).unwrap(); graph.add_edge(egress, node, was_materialized); // all references to our original parent should now refer to the egress swaps.entry(domain).or_insert_with(HashMap::new).insert(*parent, egress); // and we should now just consider the egress our parent instead *parent = egress; } // Now that we know all our foreign parents are egress nodes, we can add ingress nodes. // Note that by this time (due to the topological walk), we know that `ingresses` has been // sufficiently populated to contain any relevant existing ingress nodes. for parent in parents { // is there already an ingress node we can re-use? let mut ingress = ingresses.get(&domain).and_then(|ingresses| ingresses.get(&parent)).map(|ni| *ni); if ingress.is_none() { // nope -- create our new ingress node let mut i = graph[parent].mirror(node::Type::Ingress); i.add_to(domain); // it belongs to this domain, not that of the parent let i = graph.add_node(i); graph.add_edge(parent, i, false); // we also now need to deal with this ingress node new.insert(i); if parent == source { trace!(log, "adding source ingress"; "base" => node.index(), "ingress" => i.index()); // we don't re-use source ingress nodes } else { trace!(log, "adding cross-domain ingress"; "to" => node.index(), "from" => parent.index(), "ingress" => i.index()); ingresses.entry(domain).or_insert_with(HashMap::new).insert(parent, i); } ingress = Some(i); } else { trace!(log, "re-using cross-domain ingress"; "to" => node.index(), "from" => parent.index(), "ingress" => ingress.unwrap().index()); } let ingress = ingress.unwrap(); // we need to hook the ingress node in between us and the parent let old = graph.find_edge(parent, node).unwrap();
let was_materialized = graph.remove_edge(old).unwrap(); graph.add_edge(ingress, node, was_materialized); // tracking swaps here is a bit tricky because we've already swapped the "true" parents
random_line_split
main.rs
*response.body_mut() = Body::wrap_stream(stream_fuck); //*response.body_mut() = Body::empty(); //*response.set_body(Box::new(stream)); // type BoxFut = Box<Future<Item = Response<Body>, Error = hyper::Error> + Send>; //let future_result = future::ok(response); // let mut buf = BytesMut::with_capacity(1024); // buf.put(&b"hello world"[..]); //let mut response1 = Response::new("Fuck"); // let mut response1 = Response::new(Body::from(buf.freeze())); // let future_result: FutureResult<Response<Body>, hyper::Error> = future::ok(response1); //return Box::new( future_result); //let (method, uri, version, headers, body) = req.deconstruct(); let myresp = chunk::handle_request(Request::new(Body::from("Fuck ya to chunk::handle_request"))); return Box::new(myresp); //let future_result: FutureResult<Response<Body>, hyper::Error> = future::ok(response); //return Box::new(future_result); } // Simply echo the body back to the client. (&Method::POST, "/echo") => { *response.body_mut() = req.into_body(); } //(&Method::GET, Some("/fwd/")) => { // *response.body_mut() = Body::from("Jahahahahaha"); //} // Convert to uppercase before sending back to client. (&Method::POST, "/echo/uppercase") => { let mapping = req.into_body().map(|chunk| { chunk .iter() .map(|byte| byte.to_ascii_uppercase()) .collect::<Vec<u8>>() }); *response.body_mut() = Body::wrap_stream(mapping); } // Reverse the entire body before sending back to the client. // // Since we don't know the end yet, we can't simply stream // the chunks as they arrive. So, this returns a different // future, waiting on concatenating the full body, so that // it can be reversed. Only then can we return a `Response`. (&Method::POST, "/echo/reversed") => { let reversed = req.into_body().concat2().map(move |chunk| { let body = chunk.iter().rev().cloned().collect::<Vec<u8>>(); *response.body_mut() = Body::from(body); response }); return Box::new(reversed); } // The 404 Not Found route... _ => { println!("404 not found."); *response.status_mut() = StatusCode::NOT_FOUND; } }; Box::new(future::ok(response)) } /*struct ChunkDecoder; impl Decoder for ChunkDecoder { type Item = Chunk; type Error = io::Error; fn decode(&mut self, buf: &mut bytes::BytesMut) -> Result<Option<Chunk>, io::Error> { let len = buf.len(); if len > 0 { Ok(Some(buf.take().freeze().into())) } else { Ok(None) } } }*/ struct MyStream { current: u32, max: u32, } impl MyStream { pub fn new(max: u32) -> MyStream { MyStream { current: 0, max: max, } } } impl Stream for MyStream { type Item = u32; type Error = Box<Error>; fn poll(&mut self) -> Poll<Option<Self::Item>, Self::Error> { match self.current { ref mut x if *x < self.max => { *x = *x + 1; Ok(Async::Ready(Some(*x))) } _ => Ok(Async::Ready(None)), } } } /*fn get_local_mp3_path() -> &str { //let mut f = File::open("./p.mp3").expect("failed to open mp3 file!"); let mut filepath = "/media/hdd/jedzia/rust/p.mp3"; if cfg!(target_os = "windows") { filepath = "p.mp3"; } filepath }*/ fn main() -> Result<(), Box<dyn Error>> { println!("Hello, lovely VU Duo!"); let y: f32 = 5.0; if y < 4.0 { // https://stackoverflow.com/questions/51550167/how-to-manually-return-a-result-boxerror return Err("Bad request".into()); } // http://192.168.2.43:3000/ //let addr = ([0, 0, 0, 0], 3000).into(); //let addr = ([127, 0, 0, 1], 3000).into(); //let addr = ([192, 168, 2, 43], 3000).into(); pretty_env_logger::init(); //fun_with_ssl(); //return Ok(()); /* // helps when certificates are not found extern crate openssl_probe; let ssl = openssl_probe::init_ssl_cert_env_vars(); */ //let mut buffer = String::new(); //f.read_to_string(&mut buffer)?; //let in_addr: SocketAddr = ([127, 0, 0, 1], 3333).into(); let in_addr = get_in_addr(); /*let out_addr: SocketAddr = ([127, 0, 0, 1], 3000).into(); // google.de 216.58.208.35 //let out_addr: SocketAddr = ([216, 58, 208, 35], 443).into(); let client_main = Client::new(); let out_addr_clone = out_addr.clone(); // new_service is run for each connection, creating a 'service' // to handle requests for that specific connection. let new_service = move || { let client = client_main.clone(); // This is the `Service` that will handle the connection. // `service_fn_ok` is a helper to convert a function that // returns a Response into a `Service`. service_fn(move |mut req| { let uri_string = format!( "http://{}/{}", out_addr_clone, req.uri().path_and_query().map(|x| x.as_str()).unwrap_or("") ); let uri = uri_string.parse().unwrap(); let in_uri_string = format!("http://{}/{}", in_addr, req.uri()); let in_remove_string = format!("http://{}//", in_addr); println!("req.uri(): {}", in_uri_string); let result = in_uri_string.replace(&in_remove_string, ""); //let result = in_uri_string.split(in_remove_string.unwrap_or("")).take(1).next().unwrap_or(""); println!("result: {}", result); *req.uri_mut() = uri; client.request(req) }) }; let server = Server::bind(&in_addr) .serve(new_service) .map_err(|e| eprintln!("server error: {}", e)); println!("Listening on http://{}", in_addr); println!("Proxying on http://{}", out_addr); rt::run(server);*/ //let mut f = File::open("p.mp3")?; //let mut buffer: Vec<u8> = Vec::new(); //f.read_to_end(&mut buffer)?; //let b = buffer.clone(); let server = Server::bind(&in_addr) //.serve(|| service_fn(|req| echo(req, Vec::new()))) .serve(|| service_fn(echo)) .map_err(|e| eprintln!("server error: {}", e)); println!("Listening on http://{}", in_addr); hyper::rt::run(server); println!("finished."); Ok(()) } /*//#cfg!(target_os = "windows") fn testOpenSSL1() { extern crate openssl; use openssl::rsa::{Padding, Rsa}; let rsa = Rsa::generate(2048).unwrap(); let data = b"foobar"; println!("data {:?}", data); let mut buf = vec![0; rsa.size() as usize]; let encrypted_len = rsa.public_encrypt(data, &mut buf, Padding::PKCS1).unwrap(); println!("encripted {:?}", buf); }*/ fn testOpenSSL()
{ /* extern crate openssl; println!("===== testOpenSSL ====="); use openssl::ssl::{SslConnector, SslMethod}; use std::io::{Read, Write}; use std::net::TcpStream; let connector = SslConnector::builder(SslMethod::tls()).unwrap().build(); let stream = TcpStream::connect("google.com:443").unwrap(); let mut stream = connector.connect("google.com", stream).unwrap(); stream.write_all(b"GET / HTTP/1.0\r\n\r\n").unwrap(); let mut res = vec![]; stream.read_to_end(&mut res).unwrap(); println!( "{}", String::from_utf8_lossy(&res) .lines()
identifier_body
main.rs
(uri: &Uri) -> String { //let in_addr: SocketAddr = get_in_addr(); let uri_string = uri.path_and_query().map(|x| x.as_str()).unwrap_or(""); //let uri: String = uri_string.parse().unwrap(); //let in_uri_string = format!("http://{}/{}", in_addr, req.uri()); let in_remove_string = "/fwd/"; debug!("uri_string: {}", uri_string); let result = uri_string.replace(&in_remove_string, ""); debug!("result: {}", result); //let result = in_uri_string.split(in_remove_string.unwrap_or("")).take(1).next().unwrap_or(""); result } fn reformat_dates(before: &str) -> Cow<str> { lazy_static! { static ref ISO8601_DATE_REGEX : Regex = Regex::new( //r"(?P<y>\d{4})-(?P<m>\d{2})-(?P<d>\d{2})" //r"/^(?:(?:(?:https?|ftp):)?\/\/)(?:\S+(?::\S*)?@)?(?:(?!(?:10|127)(?:\.\d{1,3}){3})(?!(?:169\.254|192\.168)(?:\.\d{1,3}){2})(?!172\.(?:1[6-9]|2\d|3[0-1])(?:\.\d{1,3}){2})(?:[1-9]\d?|1\d\d|2[01]\d|22[0-3])(?:\.(?:1?\d{1,2}|2[0-4]\d|25[0-5])){2}(?:\.(?:[1-9]\d?|1\d\d|2[0-4]\d|25[0-4]))|(?:(?:[a-z0-9\u00a1-\uffff][a-z0-9\u00a1-\uffff_-]{0,62})?[a-z0-9\u00a1-\uffff]\.)+(?:[a-z\u00a1-\uffff]{2,}\.?))(?::\d{2,5})?(?:[/?#]\S*)?$/i" //r"(?:(?:(?:https?|ftp):)?\/\/)(?:\S+(?::\S*)?@)?(?:(?!(?:10|127)(?:\.\d{1,3}){3})(?!(?:169\.254|192\.168)(?:\.\d{1,3}){2})(?!172\.(?:1[6-9]|2\d|3[0-1])(?:\.\d{1,3}){2})(?:[1-9]\d?|1\d\d|2[01]\d|22[0-3])(?:\.(?:1?\d{1,2}|2[0-4]\d|25[0-5])){2}(?:\.(?:[1-9]\d?|1\d\d|2[0-4]\d|25[0-4]))|(?:(?:[a-z0-9\u00a1-\uffff][a-z0-9\u00a1-\uffff_-]{0,62})?[a-z0-9\u00a1-\uffff]\.)+(?:[a-z\u00a1-\uffff]{2,}\.?))(?::\d{2,5})?(?:[/?#]\S*)?$/i" r"(https?://(www\.)?[-a-zA-Z0-9@:%._\+~#=]{2,256}\.[a-z]{2,6}\b([-a-zA-Z0-9@:%_\+.~#?&//=]*)*.(\.ogg))" ).unwrap(); } ISO8601_DATE_REGEX.replace_all(before, "FUCK YA") } /// We need to return different futures depending on the route matched, /// and we can do that with an enum, such as `futures::Either`, or with /// trait objects. /// /// A boxed Future (trait object) is used as it is easier to understand /// and extend with more types. Advanced users could switch to `Either`. type BoxFut = Box<Future<Item = Response<Body>, Error = hyper::Error> + Send>; /// This is our service handler. It receives a Request, routes on its /// path, and returns a Future of a Response. //fn echo(req: Request<Body>, buf: Vec<u8>) -> BoxFut { fn echo(req: Request<Body>) -> BoxFut { let mut response = Response::new(Body::empty()); debug!("method: {}, uri: {}", req.method(), req.uri()); match req.method() { &Method::GET => { if req.uri().path().starts_with("/fwd/") { let req_uri = reduce_forwarded_uri(req.uri()); //let forwarded_uri = Uri::from_static(&req_uri); *response.body_mut() = Body::from("Lets forward: ".to_owned() + &req_uri); let body = reqwest::get(req_uri.as_str()) //.unwrap(); //.danger_disable_certs_verification() .expect(&format!("cannot get '{}'", &req_uri)) .text() //.unwrap(); .expect(&format!("cannot get text for '{}'", &req_uri)); /*let body = reqwest::Client::builder() .danger_accept_invalid_hostnames(true) .danger_accept_invalid_certs(true) .build() .unwrap() .get("https://www.google.de/") .send() .unwrap() .text() .unwrap();*/ println!("body = {}", body.lines().take(3).collect::<String>()); /*let re_weburl = Regex::new( r"/^(?:(?:(?:https?|ftp):)?\/\/)(?:\S+(?::\S*)?@)?(?:(?!(?:10|127)(?:\.\d{1,3}){3})(?!(?:169\.254|192\.168)(?:\.\d{1,3}){2})(?!172\.(?:1[6-9]|2\d|3[0-1])(?:\.\d{1,3}){2})(?:[1-9]\d?|1\d\d|2[01]\d|22[0-3])(?:\.(?:1?\d{1,2}|2[0-4]\d|25[0-5])){2}(?:\.(?:[1-9]\d?|1\d\d|2[0-4]\d|25[0-4]))|(?:(?:[a-z0-9\u00a1-\uffff][a-z0-9\u00a1-\uffff_-]{0,62})?[a-z0-9\u00a1-\uffff]\.)+(?:[a-z\u00a1-\uffff]{2,}\.?))(?::\d{2,5})?(?:[/?#]\S*)?$/i" );*/ // check if there is an alternative to the ogg-vorbis stream // when true, then prioritize the mp3 over it // else create a reference to the mp3 forwarding endpoint // SSL Certificates on the host are important. make shure: // ENV SSL_CERT_FILE=/etc/ssl/certs/ca-certificates.crt // ENV SSL_CERT_DIR=/etc/ssl/certs // are set. let after = reformat_dates(&body); //println!("body = {}", after); //let chunk = Chunk::from(after); //*response.body_mut() = Body::from(after.to_string()); *response.body_mut() = Body::from(after.to_string()); //*response.body_mut() = Body::from("got regex"); return Box::new(future::ok(response)); } } _ => {} } match (req.method(), req.uri().path()) { // Serve some instructions at / (&Method::GET, "/") => { //command_name = 'ffmpeg', //command_opts = ['-i', 'pipe:0', '-f', 'mp3', '-acodec', 'libvorbis', '-ab', '128k', '-aq', '60', '-f', 'ogg', '-']; /* let command_name = "ffmpeg"; //let command_opts = ["-i", "pipe:0", "-f", "mp3", "-acodec", "libvorbis", "-ab", "128k", "-aq", "60", "-f", "ogg", "-"]; //"D:\Program Files\ffmpeg\bin\ffmpeg" -re -i "https://cdn.netzpolitik.org/wp-upload/2019/02/NPP169-Worum-geht-es-eigentlich-bei-der-ePrivacy-Reform.ogg" // -acodec libmp3lame -ab 128k -aq 6
reduce_forwarded_uri
identifier_name
main.rs
ut { fn echo(req: Request<Body>) -> BoxFut { let mut response = Response::new(Body::empty()); debug!("method: {}, uri: {}", req.method(), req.uri()); match req.method() { &Method::GET => { if req.uri().path().starts_with("/fwd/") { let req_uri = reduce_forwarded_uri(req.uri()); //let forwarded_uri = Uri::from_static(&req_uri); *response.body_mut() = Body::from("Lets forward: ".to_owned() + &req_uri); let body = reqwest::get(req_uri.as_str()) //.unwrap(); //.danger_disable_certs_verification() .expect(&format!("cannot get '{}'", &req_uri)) .text() //.unwrap(); .expect(&format!("cannot get text for '{}'", &req_uri)); /*let body = reqwest::Client::builder() .danger_accept_invalid_hostnames(true) .danger_accept_invalid_certs(true) .build() .unwrap() .get("https://www.google.de/") .send() .unwrap() .text() .unwrap();*/ println!("body = {}", body.lines().take(3).collect::<String>()); /*let re_weburl = Regex::new( r"/^(?:(?:(?:https?|ftp):)?\/\/)(?:\S+(?::\S*)?@)?(?:(?!(?:10|127)(?:\.\d{1,3}){3})(?!(?:169\.254|192\.168)(?:\.\d{1,3}){2})(?!172\.(?:1[6-9]|2\d|3[0-1])(?:\.\d{1,3}){2})(?:[1-9]\d?|1\d\d|2[01]\d|22[0-3])(?:\.(?:1?\d{1,2}|2[0-4]\d|25[0-5])){2}(?:\.(?:[1-9]\d?|1\d\d|2[0-4]\d|25[0-4]))|(?:(?:[a-z0-9\u00a1-\uffff][a-z0-9\u00a1-\uffff_-]{0,62})?[a-z0-9\u00a1-\uffff]\.)+(?:[a-z\u00a1-\uffff]{2,}\.?))(?::\d{2,5})?(?:[/?#]\S*)?$/i" );*/ // check if there is an alternative to the ogg-vorbis stream // when true, then prioritize the mp3 over it // else create a reference to the mp3 forwarding endpoint // SSL Certificates on the host are important. make shure: // ENV SSL_CERT_FILE=/etc/ssl/certs/ca-certificates.crt // ENV SSL_CERT_DIR=/etc/ssl/certs // are set. let after = reformat_dates(&body); //println!("body = {}", after); //let chunk = Chunk::from(after); //*response.body_mut() = Body::from(after.to_string()); *response.body_mut() = Body::from(after.to_string()); //*response.body_mut() = Body::from("got regex"); return Box::new(future::ok(response)); } } _ => {} } match (req.method(), req.uri().path()) { // Serve some instructions at / (&Method::GET, "/") => { //command_name = 'ffmpeg', //command_opts = ['-i', 'pipe:0', '-f', 'mp3', '-acodec', 'libvorbis', '-ab', '128k', '-aq', '60', '-f', 'ogg', '-']; /* let command_name = "ffmpeg"; //let command_opts = ["-i", "pipe:0", "-f", "mp3", "-acodec", "libvorbis", "-ab", "128k", "-aq", "60", "-f", "ogg", "-"]; //"D:\Program Files\ffmpeg\bin\ffmpeg" -re -i "https://cdn.netzpolitik.org/wp-upload/2019/02/NPP169-Worum-geht-es-eigentlich-bei-der-ePrivacy-Reform.ogg" // -acodec libmp3lame -ab 128k -aq 60 -f mp3 - > bla.mp3 //let media_addr = "https://cdn.netzpolitik.org/wp-upload/2019/02/NPP169-Worum-geht-es-eigentlich-bei-der-ePrivacy-Reform.ogg"; let media_addr = "https://upload.wikimedia.org/wikipedia/commons/f/f2/Median_test.ogg"; let command_opts = ["-i", media_addr, "-acodec", "libmp3lame", "-ab", "128k", "-aq", "60", "-f", "mp3", "-"]; let mut ffmpeg_path = command_name; if cfg!(target_os = "windows") { ffmpeg_path = "D:/Program Files/ffmpeg/bin/ffmpeg.exe"; } // Spawn the `wc` command let process = match Command::new(ffmpeg_path) .args(&command_opts) .stdin(Stdio::piped()) .stdout(Stdio::piped()) .spawn() { Err(why) => panic!("couldn't spawn {}: {}", command_name, why.description()), Ok(process) => process, }; // The `stdout` field also has type `Option<ChildStdout>` so must be unwrapped. let mut buffer: Vec<u8> = Vec::new(); match process.stdout.unwrap().read_to_end(&mut buffer) { Err(why) => panic!("couldn't read {} stdout: {}", command_name, why.description()), Ok(_) => println!("buffer size:[{}]", buffer.len()), } *response.body_mut() = Body::from(buffer); return Box::new( future::ok(response)); */ /*let mapping = || -> Vec(u8) { }*/ //let chunks = vec!["hello", " ", "world"]; //let stream = futures::stream::iter_ok::<_, ::std::io::Error>(chunks); /*let mapping = req .into_body() .map(|chunk| { chunk.iter() .map(|byte| byte.to_ascii_uppercase()) .collect::<Vec<u8>>() });*/ let mapping1 = req.into_body().map(|chunk| { chunk .iter() .map(|byte| { println!("chunk {}", byte.to_ascii_uppercase()); byte.to_ascii_uppercase() }) .collect::<Vec<u8>>() }); let data_fuck = vec!["FUCK", " ", "YOU!"]; let chunk_fuck = Chunk::from("fuck"); let stream_fuck = futures::stream::iter_ok::<_, ::std::io::Error>(data_fuck); /* //let data2 = vec!["hello", " ", "world"]; let data2: Vec<u8> = vec![0x55, 0x20, 0x66]; //let chunk2 = Chunk::from(data2); //let conv = |x: Vec<u8>| x.iter(); let stream2 = futures::stream::iter_ok::<_, ::std::io::Error>(data2); //let stream2 = futures::stream::iter_ok::<_, ::std::io::Error>(data2); let chunks = fileio::load_local_mp3_buffer(); let c: &[u8] = &chunks; // c: &[u8] //let chunk = Chunk::from(c); let stream = futures::stream::iter_ok::<_, ::std::io::Error>(c); *response.body_mut() = Body::from(chunks); return Box::new( future::ok(response)); */ // type BoxFut = Box<Future<Item = Response<Body>, Error = hyper::Error> + Send>; //let bbb = Box::new(future::ok(Response::new("Fuck YOU"))); //let xxx: BoxFut = Box::new(future::ok(response)); //xxx let my_stream = MyStream::new(5); //let xstream = futures::stream::iter_ok::<_, ::std::io::Error>(my_stream.iter()); //let mut file = &get_local_mp3_path(); let mut filepath = "/media/hdd/jedzia/rust/p.mp3"; if cfg!(target_os = "windows") { filepath = "p.mp3"; } //let file = File::open(filepath).map(file_response).or_else(|_| status_response(StatusCode::NOT_FOUND)); //.expect("failed to open file") //let file = tokio::fs::File::open(filepath).catch_unwind(); //let fstream = FramedRead::new(file, ChunkDecoder); /*fn decode(buf: Vec<u8>) -> Result<Option<Chunk>, io::Error> { let len = buf.len();
random_line_split
mod.rs
::Always), _ => Err("invalid cache policy"), } } } impl Default for CachePolicy { fn default() -> Self { CachePolicy::Auto } } /// Options that configure the behavior of the passthrough fuse file system. #[derive(Debug, Clone, PartialEq)] pub struct Config { /// How long the FUSE client should consider directory entries to be valid. If the contents of a /// directory can only be modified by the FUSE client (i.e., the file system has exclusive /// access), then this should be a large value. /// /// The default value for this option is 5 seconds. pub entry_timeout: Duration, /// How long the FUSE client should consider file and directory attributes to be valid. If the /// attributes of a file or directory can only be modified by the FUSE client (i.e., the file /// system has exclusive access), then this should be set to a large value. /// /// The default value for this option is 5 seconds. pub attr_timeout: Duration, /// The caching policy the file system should use. See the documentation of `CachePolicy` for /// more details. pub cache_policy: CachePolicy, /// Whether the file system should enabled writeback caching. This can improve performance as it /// allows the FUSE client to cache and coalesce multiple writes before sending them to the file /// system. However, enabling this option can increase the risk of data corruption if the file /// contents can change without the knowledge of the FUSE client (i.e., the server does **NOT** /// have exclusive access). Additionally, the file system should have read access to all files /// in the directory it is serving as the FUSE client may send read requests even for files /// opened with `O_WRONLY`. /// /// Therefore callers should only enable this option when they can guarantee that: 1) the file /// system has exclusive access to the directory and 2) the file system has read permissions for /// all files in that directory. /// /// The default value for this option is `false`. pub writeback: bool, /// The path of the root directory. /// /// The default is `/`. pub root_dir: String, /// Whether the file system should support Extended Attributes (xattr). Enabling this feature may /// have a significant impact on performance, especially on write parallelism. This is the result /// of FUSE attempting to remove the special file privileges after each write request. /// /// The default value for this options is `false`. pub xattr: bool, /// To be compatible with Vfs and PseudoFs, PassthroughFs needs to prepare /// root inode before accepting INIT request. /// /// The default value for this option is `true`. pub do_import: bool, /// Control whether no_open is allowed. /// /// The default value for this option is `false`. pub no_open: bool, /// Control whether no_opendir is allowed. /// /// The default value for this option is `false`. pub no_opendir: bool, } impl Default for Config { fn default() -> Self { Config { entry_timeout: Duration::from_secs(5), attr_timeout: Duration::from_secs(5), cache_policy: Default::default(), writeback: false, root_dir: String::from("/"), xattr: false, do_import: true, no_open: false, no_opendir: false, } } } /// A file system that simply "passes through" all requests it receives to the underlying file /// system. /// /// To keep the implementation simple it servers the contents of its root directory. Users /// that wish to serve only a specific directory should set up the environment so that that /// directory ends up as the root of the file system process. One way to accomplish this is via a /// combination of mount namespaces and the pivot_root system call. pub struct PassthroughFs<D> { // File descriptors for various points in the file system tree. These fds are always opened with // the `O_PATH` option so they cannot be used for reading or writing any data. See the // documentation of the `O_PATH` flag in `open(2)` for more details on what one can and cannot // do with an fd opened with this flag. inode_map: InodeMap, next_inode: AtomicU64, // File descriptors for open files and directories. Unlike the fds in `inodes`, these _can_ be // used for reading and writing data. handle_map: HandleMap, next_handle: AtomicU64, // File descriptor pointing to the `/proc` directory. This is used to convert an fd from // `inodes` into one that can go into `handles`. This is accomplished by reading the // `self/fd/{}` symlink. We keep an open fd here in case the file system tree that we are meant // to be serving doesn't have access to `/proc`. proc: File, // Whether writeback caching is enabled for this directory. This will only be true when // `cfg.writeback` is true and `init` was called with `FsOptions::WRITEBACK_CACHE`. writeback: AtomicBool, // Whether no_open is enabled. no_open: AtomicBool, // Whether no_opendir is enabled. no_opendir: AtomicBool, cfg: Config, phantom: PhantomData<D>, } impl<D: AsyncDrive> PassthroughFs<D> { /// Create a Passthrough file system instance. pub fn new(cfg: Config) -> io::Result<PassthroughFs<D>> { // Safe because this is a constant value and a valid C string. let proc_cstr = unsafe { CStr::from_bytes_with_nul_unchecked(PROC_CSTR) }; let proc = Self::open_file( libc::AT_FDCWD, proc_cstr, libc::O_PATH | libc::O_NOFOLLOW | libc::O_CLOEXEC, 0, )?; Ok(PassthroughFs { inode_map: InodeMap::new(), next_inode: AtomicU64::new(fuse::ROOT_ID + 1), handle_map: HandleMap::new(), next_handle: AtomicU64::new(1), proc, writeback: AtomicBool::new(false), no_open: AtomicBool::new(false), no_opendir: AtomicBool::new(false), cfg, phantom: PhantomData, }) } /// Initialize the Passthrough file system. pub fn import(&self) -> io::Result<()> { let root = CString::new(self.cfg.root_dir.as_str()).expect("CString::new failed"); // We use `O_PATH` because we just want this for traversing the directory tree // and not for actually reading the contents. let f = Self::open_file( libc::AT_FDCWD, &root, libc::O_PATH | libc::O_NOFOLLOW | libc::O_CLOEXEC, 0, )?; let st = Self::stat(&f)?; // Safe because this doesn't modify any memory and there is no need to check the return // value because this system call always succeeds. We need to clear the umask here because // we want the client to be able to set all the bits in the mode. unsafe { libc::umask(0o000) }; // Not sure why the root inode gets a refcount of 2 but that's what libfuse does. self.inode_map.insert( fuse::ROOT_ID, InodeAltKey::from_stat(&st), InodeData::new(fuse::ROOT_ID, f, 2), ); Ok(()) } /// Get the list of file descriptors which should be reserved across live upgrade. pub fn keep_fds(&self) -> Vec<RawFd> { vec![self.proc.as_raw_fd()] } fn stat(f: &File) -> io::Result<libc::stat64> { // Safe because this is a constant value and a valid C string. let pathname = unsafe { CStr::from_bytes_with_nul_unchecked(EMPTY_CSTR) }; let mut st = MaybeUninit::<libc::stat64>::zeroed(); // Safe because the kernel will only write data in `st` and we check the return value. let res = unsafe { libc::fstatat64( f.as_raw_fd(), pathname.as_ptr(), st.as_mut_ptr(), libc::AT_EMPTY_PATH | libc::AT_SYMLINK_NOFOLLOW, ) }; if res >= 0 { // Safe because the kernel guarantees that the struct is now fully initialized. Ok(unsafe { st.assume_init() }) } else { Err(io::Error::last_os_error()) } } fn open_file(dfd: i32, pathname: &CStr, flags: i32, mode: u32) -> io::Result<File> { let fd = if flags & libc::O_CREAT == libc::O_CREAT { unsafe { libc::openat(dfd, pathname.as_ptr(), flags, mode) } } else {
unsafe { libc::openat(dfd, pathname.as_ptr(), flags) } };
conditional_block
mod.rs
use std::mem::MaybeUninit; use std::os::unix::io::{AsRawFd, FromRawFd, RawFd}; use std::str::FromStr; use std::sync::atomic::{AtomicBool, AtomicU64, Ordering}; use std::sync::{Arc, Mutex, MutexGuard, RwLock, RwLockWriteGuard}; use std::time::Duration; use vm_memory::ByteValued; use crate::abi::linux_abi as fuse; use crate::api::filesystem::Entry; use crate::api::{BackendFileSystem, VFS_MAX_INO}; #[cfg(feature = "async-io")] mod async_io; mod sync_io; mod multikey; use multikey::MultikeyBTreeMap; use crate::async_util::AsyncDrive; const CURRENT_DIR_CSTR: &[u8] = b".\0"; const PARENT_DIR_CSTR: &[u8] = b"..\0"; const EMPTY_CSTR: &[u8] = b"\0"; const PROC_CSTR: &[u8] = b"/proc\0"; type Inode = u64; type Handle = u64; #[derive(Clone, Copy, Debug, PartialOrd, Ord, PartialEq, Eq)] struct InodeAltKey { ino: libc::ino64_t, dev: libc::dev_t, } impl InodeAltKey { fn from_stat(st: &libc::stat64) -> Self { InodeAltKey { ino: st.st_ino, dev: st.st_dev, } } } struct InodeData { inode: Inode, // Most of these aren't actually files but ¯\_(ツ)_/¯. file: File, refcount: AtomicU64, } impl InodeData { fn new(inode: Inode, file: File, refcount: u64) -> Self { InodeData { inode, file, refcount: AtomicU64::new(refcount), } } // When making use of the underlying RawFd, the caller must ensure that the Arc<InodeData> // object is within scope. Otherwise it may cause race window to access wrong target fd. // By introducing this method, we could explicitly audit all callers making use of the // underlying RawFd. fn get_raw_fd(&self) -> RawFd { self.file.as_raw_fd() } } /// Data structures to manage accessed inodes. struct InodeMap { inodes: RwLock<MultikeyBTreeMap<Inode, InodeAltKey, Arc<InodeData>>>, } impl InodeMap { fn new() -> Self { InodeMap { inodes: RwLock::new(MultikeyBTreeMap::new()), } } fn clear(&self) { self.inodes.write().unwrap().clear(); } fn get(&self, inode: Inode) -> io::Result<Arc<InodeData>> { self.inodes .read() .unwrap() .get(&inode) .map(Arc::clone) .ok_or_else(ebadf) } fn get_alt(&self, altkey: &InodeAltKey) -> Option<Arc<InodeData>> { self.inodes.read().unwrap().get_alt(altkey).map(Arc::clone) } fn get_map_mut( &self, ) -> RwLockWriteGuard<MultikeyBTreeMap<Inode, InodeAltKey, Arc<InodeData>>> { self.inodes.write().unwrap() } fn insert(&self, inode: Inode, altkey: InodeAltKey, data: InodeData) { self.inodes .write() .unwrap() .insert(inode, altkey, Arc::new(data)); } } struct HandleData { inode: Inode, file: File, lock: Mutex<()>, } impl HandleData { fn new(inode: Inode, file: File) -> Self { HandleData { inode, file, lock: Mutex::new(()), } } fn get_file_mut(&self) -> (MutexGuard<()>, &File) { (self.lock.lock().unwrap(), &self.file) } // When making use of the underlying RawFd, the caller must ensure that the Arc<HandleData> // object is within scope. Otherwise it may cause race window to access wrong target fd. // By introducing this method, we could explicitly audit all callers making use of the // underlying RawFd. fn get_handle_raw_fd(&self) -> RawFd { self.file.as_raw_fd() } } struct HandleMap { handles: RwLock<BTreeMap<Handle, Arc<HandleData>>>, } impl HandleMap { fn new() -> Self { HandleMap { handles: RwLock::new(BTreeMap::new()), } } fn clear(&self) { self.handles.write().unwrap().clear(); } fn insert(&self, handle: Handle, data: HandleData) { self.handles.write().unwrap().insert(handle, Arc::new(data)); } fn release(&self, handle: Handle, inode: Inode) -> io::Result<()> { let mut handles = self.handles.write().unwrap(); if let btree_map::Entry::Occupied(e) = handles.entry(handle) { if e.get().inode == inode { // We don't need to close the file here because that will happen automatically when // the last `Arc` is dropped. e.remove(); return Ok(()); } } Err(ebadf()) } fn get(&self, handle: Handle, inode: Inode) -> io::Result<Arc<HandleData>> { self.handles .read() .unwrap() .get(&handle) .filter(|hd| hd.inode == inode) .map(Arc::clone) .ok_or_else(ebadf) } } #[repr(C, packed)] #[derive(Clone, Copy, Debug, Default)] struct LinuxDirent64 { d_ino: libc::ino64_t, d_off: libc::off64_t, d_reclen: libc::c_ushort, d_ty: libc::c_uchar, } unsafe impl ByteValued for LinuxDirent64 {} /// The caching policy that the file system should report to the FUSE client. By default the FUSE /// protocol uses close-to-open consistency. This means that any cached contents of the file are /// invalidated the next time that file is opened. #[derive(Debug, Clone, PartialEq)] pub enum CachePolicy { /// The client should never cache file data and all I/O should be directly forwarded to the /// server. This policy must be selected when file contents may change without the knowledge of /// the FUSE client (i.e., the file system does not have exclusive access to the directory). Never, /// The client is free to choose when and how to cache file data. This is the default policy and /// uses close-to-open consistency as described in the enum documentation. Auto, /// The client should always cache file data. This means that the FUSE client will not /// invalidate any cached data that was returned by the file system the last time the file was /// opened. This policy should only be selected when the file system has exclusive access to the /// directory. Always, } impl FromStr for CachePolicy { type Err = &'static str; fn from_str(s: &str) -> Result<Self, Self::Err> { match s { "never" | "Never" | "NEVER" | "none" | "None" | "NONE" => Ok(CachePolicy::Never), "auto" | "Auto" | "AUTO" => Ok(CachePolicy::Auto), "always" | "Always" | "ALWAYS" => Ok(CachePolicy::Always), _ => Err("invalid cache policy"), } } } impl Default for CachePolicy { fn default() -> Self { CachePolicy::Auto } } /// Options that configure the behavior of the passthrough fuse file system. #[derive(Debug, Clone, PartialEq)] pub struct Config { /// How long the FUSE client should consider directory entries to be valid. If the contents of a /// directory can only be modified by the FUSE client (i.e., the file system has exclusive /// access), then this should be a large value. /// /// The default value for this option is 5 seconds. pub entry_timeout: Duration, /// How long the FUSE client should consider file and directory attributes to be valid. If the /// attributes of a file or directory can only be modified by the FUSE client (i.e., the file /// system has exclusive access), then this should be set to a large value. /// /// The default value for this option is 5 seconds. pub attr_timeout: Duration, /// The caching policy the file system should use. See the documentation of `CachePolicy` for /// more details. pub cache_policy: CachePolicy, /// Whether the file system should enabled writeback caching. This can improve performance as it
use std::ffi::{CStr, CString}; use std::fs::File; use std::io; use std::marker::PhantomData;
random_line_split
mod.rs
fd from // `inodes` into one that can go into `handles`. This is accomplished by reading the // `self/fd/{}` symlink. We keep an open fd here in case the file system tree that we are meant // to be serving doesn't have access to `/proc`. proc: File, // Whether writeback caching is enabled for this directory. This will only be true when // `cfg.writeback` is true and `init` was called with `FsOptions::WRITEBACK_CACHE`. writeback: AtomicBool, // Whether no_open is enabled. no_open: AtomicBool, // Whether no_opendir is enabled. no_opendir: AtomicBool, cfg: Config, phantom: PhantomData<D>, } impl<D: AsyncDrive> PassthroughFs<D> { /// Create a Passthrough file system instance. pub fn new(cfg: Config) -> io::Result<PassthroughFs<D>> { // Safe because this is a constant value and a valid C string. let proc_cstr = unsafe { CStr::from_bytes_with_nul_unchecked(PROC_CSTR) }; let proc = Self::open_file( libc::AT_FDCWD, proc_cstr, libc::O_PATH | libc::O_NOFOLLOW | libc::O_CLOEXEC, 0, )?; Ok(PassthroughFs { inode_map: InodeMap::new(), next_inode: AtomicU64::new(fuse::ROOT_ID + 1), handle_map: HandleMap::new(), next_handle: AtomicU64::new(1), proc, writeback: AtomicBool::new(false), no_open: AtomicBool::new(false), no_opendir: AtomicBool::new(false), cfg, phantom: PhantomData, }) } /// Initialize the Passthrough file system. pub fn import(&self) -> io::Result<()> { let root = CString::new(self.cfg.root_dir.as_str()).expect("CString::new failed"); // We use `O_PATH` because we just want this for traversing the directory tree // and not for actually reading the contents. let f = Self::open_file( libc::AT_FDCWD, &root, libc::O_PATH | libc::O_NOFOLLOW | libc::O_CLOEXEC, 0, )?; let st = Self::stat(&f)?; // Safe because this doesn't modify any memory and there is no need to check the return // value because this system call always succeeds. We need to clear the umask here because // we want the client to be able to set all the bits in the mode. unsafe { libc::umask(0o000) }; // Not sure why the root inode gets a refcount of 2 but that's what libfuse does. self.inode_map.insert( fuse::ROOT_ID, InodeAltKey::from_stat(&st), InodeData::new(fuse::ROOT_ID, f, 2), ); Ok(()) } /// Get the list of file descriptors which should be reserved across live upgrade. pub fn keep_fds(&self) -> Vec<RawFd> { vec![self.proc.as_raw_fd()] } fn stat(f: &File) -> io::Result<libc::stat64> { // Safe because this is a constant value and a valid C string. let pathname = unsafe { CStr::from_bytes_with_nul_unchecked(EMPTY_CSTR) }; let mut st = MaybeUninit::<libc::stat64>::zeroed(); // Safe because the kernel will only write data in `st` and we check the return value. let res = unsafe { libc::fstatat64( f.as_raw_fd(), pathname.as_ptr(), st.as_mut_ptr(), libc::AT_EMPTY_PATH | libc::AT_SYMLINK_NOFOLLOW, ) }; if res >= 0 { // Safe because the kernel guarantees that the struct is now fully initialized. Ok(unsafe { st.assume_init() }) } else { Err(io::Error::last_os_error()) } } fn open_file(dfd: i32, pathname: &CStr, flags: i32, mode: u32) -> io::Result<File> { let fd = if flags & libc::O_CREAT == libc::O_CREAT { unsafe { libc::openat(dfd, pathname.as_ptr(), flags, mode) } } else { unsafe { libc::openat(dfd, pathname.as_ptr(), flags) } }; if fd < 0 { return Err(io::Error::last_os_error()); } // Safe because we just opened this fd. Ok(unsafe { File::from_raw_fd(fd) }) } fn do_lookup(&self, parent: Inode, name: &CStr) -> io::Result<Entry> { let p = self.inode_map.get(parent)?; let f = Self::open_file( p.get_raw_fd(), name, libc::O_PATH | libc::O_NOFOLLOW | libc::O_CLOEXEC, 0, )?; let st = Self::stat(&f)?; let altkey = InodeAltKey::from_stat(&st); let mut found = None; 'search: loop { match self.inode_map.get_alt(&altkey) { // No existing entry found None => break 'search, Some(data) => { let curr = data.refcount.load(Ordering::Acquire); // forgot_one() has just destroyed the entry, retry... if curr == 0 { continue 'search; } // Saturating add to avoid integer overflow, it's not realistic to saturate u64. let new = curr.saturating_add(1); // Synchronizes with the forgot_one() if data .refcount .compare_exchange(curr, new, Ordering::AcqRel, Ordering::Acquire) .is_ok() { found = Some(data.inode); break; } } } } let inode = if let Some(v) = found { v } else { let mut inodes = self.inode_map.get_map_mut(); // Lookup inode_map again after acquiring the inode_map lock, as there might be another // racing thread already added an inode with the same altkey while we're not holding // the lock. If so just use the newly added inode, otherwise the inode will be replaced // and results in EBADF. match inodes.get_alt(&altkey).map(Arc::clone) { Some(data) => { trace!( "fuse: do_lookup sees existing inode {} altkey {:?}", data.inode, altkey ); data.refcount.fetch_add(1, Ordering::Relaxed); data.inode } None => { let inode = self.next_inode.fetch_add(1, Ordering::Relaxed); if inode > VFS_MAX_INO { return Err(io::Error::new( io::ErrorKind::Other, format!("max inode number reached: {}", VFS_MAX_INO), )); } trace!( "fuse: do_lookup adds new inode {} altkey {:?}", inode, altkey ); inodes.insert(inode, altkey, Arc::new(InodeData::new(inode, f, 1))); inode } } }; Ok(Entry { inode, generation: 0, attr: st, attr_timeout: self.cfg.attr_timeout, entry_timeout: self.cfg.entry_timeout, }) } fn forget_one( inodes: &mut MultikeyBTreeMap<Inode, InodeAltKey, Arc<InodeData>>, inode: Inode, count: u64, ) { // ROOT_ID should not be forgotten, or we're not able to access to files any more. if inode == fuse::ROOT_ID { return; } if let Some(data) = inodes.get(&inode) { // Acquiring the write lock on the inode map prevents new lookups from incrementing the // refcount but there is the possibility that a previous lookup already acquired a // reference to the inode data and is in the process of updating the refcount so we need // to loop here until we can decrement successfully. loop { let curr = data.refcount.load(Ordering::Acquire); // Saturating sub because it doesn't make sense for a refcount to go below zero and // we don't want misbehaving clients to cause integer overflow. let new = curr.saturating_sub(count); trace!( "fuse: forget inode {} refcount {}, count {}, new_count {}", inode, curr, count, new ); // Synchronizes with the acquire load in `do_lookup`. if data .refcount .compare_exchange(curr, new, Ordering::AcqRel, Ordering::Acquire) .is_ok() { if new == 0 { // We just removed the last refcount for this inode. inodes.remove(&inode); } break; } } } } fn do_r
elease(&se
identifier_name
mod.rs
keyBTreeMap<Inode, InodeAltKey, Arc<InodeData>>> { self.inodes.write().unwrap() } fn insert(&self, inode: Inode, altkey: InodeAltKey, data: InodeData) { self.inodes .write() .unwrap() .insert(inode, altkey, Arc::new(data)); } } struct HandleData { inode: Inode, file: File, lock: Mutex<()>, } impl HandleData { fn new(inode: Inode, file: File) -> Self { HandleData { inode, file, lock: Mutex::new(()), } } fn get_file_mut(&self) -> (MutexGuard<()>, &File) { (self.lock.lock().unwrap(), &self.file) } // When making use of the underlying RawFd, the caller must ensure that the Arc<HandleData> // object is within scope. Otherwise it may cause race window to access wrong target fd. // By introducing this method, we could explicitly audit all callers making use of the // underlying RawFd. fn get_handle_raw_fd(&self) -> RawFd { self.file.as_raw_fd() } } struct HandleMap { handles: RwLock<BTreeMap<Handle, Arc<HandleData>>>, } impl HandleMap { fn new() -> Self { HandleMap { handles: RwLock::new(BTreeMap::new()), } } fn clear(&self) { self.handles.write().unwrap().clear(); } fn insert(&self, handle: Handle, data: HandleData) { self.handles.write().unwrap().insert(handle, Arc::new(data)); } fn release(&self, handle: Handle, inode: Inode) -> io::Result<()> { let mut handles = self.handles.write().unwrap(); if let btree_map::Entry::Occupied(e) = handles.entry(handle) { if e.get().inode == inode { // We don't need to close the file here because that will happen automatically when // the last `Arc` is dropped. e.remove(); return Ok(()); } } Err(ebadf()) } fn get(&self, handle: Handle, inode: Inode) -> io::Result<Arc<HandleData>> { self.handles .read() .unwrap() .get(&handle) .filter(|hd| hd.inode == inode) .map(Arc::clone) .ok_or_else(ebadf) } } #[repr(C, packed)] #[derive(Clone, Copy, Debug, Default)] struct LinuxDirent64 { d_ino: libc::ino64_t, d_off: libc::off64_t, d_reclen: libc::c_ushort, d_ty: libc::c_uchar, } unsafe impl ByteValued for LinuxDirent64 {} /// The caching policy that the file system should report to the FUSE client. By default the FUSE /// protocol uses close-to-open consistency. This means that any cached contents of the file are /// invalidated the next time that file is opened. #[derive(Debug, Clone, PartialEq)] pub enum CachePolicy { /// The client should never cache file data and all I/O should be directly forwarded to the /// server. This policy must be selected when file contents may change without the knowledge of /// the FUSE client (i.e., the file system does not have exclusive access to the directory). Never, /// The client is free to choose when and how to cache file data. This is the default policy and /// uses close-to-open consistency as described in the enum documentation. Auto, /// The client should always cache file data. This means that the FUSE client will not /// invalidate any cached data that was returned by the file system the last time the file was /// opened. This policy should only be selected when the file system has exclusive access to the /// directory. Always, } impl FromStr for CachePolicy { type Err = &'static str; fn from_str(s: &str) -> Result<Self, Self::Err> {
impl Default for CachePolicy { fn default() -> Self { CachePolicy::Auto } } /// Options that configure the behavior of the passthrough fuse file system. #[derive(Debug, Clone, PartialEq)] pub struct Config { /// How long the FUSE client should consider directory entries to be valid. If the contents of a /// directory can only be modified by the FUSE client (i.e., the file system has exclusive /// access), then this should be a large value. /// /// The default value for this option is 5 seconds. pub entry_timeout: Duration, /// How long the FUSE client should consider file and directory attributes to be valid. If the /// attributes of a file or directory can only be modified by the FUSE client (i.e., the file /// system has exclusive access), then this should be set to a large value. /// /// The default value for this option is 5 seconds. pub attr_timeout: Duration, /// The caching policy the file system should use. See the documentation of `CachePolicy` for /// more details. pub cache_policy: CachePolicy, /// Whether the file system should enabled writeback caching. This can improve performance as it /// allows the FUSE client to cache and coalesce multiple writes before sending them to the file /// system. However, enabling this option can increase the risk of data corruption if the file /// contents can change without the knowledge of the FUSE client (i.e., the server does **NOT** /// have exclusive access). Additionally, the file system should have read access to all files /// in the directory it is serving as the FUSE client may send read requests even for files /// opened with `O_WRONLY`. /// /// Therefore callers should only enable this option when they can guarantee that: 1) the file /// system has exclusive access to the directory and 2) the file system has read permissions for /// all files in that directory. /// /// The default value for this option is `false`. pub writeback: bool, /// The path of the root directory. /// /// The default is `/`. pub root_dir: String, /// Whether the file system should support Extended Attributes (xattr). Enabling this feature may /// have a significant impact on performance, especially on write parallelism. This is the result /// of FUSE attempting to remove the special file privileges after each write request. /// /// The default value for this options is `false`. pub xattr: bool, /// To be compatible with Vfs and PseudoFs, PassthroughFs needs to prepare /// root inode before accepting INIT request. /// /// The default value for this option is `true`. pub do_import: bool, /// Control whether no_open is allowed. /// /// The default value for this option is `false`. pub no_open: bool, /// Control whether no_opendir is allowed. /// /// The default value for this option is `false`. pub no_opendir: bool, } impl Default for Config { fn default() -> Self { Config { entry_timeout: Duration::from_secs(5), attr_timeout: Duration::from_secs(5), cache_policy: Default::default(), writeback: false, root_dir: String::from("/"), xattr: false, do_import: true, no_open: false, no_opendir: false, } } } /// A file system that simply "passes through" all requests it receives to the underlying file /// system. /// /// To keep the implementation simple it servers the contents of its root directory. Users /// that wish to serve only a specific directory should set up the environment so that that /// directory ends up as the root of the file system process. One way to accomplish this is via a /// combination of mount namespaces and the pivot_root system call. pub struct PassthroughFs<D> { // File descriptors for various points in the file system tree. These fds are always opened with // the `O_PATH` option so they cannot be used for reading or writing any data. See the // documentation of the `O_PATH` flag in `open(2)` for more details on what one can and cannot // do with an fd opened with this flag. inode_map: InodeMap, next_inode: AtomicU64, // File descriptors for open files and directories. Unlike the fds in `inodes`, these _can_ be // used for reading and writing data. handle_map: HandleMap, next_handle: AtomicU64, // File descriptor pointing to the `/proc` directory. This is used to convert an fd from // `inodes` into one that can go into `handles`. This is accomplished by reading
match s { "never" | "Never" | "NEVER" | "none" | "None" | "NONE" => Ok(CachePolicy::Never), "auto" | "Auto" | "AUTO" => Ok(CachePolicy::Auto), "always" | "Always" | "ALWAYS" => Ok(CachePolicy::Always), _ => Err("invalid cache policy"), } } }
identifier_body
CollectGame.js
cookie('collect_high_score') != '') { this.high_score = $.cookie('collect_high_score'); } }, create: function() { console.log('Collect game'); this.background_music = this.game.add.audio('boss_background_music'); this.right_answer_sound = this.game.add.audio('right_answer_sound'); this.wrong_answer_sound = this.game.add.audio('wrong_answer_sound'); this.win_sound = this.game.add.audio('win_sound'); // Manage Layers this.background_layer = this.game.add.group(); this.background_layer.z = 0; // move button group this.move_button_group = this.game.add.group(); this.move_button_group.z = 1; this.up_button = this.game.add.sprite(50, 175, 'arrowUp'); this.up_button.inputEnabled = true; this.down_button = this.game.add.sprite(50, 425, 'arrowDown'); this.down_button.inputEnabled = true; this.game.input.onDown.add(this.activeInput, this); this.game.input.onUp.add(this.releaseInput, this); // Pause button / ui group this.ui_layer = this.game.add.group(); this.ui_layer.z = 2; // Create rock group this.boulders = this.game.add.group(); // Score display this.score_text = this.game.add.text(this.game.width - 145, 9, this.score + '', this.text_style); this.score_text.anchor.setTo(1, 0); //this.score_text.fixedToCamera = true; this.ui_layer.add(this.score_text); // Create zcar this.zcar = this.game.add.sprite(337, 443, 'zcar'); this.zcar.anchor.setTo(.5, 1); // Instructions this.instructions = this.game.add.sprite(0, 0, 'collect_instructions'); this.start_button = this.game.add.button(this.game.world.centerX, this.game.world.height - 90, 'yellow_buttons', this.killInstructions, this, 3, 3, 4); this.start_button.alpha = 0; this.start_button.anchor.setTo(0.5, 0.5); this.start_text = this.game.add.text(4, 0, 'START', {font: '30pt kenvector_future', fill: '#000', align: 'center'}); this.start_text.anchor.setTo(0.5, 0.5); this.start_button.addChild(this.start_text); // Initialize scoreboard this.scoreboard = this.game.add.group(); var instruction_audio = this.game.add.audio('collect_instruction_sound'); instruction_audio.onStop.add(function(){ this.game.add.tween(this.start_button).to({alpha: 1}, 500, null, true); }, this); instruction_audio.play(); }, activeInput: function() { this.inputActive = true; }, releaseInput: function() { this.inputActive = false; }, killInstructions: function() { this.start_button.destroy(); this.instructions.destroy(); this.startLevel(); }, update: function() { if (!this.started) { return; } if (this.inputActive) { if (this.up_button.input.pointerOver()) { //this.zcar.body.velocity.y = 200; this.zcar.y -= 5; if (this.zcar.y <= 200) { this.zcar.y = 200; } } if (this.down_button.input.pointerOver()) { //this.zcar.body.velocity.y = 200; this.zcar.y += 5; if (this.zcar.y >= 650) { this.zcar.y = 650; } } } this.gametime = this.gametime+1; if (this.gametime >= 150) { this.gametime = this.gametime-150; this.generateBoulder(); } this.game.physics.overlap(this.zcar, this.boulders, this.touchBoulder, null, this); if (this.numCollected == 3) { this.winLevel(); } }, generateBoulder: function() { boulderPos = this.game.rnd.integerInRange(1, 6) while (boulderPos == this.lastBoulderPosition) { boulderPos = this.game.rnd.integerInRange(1, 6) } this.lastBoulderPosition = boulderPos; boulder = this.boulders.create(1220, 100*boulderPos, 'boulder'); boulder.body.gravity.x = -80; boulder.anchor.setTo(0,0); // 1 in 4 change of generating the right answer boulderrnd = this.game.rnd.integerInRange(1, 4); console.log('boulderrnd' + boulderrnd); if (boulderrnd == 2) { boulderVal = this.answer; } else { boulderVal = this.game.rnd.integerInRange(1, 20); } while (boulderVal == this.lastBoulderValue) { boulderVal = this.game.rnd.integerInRange(1, 20); } this.lastBoulderValue = boulderVal; boulder_text = this.game.add.text(15, 15, boulderVal, {font: '40px kenvector_future', fill: '#fff', align: 'center'}); boulder.value = boulderVal; boulder_text.anchor.setTo(0, 0); boulder.addChild(boulder_text); /*theB = this.boulders.create(1000, 800, 'boulder'); rock = this.rocks.create(this.game.rnd.integerInRange(1, 4)*280 +82, 1, 'rock'); theB.body.gravity.x = -40; theB.anchor.setTo(0,0);*/ //boulder.value = this.boulderVal; //console.log('VALUE: ' + boulder.value); //console.log('boulder'); //boulder_text = this.game.add.text(15, 15, boulder.value, {font: '40px kenvector_future', fill: '#fff', align: 'center'}); }, /*reset: function() { this.zcar.x = 337; this.started = false; this.boulders.destroy(); this.boulders = this.game.add.group(); },*/ startLevel: function() { this.scoreboard.destroy(); /*if (this.game.global_vars.load_saved_state) { this.current_level = this.game.global_vars.saved_level; this.game.global_vars.load_saved_state = false; }*/ // Load level specific things this.background = this.game.add.sprite(0, 0, this.level_images[this.current_level]['background']); this.background_layer.add(this.background); this.difficulty_text = this.game.add.text(80, 10, 'Difficulty ' + this.game.global_vars.diff_score, {font: '20px kenvector_future', fill: '#fff'}); //this.started = false; //this.boulders = this.game.add.group(); //this.startRace(); this.started = true; this.displayNewProblem(); // commented out to test 10/17 //this.background_music.play(0,0,1,true); }, startRace: function() { this.started = true; this.displayNewProblem(); // commented out to test 10/17 //this.background_music.play(0,0,1,true); }, displayNewProblem: function() { var problem = this.game.getMathProblem('addSolution', 'easy'); this.answer = problem.answer; if (this.problem_text == null || !this.problem_text.exists) { this.problem_text = this.game.add.text(640, 100, problem.text, {font: '40px kenvector_future', fill: '#fff'}); this.problem_text.anchor.setTo(0.5, 0.5); } else { this.problem_text.setText(problem.text); } }, moveUp: function() { this.zcar.body.velocity.y = 200; }, moveDown: function() { this.zcar.body.velocity.y = -200; }, checkAnswer: function(answer) { console.log('checking: ' + answer + " and " + this.answer); if (this.answer != answer){ this.wrong_answer_sound.play(); this.game.global_vars.diff_score -= 1; if (this.game.global_vars.diff_score < 0) { this.game.global_vars.diff_score = 0; } this.displayNewProblem(); this.score -= 30; if (this.score < 0)
} if (this.answer == answer) {
{ this.score = 0; }
conditional_block
CollectGame.js
$.cookie('collect_high_score') != '') { this.high_score = $.cookie('collect_high_score'); } }, create: function() { console.log('Collect game'); this.background_music = this.game.add.audio('boss_background_music'); this.right_answer_sound = this.game.add.audio('right_answer_sound'); this.wrong_answer_sound = this.game.add.audio('wrong_answer_sound'); this.win_sound = this.game.add.audio('win_sound'); // Manage Layers this.background_layer = this.game.add.group(); this.background_layer.z = 0; // move button group this.move_button_group = this.game.add.group(); this.move_button_group.z = 1; this.up_button = this.game.add.sprite(50, 175, 'arrowUp'); this.up_button.inputEnabled = true; this.down_button = this.game.add.sprite(50, 425, 'arrowDown'); this.down_button.inputEnabled = true; this.game.input.onDown.add(this.activeInput, this); this.game.input.onUp.add(this.releaseInput, this); // Pause button / ui group this.ui_layer = this.game.add.group(); this.ui_layer.z = 2; // Create rock group this.boulders = this.game.add.group(); // Score display this.score_text = this.game.add.text(this.game.width - 145, 9, this.score + '', this.text_style); this.score_text.anchor.setTo(1, 0); //this.score_text.fixedToCamera = true; this.ui_layer.add(this.score_text); // Create zcar this.zcar = this.game.add.sprite(337, 443, 'zcar'); this.zcar.anchor.setTo(.5, 1); // Instructions this.instructions = this.game.add.sprite(0, 0, 'collect_instructions'); this.start_button = this.game.add.button(this.game.world.centerX, this.game.world.height - 90, 'yellow_buttons', this.killInstructions, this, 3, 3, 4); this.start_button.alpha = 0; this.start_button.anchor.setTo(0.5, 0.5); this.start_text = this.game.add.text(4, 0, 'START', {font: '30pt kenvector_future', fill: '#000', align: 'center'}); this.start_text.anchor.setTo(0.5, 0.5); this.start_button.addChild(this.start_text); // Initialize scoreboard this.scoreboard = this.game.add.group(); var instruction_audio = this.game.add.audio('collect_instruction_sound'); instruction_audio.onStop.add(function(){ this.game.add.tween(this.start_button).to({alpha: 1}, 500, null, true); }, this); instruction_audio.play(); }, activeInput: function() { this.inputActive = true; }, releaseInput: function() { this.inputActive = false; }, killInstructions: function() { this.start_button.destroy(); this.instructions.destroy(); this.startLevel(); }, update: function() { if (!this.started) { return; } if (this.inputActive) { if (this.up_button.input.pointerOver()) { //this.zcar.body.velocity.y = 200; this.zcar.y -= 5; if (this.zcar.y <= 200) { this.zcar.y = 200; } } if (this.down_button.input.pointerOver()) { //this.zcar.body.velocity.y = 200; this.zcar.y += 5; if (this.zcar.y >= 650) { this.zcar.y = 650; } } } this.gametime = this.gametime+1; if (this.gametime >= 150) { this.gametime = this.gametime-150; this.generateBoulder(); } this.game.physics.overlap(this.zcar, this.boulders, this.touchBoulder, null, this); if (this.numCollected == 3) { this.winLevel(); } }, generateBoulder: function() { boulderPos = this.game.rnd.integerInRange(1, 6) while (boulderPos == this.lastBoulderPosition) { boulderPos = this.game.rnd.integerInRange(1, 6) } this.lastBoulderPosition = boulderPos; boulder = this.boulders.create(1220, 100*boulderPos, 'boulder'); boulder.body.gravity.x = -80; boulder.anchor.setTo(0,0); // 1 in 4 change of generating the right answer boulderrnd = this.game.rnd.integerInRange(1, 4); console.log('boulderrnd' + boulderrnd); if (boulderrnd == 2) { boulderVal = this.answer; } else { boulderVal = this.game.rnd.integerInRange(1, 20); } while (boulderVal == this.lastBoulderValue) { boulderVal = this.game.rnd.integerInRange(1, 20); } this.lastBoulderValue = boulderVal; boulder_text = this.game.add.text(15, 15, boulderVal, {font: '40px kenvector_future', fill: '#fff', align: 'center'}); boulder.value = boulderVal; boulder_text.anchor.setTo(0, 0); boulder.addChild(boulder_text); /*theB = this.boulders.create(1000, 800, 'boulder'); rock = this.rocks.create(this.game.rnd.integerInRange(1, 4)*280 +82, 1, 'rock'); theB.body.gravity.x = -40; theB.anchor.setTo(0,0);*/
//boulder.value = this.boulderVal; //console.log('VALUE: ' + boulder.value); //console.log('boulder'); //boulder_text = this.game.add.text(15, 15, boulder.value, {font: '40px kenvector_future', fill: '#fff', align: 'center'}); }, /*reset: function() { this.zcar.x = 337; this.started = false; this.boulders.destroy(); this.boulders = this.game.add.group(); },*/ startLevel: function() { this.scoreboard.destroy(); /*if (this.game.global_vars.load_saved_state) { this.current_level = this.game.global_vars.saved_level; this.game.global_vars.load_saved_state = false; }*/ // Load level specific things this.background = this.game.add.sprite(0, 0, this.level_images[this.current_level]['background']); this.background_layer.add(this.background); this.difficulty_text = this.game.add.text(80, 10, 'Difficulty ' + this.game.global_vars.diff_score, {font: '20px kenvector_future', fill: '#fff'}); //this.started = false; //this.boulders = this.game.add.group(); //this.startRace(); this.started = true; this.displayNewProblem(); // commented out to test 10/17 //this.background_music.play(0,0,1,true); }, startRace: function() { this.started = true; this.displayNewProblem(); // commented out to test 10/17 //this.background_music.play(0,0,1,true); }, displayNewProblem: function() { var problem = this.game.getMathProblem('addSolution', 'easy'); this.answer = problem.answer; if (this.problem_text == null || !this.problem_text.exists) { this.problem_text = this.game.add.text(640, 100, problem.text, {font: '40px kenvector_future', fill: '#fff'}); this.problem_text.anchor.setTo(0.5, 0.5); } else { this.problem_text.setText(problem.text); } }, moveUp: function() { this.zcar.body.velocity.y = 200; }, moveDown: function() { this.zcar.body.velocity.y = -200; }, checkAnswer: function(answer) { console.log('checking: ' + answer + " and " + this.answer); if (this.answer != answer){ this.wrong_answer_sound.play(); this.game.global_vars.diff_score -= 1; if (this.game.global_vars.diff_score < 0) { this.game.global_vars.diff_score = 0; } this.displayNewProblem(); this.score -= 30; if (this.score < 0) { this.score = 0; } } if (this.answer == answer) {
random_line_split
app.js
= { x: 0, y: 0 }; var imageX = 50; var imageY = 50; var tableGlobal; // var numXscale = canvas.clientWidth / canvas.width; // var numYscale = canvas.clientHeight / canvas.height; var imageWidth, imageHeight, imageRight, imageBottom; var draggingImage = false; var startX; var startY; // var img = new Image(); // img.onload = function () { // imageWidth = img.width; // imageHeight = img.height; // imageRight = imageX + imageWidth; // imageBottom = imageY + imageHeight; // draw(true, false); // }; // img.src = "https://dl.dropboxusercontent.com/u/139992952/stackoverflow/facesSmall.png"; var imagDrop = document.createElement("img"); imagDrop.addEventListener("load", function() { //clearCanvas(); imageWidth = imagDrop.width; imageHeight = imagDrop.height; imageRight = imageX + imageWidth; imageBottom = imageY + imageHeight; draw(true, false); console.log("test"); //ctx.drawImage(imagDrop, 0, 0); }, false); // To enable drag and drop canvas.addEventListener("dragover", function(evt) { evt.preventDefault(); }, false); // Handle dropped image file - only Firefox and Google Chrome canvas.addEventListener("drop", function(evt) { var files = evt.dataTransfer.files; if (files.length > 0) { var file = files[0]; if (typeof FileReader !== "undefined" && file.type.indexOf("image") != -1) { var reader = new FileReader(); // Note: addEventListener doesn't work in Google Chrome for this event reader.onload = function(evt) { imagDrop.src = evt.target.result; }; reader.readAsDataURL(file); } } evt.preventDefault(); }, false); updateAxis(); addTable(); function togglePin() { if (flagPin) { document.getElementById('idButtonPin').innerHTML = 'Start pinning'; flagPin = false; } else { document.getElementById('idButtonPin').innerHTML = 'Stop pinning'; flagPin = true; } } function addTable() { var newTable = document.createElement('table'); newTable.setAttribute("style", "float:left;position:relative"); var header = newTable.createTHead(); var row = header.insertRow(0); var cell = row.insertCell(0); cell.innerHTML = "X"; var cell = row.insertCell(1); cell.innerHTML = "Y"; document.getElementById('tables').appendChild(newTable); tableGlobal = newTable; } function canvas2realX(numX) { return numX*canvas.clientWidth / canvas.width; } function canvas2realY(numY) { return numY*canvas.clientHeight / canvas.height; } function drawAxis() { ctx.lineWidth = 3; ctx.strokeStyle = '#333'; ctx.font = 'italic 8pt sans-serif'; ctx.textAlign = "center"; // Draw the axises ctx.beginPath(); ctx.moveTo(xPadding, canvas.height - yPadding); ctx.lineTo(xPadding, 0); ctx.stroke(); ctx.beginPath(); ctx.moveTo(xPadding, canvas.height - yPadding); ctx.lineTo(canvas.width, canvas.height - yPadding); ctx.stroke(); // // Draw the X value texts ctx.lineWidth = 1; for (var i = 0; i <= numMaxX - numMinX; i += numStepX) { ctx.fillText(i, getXPixel(i), canvas.height - yPadding + 20); ctx.beginPath(); ctx.moveTo(getXPixel(i), canvas.height - yPadding); ctx.lineTo(getXPixel(i), 0); ctx.stroke(); } ctx.textAlign = "right"; ctx.textBaseline = "middle"; for (var i = 0; i <= numMaxY - numMinY; i += numStepY) { ctx.fillText(i, xPadding - 10, getYPixel(i)); ctx.beginPath(); ctx.moveTo(xPadding, getYPixel(i)); ctx.lineTo(canvas.width, getYPixel(i)); ctx.stroke(); } ctx.strokeStyle = '#f00'; } function updateAxis() { numMaxX = parseInt(document.getElementById('MaxX').value); numMaxY = parseInt(document.getElementById('MaxY').value); numStepX = parseInt(document.getElementById('StepX').value); numStepY = parseInt(document.getElementById('StepY').value); ctx.clearRect(0, 0, canvas.width, canvas.height); drawAxis(); } function getXPixel(val) { return ((canvas.width - 2 * xPadding) / numMaxX) * val + (xPadding); } // Return the y pixel for a graph point function getYPixel(val) { return canvas.height - (((canvas.height - 2 * yPadding) / numMaxY) * val) - yPadding; } function draw(withAnchors, withBorders) { // clear the canvas ctx.clearRect(0, 0, canvas.width, canvas.height); updateAxis(); // draw the image ctx.globalAlpha = 0.4; ctx.drawImage(imagDrop, 0, 0, imagDrop.width, imagDrop.height, imageX, imageY, imageWidth, imageHeight); // optionally draw the draggable anchors if (withAnchors) { drawDragAnchor(imageX, imageY); drawDragAnchor(imageRight, imageY); drawDragAnchor(imageRight, imageBottom); drawDragAnchor(imageX, imageBottom); } // optionally draw the connecting anchor lines if (withBorders) { ctx.beginPath(); ctx.moveTo(imageX, imageY); ctx.lineTo(imageRight, imageY); ctx.lineTo(imageRight, imageBottom); ctx.lineTo(imageX, imageBottom); ctx.closePath(); ctx.stroke(); } } function drawDragAnchor(x, y) { ctx.beginPath(); ctx.arc(x, y, resizerRadius, 0, pi2, false); ctx.closePath(); ctx.fill(); } function anchorHitTest(x, y) { var dx, dy; // top-left dx = x - canvas2realX(imageX); dy = y - canvas2realY(imageY); if (dx * dx + dy * dy <= rr) { return (0); } // top-right dx = x - canvas2realX(imageRight); dy = y - canvas2realY(imageY); console.log(dx.toString()+' '+dy.toString()); if (dx * dx + dy * dy <= rr) { return (1); } // bottom-right dx = x - canvas2realX(imageRight); dy = y - canvas2realY(imageBottom); if (dx * dx + dy * dy <= rr) { return (2); } // bottom-left dx = x - canvas2realX(imageX); dy = y - canvas2realY(imageBottom); if (dx * dx + dy * dy <= rr) { return (3); } return (-1); } function hitImage(x, y) { return (x > canvas2realX(imageX) && x < canvas2realX(imageX + imageWidth) && y > canvas2realY(imageY) && y < canvas2realY(imageY + imageHeight)); } function handleMouseDown(e) { if (!flagPin) { startX = parseInt(e.clientX - document.getElementById('canvas').offsetLeft); startY = parseInt(e.clientY - document.getElementById('canvas').offsetTop); console.log(startX.toString()+'_'+startY.toString()); draggingResizer = anchorHitTest(startX, startY); draggingImage = draggingResizer < 0 && hitImage(startX, startY); } else { var row = tableGlobal.insertRow(tableGlobal.rows.length); var cell1 = row.insertCell(0); var cell2 = row.insertCell(1); cell2.innerHTML = Math.round((canvas.clientHeight - parseInt(e.clientY - document.getElementById('canvas').offsetTop) - canvas2realY(yPadding)) / ((canvas.clientHeight - 2 * canvas2realY(yPadding)) / numMaxY)); cell1.innerHTML = Math.round((parseInt(e.clientX - document.getElementById('canvas').offsetLeft) - canvas2realX(xPadding)) / ((canvas.clientWidth - 2 * canvas2realX(xPadding)) / numMaxX)); } } function handleMouseUp(e)
function handleMouseOut(e) { if (!flagPin) { handleMouseUp(e); } } function handleMouseMove(e) { if (!flagPin) { if (draggingResizer > -1) { mouseX = parseInt(e.clientX - document.getElementById('canvas').offsetLeft)*canvas.width/canvas.clientWidth; mouseY = parseInt(e.clientY - document.getElementById('canvas').offsetTop)*canvas.height/canvas.clientHeight; console.log(mouseX.toString()+','+mouseY.toString()); // resize the image switch (draggingResizer) { case 0: //top-left imageX = mouseX; imageWidth = imageRight - mouseX; imageY = mouseY; imageHeight =
{ if (!flagPin) { draggingResizer = -1; draggingImage = false; draw(true, false); } }
identifier_body
app.js
= { x: 0, y: 0 }; var imageX = 50; var imageY = 50; var tableGlobal; // var numXscale = canvas.clientWidth / canvas.width; // var numYscale = canvas.clientHeight / canvas.height; var imageWidth, imageHeight, imageRight, imageBottom; var draggingImage = false; var startX; var startY; // var img = new Image(); // img.onload = function () { // imageWidth = img.width; // imageHeight = img.height; // imageRight = imageX + imageWidth; // imageBottom = imageY + imageHeight; // draw(true, false); // }; // img.src = "https://dl.dropboxusercontent.com/u/139992952/stackoverflow/facesSmall.png"; var imagDrop = document.createElement("img"); imagDrop.addEventListener("load", function() { //clearCanvas(); imageWidth = imagDrop.width; imageHeight = imagDrop.height; imageRight = imageX + imageWidth; imageBottom = imageY + imageHeight; draw(true, false); console.log("test"); //ctx.drawImage(imagDrop, 0, 0); }, false); // To enable drag and drop canvas.addEventListener("dragover", function(evt) { evt.preventDefault(); }, false); // Handle dropped image file - only Firefox and Google Chrome canvas.addEventListener("drop", function(evt) { var files = evt.dataTransfer.files; if (files.length > 0) { var file = files[0]; if (typeof FileReader !== "undefined" && file.type.indexOf("image") != -1) { var reader = new FileReader(); // Note: addEventListener doesn't work in Google Chrome for this event reader.onload = function(evt) { imagDrop.src = evt.target.result; }; reader.readAsDataURL(file); } } evt.preventDefault(); }, false); updateAxis(); addTable(); function togglePin() { if (flagPin) { document.getElementById('idButtonPin').innerHTML = 'Start pinning'; flagPin = false; } else { document.getElementById('idButtonPin').innerHTML = 'Stop pinning'; flagPin = true; } } function addTable() { var newTable = document.createElement('table'); newTable.setAttribute("style", "float:left;position:relative"); var header = newTable.createTHead(); var row = header.insertRow(0); var cell = row.insertCell(0); cell.innerHTML = "X"; var cell = row.insertCell(1); cell.innerHTML = "Y"; document.getElementById('tables').appendChild(newTable); tableGlobal = newTable; } function canvas2realX(numX) { return numX*canvas.clientWidth / canvas.width; } function canvas2realY(numY) { return numY*canvas.clientHeight / canvas.height; } function drawAxis() { ctx.lineWidth = 3; ctx.strokeStyle = '#333'; ctx.font = 'italic 8pt sans-serif'; ctx.textAlign = "center"; // Draw the axises ctx.beginPath(); ctx.moveTo(xPadding, canvas.height - yPadding); ctx.lineTo(xPadding, 0); ctx.stroke(); ctx.beginPath(); ctx.moveTo(xPadding, canvas.height - yPadding); ctx.lineTo(canvas.width, canvas.height - yPadding); ctx.stroke(); // // Draw the X value texts ctx.lineWidth = 1; for (var i = 0; i <= numMaxX - numMinX; i += numStepX) { ctx.fillText(i, getXPixel(i), canvas.height - yPadding + 20); ctx.beginPath(); ctx.moveTo(getXPixel(i), canvas.height - yPadding); ctx.lineTo(getXPixel(i), 0); ctx.stroke(); } ctx.textAlign = "right"; ctx.textBaseline = "middle"; for (var i = 0; i <= numMaxY - numMinY; i += numStepY) { ctx.fillText(i, xPadding - 10, getYPixel(i)); ctx.beginPath(); ctx.moveTo(xPadding, getYPixel(i)); ctx.lineTo(canvas.width, getYPixel(i)); ctx.stroke(); } ctx.strokeStyle = '#f00'; } function updateAxis() { numMaxX = parseInt(document.getElementById('MaxX').value); numMaxY = parseInt(document.getElementById('MaxY').value); numStepX = parseInt(document.getElementById('StepX').value); numStepY = parseInt(document.getElementById('StepY').value); ctx.clearRect(0, 0, canvas.width, canvas.height); drawAxis(); } function getXPixel(val) { return ((canvas.width - 2 * xPadding) / numMaxX) * val + (xPadding); } // Return the y pixel for a graph point function getYPixel(val) { return canvas.height - (((canvas.height - 2 * yPadding) / numMaxY) * val) - yPadding; } function draw(withAnchors, withBorders) { // clear the canvas ctx.clearRect(0, 0, canvas.width, canvas.height); updateAxis(); // draw the image ctx.globalAlpha = 0.4; ctx.drawImage(imagDrop, 0, 0, imagDrop.width, imagDrop.height, imageX, imageY, imageWidth, imageHeight); // optionally draw the draggable anchors if (withAnchors) { drawDragAnchor(imageX, imageY); drawDragAnchor(imageRight, imageY); drawDragAnchor(imageRight, imageBottom); drawDragAnchor(imageX, imageBottom); } // optionally draw the connecting anchor lines if (withBorders) { ctx.beginPath(); ctx.moveTo(imageX, imageY); ctx.lineTo(imageRight, imageY); ctx.lineTo(imageRight, imageBottom); ctx.lineTo(imageX, imageBottom); ctx.closePath(); ctx.stroke(); } } function drawDragAnchor(x, y) { ctx.beginPath(); ctx.arc(x, y, resizerRadius, 0, pi2, false); ctx.closePath(); ctx.fill(); } function anchorHitTest(x, y) { var dx, dy; // top-left dx = x - canvas2realX(imageX); dy = y - canvas2realY(imageY); if (dx * dx + dy * dy <= rr) { return (0); } // top-right dx = x - canvas2realX(imageRight); dy = y - canvas2realY(imageY); console.log(dx.toString()+' '+dy.toString()); if (dx * dx + dy * dy <= rr) { return (1); } // bottom-right dx = x - canvas2realX(imageRight); dy = y - canvas2realY(imageBottom); if (dx * dx + dy * dy <= rr)
// bottom-left dx = x - canvas2realX(imageX); dy = y - canvas2realY(imageBottom); if (dx * dx + dy * dy <= rr) { return (3); } return (-1); } function hitImage(x, y) { return (x > canvas2realX(imageX) && x < canvas2realX(imageX + imageWidth) && y > canvas2realY(imageY) && y < canvas2realY(imageY + imageHeight)); } function handleMouseDown(e) { if (!flagPin) { startX = parseInt(e.clientX - document.getElementById('canvas').offsetLeft); startY = parseInt(e.clientY - document.getElementById('canvas').offsetTop); console.log(startX.toString()+'_'+startY.toString()); draggingResizer = anchorHitTest(startX, startY); draggingImage = draggingResizer < 0 && hitImage(startX, startY); } else { var row = tableGlobal.insertRow(tableGlobal.rows.length); var cell1 = row.insertCell(0); var cell2 = row.insertCell(1); cell2.innerHTML = Math.round((canvas.clientHeight - parseInt(e.clientY - document.getElementById('canvas').offsetTop) - canvas2realY(yPadding)) / ((canvas.clientHeight - 2 * canvas2realY(yPadding)) / numMaxY)); cell1.innerHTML = Math.round((parseInt(e.clientX - document.getElementById('canvas').offsetLeft) - canvas2realX(xPadding)) / ((canvas.clientWidth - 2 * canvas2realX(xPadding)) / numMaxX)); } } function handleMouseUp(e) { if (!flagPin) { draggingResizer = -1; draggingImage = false; draw(true, false); } } function handleMouseOut(e) { if (!flagPin) { handleMouseUp(e); } } function handleMouseMove(e) { if (!flagPin) { if (draggingResizer > -1) { mouseX = parseInt(e.clientX - document.getElementById('canvas').offsetLeft)*canvas.width/canvas.clientWidth; mouseY = parseInt(e.clientY - document.getElementById('canvas').offsetTop)*canvas.height/canvas.clientHeight; console.log(mouseX.toString()+','+mouseY.toString()); // resize the image switch (draggingResizer) { case 0: //top-left imageX = mouseX; imageWidth = imageRight - mouseX; imageY = mouseY; imageHeight =
{ return (2); }
conditional_block
app.js
izer = { x: 0, y: 0 }; var imageX = 50; var imageY = 50; var tableGlobal; // var numXscale = canvas.clientWidth / canvas.width; // var numYscale = canvas.clientHeight / canvas.height; var imageWidth, imageHeight, imageRight, imageBottom; var draggingImage = false; var startX; var startY; // var img = new Image(); // img.onload = function () { // imageWidth = img.width; // imageHeight = img.height; // imageRight = imageX + imageWidth; // imageBottom = imageY + imageHeight; // draw(true, false); // }; // img.src = "https://dl.dropboxusercontent.com/u/139992952/stackoverflow/facesSmall.png"; var imagDrop = document.createElement("img"); imagDrop.addEventListener("load", function() { //clearCanvas(); imageWidth = imagDrop.width; imageHeight = imagDrop.height; imageRight = imageX + imageWidth; imageBottom = imageY + imageHeight; draw(true, false); console.log("test"); //ctx.drawImage(imagDrop, 0, 0); }, false); // To enable drag and drop canvas.addEventListener("dragover", function(evt) { evt.preventDefault(); }, false); // Handle dropped image file - only Firefox and Google Chrome canvas.addEventListener("drop", function(evt) { var files = evt.dataTransfer.files; if (files.length > 0) { var file = files[0]; if (typeof FileReader !== "undefined" && file.type.indexOf("image") != -1) { var reader = new FileReader(); // Note: addEventListener doesn't work in Google Chrome for this event reader.onload = function(evt) { imagDrop.src = evt.target.result;
}; reader.readAsDataURL(file); } } evt.preventDefault(); }, false); updateAxis(); addTable(); function togglePin() { if (flagPin) { document.getElementById('idButtonPin').innerHTML = 'Start pinning'; flagPin = false; } else { document.getElementById('idButtonPin').innerHTML = 'Stop pinning'; flagPin = true; } } function addTable() { var newTable = document.createElement('table'); newTable.setAttribute("style", "float:left;position:relative"); var header = newTable.createTHead(); var row = header.insertRow(0); var cell = row.insertCell(0); cell.innerHTML = "X"; var cell = row.insertCell(1); cell.innerHTML = "Y"; document.getElementById('tables').appendChild(newTable); tableGlobal = newTable; } function canvas2realX(numX) { return numX*canvas.clientWidth / canvas.width; } function canvas2realY(numY) { return numY*canvas.clientHeight / canvas.height; } function drawAxis() { ctx.lineWidth = 3; ctx.strokeStyle = '#333'; ctx.font = 'italic 8pt sans-serif'; ctx.textAlign = "center"; // Draw the axises ctx.beginPath(); ctx.moveTo(xPadding, canvas.height - yPadding); ctx.lineTo(xPadding, 0); ctx.stroke(); ctx.beginPath(); ctx.moveTo(xPadding, canvas.height - yPadding); ctx.lineTo(canvas.width, canvas.height - yPadding); ctx.stroke(); // // Draw the X value texts ctx.lineWidth = 1; for (var i = 0; i <= numMaxX - numMinX; i += numStepX) { ctx.fillText(i, getXPixel(i), canvas.height - yPadding + 20); ctx.beginPath(); ctx.moveTo(getXPixel(i), canvas.height - yPadding); ctx.lineTo(getXPixel(i), 0); ctx.stroke(); } ctx.textAlign = "right"; ctx.textBaseline = "middle"; for (var i = 0; i <= numMaxY - numMinY; i += numStepY) { ctx.fillText(i, xPadding - 10, getYPixel(i)); ctx.beginPath(); ctx.moveTo(xPadding, getYPixel(i)); ctx.lineTo(canvas.width, getYPixel(i)); ctx.stroke(); } ctx.strokeStyle = '#f00'; } function updateAxis() { numMaxX = parseInt(document.getElementById('MaxX').value); numMaxY = parseInt(document.getElementById('MaxY').value); numStepX = parseInt(document.getElementById('StepX').value); numStepY = parseInt(document.getElementById('StepY').value); ctx.clearRect(0, 0, canvas.width, canvas.height); drawAxis(); } function getXPixel(val) { return ((canvas.width - 2 * xPadding) / numMaxX) * val + (xPadding); } // Return the y pixel for a graph point function getYPixel(val) { return canvas.height - (((canvas.height - 2 * yPadding) / numMaxY) * val) - yPadding; } function draw(withAnchors, withBorders) { // clear the canvas ctx.clearRect(0, 0, canvas.width, canvas.height); updateAxis(); // draw the image ctx.globalAlpha = 0.4; ctx.drawImage(imagDrop, 0, 0, imagDrop.width, imagDrop.height, imageX, imageY, imageWidth, imageHeight); // optionally draw the draggable anchors if (withAnchors) { drawDragAnchor(imageX, imageY); drawDragAnchor(imageRight, imageY); drawDragAnchor(imageRight, imageBottom); drawDragAnchor(imageX, imageBottom); } // optionally draw the connecting anchor lines if (withBorders) { ctx.beginPath(); ctx.moveTo(imageX, imageY); ctx.lineTo(imageRight, imageY); ctx.lineTo(imageRight, imageBottom); ctx.lineTo(imageX, imageBottom); ctx.closePath(); ctx.stroke(); } } function drawDragAnchor(x, y) { ctx.beginPath(); ctx.arc(x, y, resizerRadius, 0, pi2, false); ctx.closePath(); ctx.fill(); } function anchorHitTest(x, y) { var dx, dy; // top-left dx = x - canvas2realX(imageX); dy = y - canvas2realY(imageY); if (dx * dx + dy * dy <= rr) { return (0); } // top-right dx = x - canvas2realX(imageRight); dy = y - canvas2realY(imageY); console.log(dx.toString()+' '+dy.toString()); if (dx * dx + dy * dy <= rr) { return (1); } // bottom-right dx = x - canvas2realX(imageRight); dy = y - canvas2realY(imageBottom); if (dx * dx + dy * dy <= rr) { return (2); } // bottom-left dx = x - canvas2realX(imageX); dy = y - canvas2realY(imageBottom); if (dx * dx + dy * dy <= rr) { return (3); } return (-1); } function hitImage(x, y) { return (x > canvas2realX(imageX) && x < canvas2realX(imageX + imageWidth) && y > canvas2realY(imageY) && y < canvas2realY(imageY + imageHeight)); } function handleMouseDown(e) { if (!flagPin) { startX = parseInt(e.clientX - document.getElementById('canvas').offsetLeft); startY = parseInt(e.clientY - document.getElementById('canvas').offsetTop); console.log(startX.toString()+'_'+startY.toString()); draggingResizer = anchorHitTest(startX, startY); draggingImage = draggingResizer < 0 && hitImage(startX, startY); } else { var row = tableGlobal.insertRow(tableGlobal.rows.length); var cell1 = row.insertCell(0); var cell2 = row.insertCell(1); cell2.innerHTML = Math.round((canvas.clientHeight - parseInt(e.clientY - document.getElementById('canvas').offsetTop) - canvas2realY(yPadding)) / ((canvas.clientHeight - 2 * canvas2realY(yPadding)) / numMaxY)); cell1.innerHTML = Math.round((parseInt(e.clientX - document.getElementById('canvas').offsetLeft) - canvas2realX(xPadding)) / ((canvas.clientWidth - 2 * canvas2realX(xPadding)) / numMaxX)); } } function handleMouseUp(e) { if (!flagPin) { draggingResizer = -1; draggingImage = false; draw(true, false); } } function handleMouseOut(e) { if (!flagPin) { handleMouseUp(e); } } function handleMouseMove(e) { if (!flagPin) { if (draggingResizer > -1) { mouseX = parseInt(e.clientX - document.getElementById('canvas').offsetLeft)*canvas.width/canvas.clientWidth; mouseY = parseInt(e.clientY - document.getElementById('canvas').offsetTop)*canvas.height/canvas.clientHeight; console.log(mouseX.toString()+','+mouseY.toString()); // resize the image switch (draggingResizer) { case 0: //top-left imageX = mouseX; imageWidth = imageRight - mouseX; imageY = mouseY; imageHeight = image
random_line_split
app.js
= { x: 0, y: 0 }; var imageX = 50; var imageY = 50; var tableGlobal; // var numXscale = canvas.clientWidth / canvas.width; // var numYscale = canvas.clientHeight / canvas.height; var imageWidth, imageHeight, imageRight, imageBottom; var draggingImage = false; var startX; var startY; // var img = new Image(); // img.onload = function () { // imageWidth = img.width; // imageHeight = img.height; // imageRight = imageX + imageWidth; // imageBottom = imageY + imageHeight; // draw(true, false); // }; // img.src = "https://dl.dropboxusercontent.com/u/139992952/stackoverflow/facesSmall.png"; var imagDrop = document.createElement("img"); imagDrop.addEventListener("load", function() { //clearCanvas(); imageWidth = imagDrop.width; imageHeight = imagDrop.height; imageRight = imageX + imageWidth; imageBottom = imageY + imageHeight; draw(true, false); console.log("test"); //ctx.drawImage(imagDrop, 0, 0); }, false); // To enable drag and drop canvas.addEventListener("dragover", function(evt) { evt.preventDefault(); }, false); // Handle dropped image file - only Firefox and Google Chrome canvas.addEventListener("drop", function(evt) { var files = evt.dataTransfer.files; if (files.length > 0) { var file = files[0]; if (typeof FileReader !== "undefined" && file.type.indexOf("image") != -1) { var reader = new FileReader(); // Note: addEventListener doesn't work in Google Chrome for this event reader.onload = function(evt) { imagDrop.src = evt.target.result; }; reader.readAsDataURL(file); } } evt.preventDefault(); }, false); updateAxis(); addTable(); function togglePin() { if (flagPin) { document.getElementById('idButtonPin').innerHTML = 'Start pinning'; flagPin = false; } else { document.getElementById('idButtonPin').innerHTML = 'Stop pinning'; flagPin = true; } } function addTable() { var newTable = document.createElement('table'); newTable.setAttribute("style", "float:left;position:relative"); var header = newTable.createTHead(); var row = header.insertRow(0); var cell = row.insertCell(0); cell.innerHTML = "X"; var cell = row.insertCell(1); cell.innerHTML = "Y"; document.getElementById('tables').appendChild(newTable); tableGlobal = newTable; } function canvas2realX(numX) { return numX*canvas.clientWidth / canvas.width; } function canvas2realY(numY) { return numY*canvas.clientHeight / canvas.height; } function drawAxis() { ctx.lineWidth = 3; ctx.strokeStyle = '#333'; ctx.font = 'italic 8pt sans-serif'; ctx.textAlign = "center"; // Draw the axises ctx.beginPath(); ctx.moveTo(xPadding, canvas.height - yPadding); ctx.lineTo(xPadding, 0); ctx.stroke(); ctx.beginPath(); ctx.moveTo(xPadding, canvas.height - yPadding); ctx.lineTo(canvas.width, canvas.height - yPadding); ctx.stroke(); // // Draw the X value texts ctx.lineWidth = 1; for (var i = 0; i <= numMaxX - numMinX; i += numStepX) { ctx.fillText(i, getXPixel(i), canvas.height - yPadding + 20); ctx.beginPath(); ctx.moveTo(getXPixel(i), canvas.height - yPadding); ctx.lineTo(getXPixel(i), 0); ctx.stroke(); } ctx.textAlign = "right"; ctx.textBaseline = "middle"; for (var i = 0; i <= numMaxY - numMinY; i += numStepY) { ctx.fillText(i, xPadding - 10, getYPixel(i)); ctx.beginPath(); ctx.moveTo(xPadding, getYPixel(i)); ctx.lineTo(canvas.width, getYPixel(i)); ctx.stroke(); } ctx.strokeStyle = '#f00'; } function updateAxis() { numMaxX = parseInt(document.getElementById('MaxX').value); numMaxY = parseInt(document.getElementById('MaxY').value); numStepX = parseInt(document.getElementById('StepX').value); numStepY = parseInt(document.getElementById('StepY').value); ctx.clearRect(0, 0, canvas.width, canvas.height); drawAxis(); } function getXPixel(val) { return ((canvas.width - 2 * xPadding) / numMaxX) * val + (xPadding); } // Return the y pixel for a graph point function getYPixel(val) { return canvas.height - (((canvas.height - 2 * yPadding) / numMaxY) * val) - yPadding; } function
(withAnchors, withBorders) { // clear the canvas ctx.clearRect(0, 0, canvas.width, canvas.height); updateAxis(); // draw the image ctx.globalAlpha = 0.4; ctx.drawImage(imagDrop, 0, 0, imagDrop.width, imagDrop.height, imageX, imageY, imageWidth, imageHeight); // optionally draw the draggable anchors if (withAnchors) { drawDragAnchor(imageX, imageY); drawDragAnchor(imageRight, imageY); drawDragAnchor(imageRight, imageBottom); drawDragAnchor(imageX, imageBottom); } // optionally draw the connecting anchor lines if (withBorders) { ctx.beginPath(); ctx.moveTo(imageX, imageY); ctx.lineTo(imageRight, imageY); ctx.lineTo(imageRight, imageBottom); ctx.lineTo(imageX, imageBottom); ctx.closePath(); ctx.stroke(); } } function drawDragAnchor(x, y) { ctx.beginPath(); ctx.arc(x, y, resizerRadius, 0, pi2, false); ctx.closePath(); ctx.fill(); } function anchorHitTest(x, y) { var dx, dy; // top-left dx = x - canvas2realX(imageX); dy = y - canvas2realY(imageY); if (dx * dx + dy * dy <= rr) { return (0); } // top-right dx = x - canvas2realX(imageRight); dy = y - canvas2realY(imageY); console.log(dx.toString()+' '+dy.toString()); if (dx * dx + dy * dy <= rr) { return (1); } // bottom-right dx = x - canvas2realX(imageRight); dy = y - canvas2realY(imageBottom); if (dx * dx + dy * dy <= rr) { return (2); } // bottom-left dx = x - canvas2realX(imageX); dy = y - canvas2realY(imageBottom); if (dx * dx + dy * dy <= rr) { return (3); } return (-1); } function hitImage(x, y) { return (x > canvas2realX(imageX) && x < canvas2realX(imageX + imageWidth) && y > canvas2realY(imageY) && y < canvas2realY(imageY + imageHeight)); } function handleMouseDown(e) { if (!flagPin) { startX = parseInt(e.clientX - document.getElementById('canvas').offsetLeft); startY = parseInt(e.clientY - document.getElementById('canvas').offsetTop); console.log(startX.toString()+'_'+startY.toString()); draggingResizer = anchorHitTest(startX, startY); draggingImage = draggingResizer < 0 && hitImage(startX, startY); } else { var row = tableGlobal.insertRow(tableGlobal.rows.length); var cell1 = row.insertCell(0); var cell2 = row.insertCell(1); cell2.innerHTML = Math.round((canvas.clientHeight - parseInt(e.clientY - document.getElementById('canvas').offsetTop) - canvas2realY(yPadding)) / ((canvas.clientHeight - 2 * canvas2realY(yPadding)) / numMaxY)); cell1.innerHTML = Math.round((parseInt(e.clientX - document.getElementById('canvas').offsetLeft) - canvas2realX(xPadding)) / ((canvas.clientWidth - 2 * canvas2realX(xPadding)) / numMaxX)); } } function handleMouseUp(e) { if (!flagPin) { draggingResizer = -1; draggingImage = false; draw(true, false); } } function handleMouseOut(e) { if (!flagPin) { handleMouseUp(e); } } function handleMouseMove(e) { if (!flagPin) { if (draggingResizer > -1) { mouseX = parseInt(e.clientX - document.getElementById('canvas').offsetLeft)*canvas.width/canvas.clientWidth; mouseY = parseInt(e.clientY - document.getElementById('canvas').offsetTop)*canvas.height/canvas.clientHeight; console.log(mouseX.toString()+','+mouseY.toString()); // resize the image switch (draggingResizer) { case 0: //top-left imageX = mouseX; imageWidth = imageRight - mouseX; imageY = mouseY; imageHeight = image
draw
identifier_name
Z_normal_8_2.py
1.15]), 9: np.array([ -1.22, -0.76, -0.43, -0.14, 0.14, 0.43, 0.76, 1.22]), 10: np.array([ -1.28, -0.84, -0.52, -0.25, 0, 0.25, 0.52, 0.84, 1.28]), 11: np.array([ -1.34, -0.91, -0.6, -0.35, -0.11, 0.11, 0.35, 0.6, 0.91, 1.34]), 12: np.array([ -1.38, -0.97, -0.67, -0.43, -0.21, 0, 0.21, 0.43, 0.67, 0.97, 1.38]), 13: np.array([ -1.43, -1.02, -0.74, -0.5, -0.29, -0.1, 0.1, 0.29, 0.5, 0.74, 1.02, 1.43]), 14: np.array([ -1.47, -1.07, -0.79, -0.57, -0.37, -0.18, 0, 0.18, 0.37, 0.57, 0.79, 1.07, 1.47]), 15: np.array([ -1.5, -1.11, -0.84, -0.62, -0.43, -0.25, -0.08, 0.08, 0.25, 0.43, 0.62, 0.84, 1.11, 1.5]), 16: np.array([ -1.53, -1.15, -0.89, -0.67, -0.49, -0.32, -0.16, 0, 0.16, 0.32, 0.49, 0.67, 0.89, 1.15, 1.53]), 17: np.array([ -1.56, -1.19, -0.93, -0.72, -0.54, -0.38, -0.22, -0.07, 0.07, 0.22, 0.38, 0.54, 0.72, 0.93, 1.19, 1.56]), 18: np.array([ -1.59, -1.22, -0.97, -0.76, -0.59, -0.43, -0.28, -0.14, 0, 0.14, 0.28, 0.43, 0.59, 0.76, 0.97, 1.22, 1.59]), 19: np.array([ -1.62, -1.25, -1, -0.8, -0.63, -0.48, -0.34, -0.2, -0.07, 0.07, 0.2, 0.34, 0.48, 0.63, 0.8, 1, 1.25, 1.62]), 20: np.array([ -1.64, -1.28, -1.04, -0.84, -0.67, -0.52, -0.39, -0.25, -0.13, 0, 0.13, 0.25, 0.39, 0.52, 0.67, 0.84, 1.04, 1.28, 1.64]), } return options[size] def break_points_quantiles(size): options=np.linspace(0, 1, size+1)[1:] return options #y_alphabets = break_points_quantiles(y_alphabet_size).tolist() y_alphabets = break_points_gaussian(y_alphabet_size).tolist() def hamming_distance1(string1, string2): distance = 0 L = len(string1) for i in range(L): if string1[i] != string2[i]: distance += 1 return distance def hamming_distance(s1, s2): if len(s1) != len(s2): raise ValueError("Undefined for sequences of unequal length") return sum(el1 != el2 for el1, el2 in zip(s1, s2)) """------------- X-axis Distribution ------------- """ def x_distrubted_values(series): mean=np.mean(series) #median=sorted(series)[len(series) // 2] return mean """------------- Index to Letter conversion ------------- """ def index_to_letter(idx): """Convert a numerical index to a char.""" if 0 <= idx < 20: return chr(97 + idx) else: raise ValueError('A wrong idx value supplied.') def normalize(x): X = np.asanyarray(x) if np.nanstd(X) < epsilon: res = [] for entry in X: if not np.isnan(entry): res.append(0) else: res.append(np.nan) return res return (X - np.nanmean(X)) / np.nanstd(X) def normal_distribution(x): x = (x-min(x))/(max(x)-min(x)) return x """------------- 1- Normalize Data ------------- """ x1=normalize(x1) plt.plot(x1) plt.show() """------------- 5.2- Y_Alphabetize ------------- """ def alphabetize_ts(sub_section): mean_val=x_distrubted_values(sub_section) y_alpha_val=min(y_alphabets, key=lambda x:abs(x-mean_val)) y_alpha_idx=y_alphabets.index(y_alpha_val) curr_word = index_to_letter(y_alpha_idx) return(curr_word) """------------- 2- Segmentization Data ------------- """ def segment_ts(series,windowSize=window_size,skip_offset=skip_offset): ts_len=len(x1) mod = ts_len%windowSize rnge=0 if(skip_offset==0): ts_len=int((ts_len-mod-window_size)/1) rnge=int(ts_len/window_size) else: ts_len=int(math.ceil((ts_len-mod-window_size)/skip_offset)) rnge=int(ts_len) curr_count=0 words=list() indices=list() complete_indices=list() for i in range(0, rnge): sub_section = series[curr_count:(curr_count+windowSize)] sub_section=normalize(sub_section) #print(curr_count,(curr_count+windowSize)) #print(sub_section) curr_word="" chunk_size=int(len(sub_section)/word_lenth) num=0 zlp=""
num+=chunk_size words.append(zlp) indices.append(curr_count) curr_count=curr_count+skip_offset-1 temp_list=[] temp_list.append(sub_section) temp_df = pd.DataFrame(temp_list) temp_df.insert(loc=0, column='keys', value=zlp) temp_df.insert(loc=1, column='position', value=sorted(sub_section)[len(sub_section) // 2]) temp_df.insert(loc=2, column='scale_high', value=np.max(sub_section)) temp_df.insert(loc=3, column='scale_low', value=np.min(sub_section)) if(i==0): df_sax =temp_df.copy() else: df_sax=df_sax.append(temp_df, ignore_index=True) return (words,indices,df_sax) alphabetize11,indices,df_sax=segment_ts(x1) """------------- SAX ------------- """ """ Complete Words """ def complete_word(series=x1,word_len=word_lenth,skip_len=skip_offset): alphabetize,indices,df_sax=segment_ts(series) complete_word=list()
for j in range(0,word_lenth): chunk = sub_section[num:num + chunk_size] curr_word=alphabetize_ts(chunk) zlp+=str(curr_word) complete_indices.append(curr_count)
random_line_split
Z_normal_8_2.py
1.15]), 9: np.array([ -1.22, -0.76, -0.43, -0.14, 0.14, 0.43, 0.76, 1.22]), 10: np.array([ -1.28, -0.84, -0.52, -0.25, 0, 0.25, 0.52, 0.84, 1.28]), 11: np.array([ -1.34, -0.91, -0.6, -0.35, -0.11, 0.11, 0.35, 0.6, 0.91, 1.34]), 12: np.array([ -1.38, -0.97, -0.67, -0.43, -0.21, 0, 0.21, 0.43, 0.67, 0.97, 1.38]), 13: np.array([ -1.43, -1.02, -0.74, -0.5, -0.29, -0.1, 0.1, 0.29, 0.5, 0.74, 1.02, 1.43]), 14: np.array([ -1.47, -1.07, -0.79, -0.57, -0.37, -0.18, 0, 0.18, 0.37, 0.57, 0.79, 1.07, 1.47]), 15: np.array([ -1.5, -1.11, -0.84, -0.62, -0.43, -0.25, -0.08, 0.08, 0.25, 0.43, 0.62, 0.84, 1.11, 1.5]), 16: np.array([ -1.53, -1.15, -0.89, -0.67, -0.49, -0.32, -0.16, 0, 0.16, 0.32, 0.49, 0.67, 0.89, 1.15, 1.53]), 17: np.array([ -1.56, -1.19, -0.93, -0.72, -0.54, -0.38, -0.22, -0.07, 0.07, 0.22, 0.38, 0.54, 0.72, 0.93, 1.19, 1.56]), 18: np.array([ -1.59, -1.22, -0.97, -0.76, -0.59, -0.43, -0.28, -0.14, 0, 0.14, 0.28, 0.43, 0.59, 0.76, 0.97, 1.22, 1.59]), 19: np.array([ -1.62, -1.25, -1, -0.8, -0.63, -0.48, -0.34, -0.2, -0.07, 0.07, 0.2, 0.34, 0.48, 0.63, 0.8, 1, 1.25, 1.62]), 20: np.array([ -1.64, -1.28, -1.04, -0.84, -0.67, -0.52, -0.39, -0.25, -0.13, 0, 0.13, 0.25, 0.39, 0.52, 0.67, 0.84, 1.04, 1.28, 1.64]), } return options[size] def break_points_quantiles(size): options=np.linspace(0, 1, size+1)[1:] return options #y_alphabets = break_points_quantiles(y_alphabet_size).tolist() y_alphabets = break_points_gaussian(y_alphabet_size).tolist() def hamming_distance1(string1, string2): distance = 0 L = len(string1) for i in range(L): if string1[i] != string2[i]: distance += 1 return distance def hamming_distance(s1, s2): if len(s1) != len(s2): raise ValueError("Undefined for sequences of unequal length") return sum(el1 != el2 for el1, el2 in zip(s1, s2)) """------------- X-axis Distribution ------------- """ def x_distrubted_values(series): mean=np.mean(series) #median=sorted(series)[len(series) // 2] return mean """------------- Index to Letter conversion ------------- """ def index_to_letter(idx): """Convert a numerical index to a char.""" if 0 <= idx < 20:
else: raise ValueError('A wrong idx value supplied.') def normalize(x): X = np.asanyarray(x) if np.nanstd(X) < epsilon: res = [] for entry in X: if not np.isnan(entry): res.append(0) else: res.append(np.nan) return res return (X - np.nanmean(X)) / np.nanstd(X) def normal_distribution(x): x = (x-min(x))/(max(x)-min(x)) return x """------------- 1- Normalize Data ------------- """ x1=normalize(x1) plt.plot(x1) plt.show() """------------- 5.2- Y_Alphabetize ------------- """ def alphabetize_ts(sub_section): mean_val=x_distrubted_values(sub_section) y_alpha_val=min(y_alphabets, key=lambda x:abs(x-mean_val)) y_alpha_idx=y_alphabets.index(y_alpha_val) curr_word = index_to_letter(y_alpha_idx) return(curr_word) """------------- 2- Segmentization Data ------------- """ def segment_ts(series,windowSize=window_size,skip_offset=skip_offset): ts_len=len(x1) mod = ts_len%windowSize rnge=0 if(skip_offset==0): ts_len=int((ts_len-mod-window_size)/1) rnge=int(ts_len/window_size) else: ts_len=int(math.ceil((ts_len-mod-window_size)/skip_offset)) rnge=int(ts_len) curr_count=0 words=list() indices=list() complete_indices=list() for i in range(0, rnge): sub_section = series[curr_count:(curr_count+windowSize)] sub_section=normalize(sub_section) #print(curr_count,(curr_count+windowSize)) #print(sub_section) curr_word="" chunk_size=int(len(sub_section)/word_lenth) num=0 zlp="" for j in range(0,word_lenth): chunk = sub_section[num:num + chunk_size] curr_word=alphabetize_ts(chunk) zlp+=str(curr_word) complete_indices.append(curr_count) num+=chunk_size words.append(zlp) indices.append(curr_count) curr_count=curr_count+skip_offset-1 temp_list=[] temp_list.append(sub_section) temp_df = pd.DataFrame(temp_list) temp_df.insert(loc=0, column='keys', value=zlp) temp_df.insert(loc=1, column='position', value=sorted(sub_section)[len(sub_section) // 2]) temp_df.insert(loc=2, column='scale_high', value=np.max(sub_section)) temp_df.insert(loc=3, column='scale_low', value=np.min(sub_section)) if(i==0): df_sax =temp_df.copy() else: df_sax=df_sax.append(temp_df, ignore_index=True) return (words,indices,df_sax) alphabetize11,indices,df_sax=segment_ts(x1) """------------- SAX ------------- """ """ Complete Words """ def complete_word(series=x1,word_len=word_lenth,skip_len=skip_offset): alphabetize,indices,df_sax=segment_ts(series) complete_word=list()
return chr(97 + idx)
conditional_block