file_name
large_stringlengths 4
140
| prefix
large_stringlengths 0
12.1k
| suffix
large_stringlengths 0
12k
| middle
large_stringlengths 0
7.51k
| fim_type
large_stringclasses 4
values |
---|---|---|---|---|
detect_outlier.py | import numpy as np
import sys
sys.path.append("../Pipeline/Audio/Pipeline/")
from AudioPipe.features import mfcc # Feature Extraction Module, part of the shared preprocessing
import AudioPipe.speaker.recognition as SR # Speaker Recognition Module
import scipy.io.wavfile as wav
import commands, os
import json
import argparse
import warnings
from scipy import stats
def outlier_detect(audio_dir, spk_name):
spk_dir = os.path.join(audio_dir,spk_name)
list_fn = os.path.join(spk_dir,"clip_list.txt")
clip_ls = from_jsonfile(list_fn)
audio_merge = merge_clips(spk_dir, clip_ls)
# Training a model based on the merged audio
Model = SR.GMMRec()
Model.enroll_file(spk_name, audio_merge)
Model.train()
# Score each utterance in the training set
llhd_ls = []
new_ls = []
stat_fn = os.path.join(audio_dir,"stats.json")
if os.path.exists(stat_fn) and os.path.getsize(stat_fn) > 0:
stat_dict = from_jsonfile(stat_fn)
else:
stat_dict = {}
if spk_name not in stat_dict:
stat_dict[spk_name]={}
for clip in clip_ls:
audio_test = os.path.join(spk_dir,clip["name"])
#commands.getstatusoutput("ffmpeg -i "+audio_test+" -vn -f wav -ab 16k "+audio_test)
try:
llhd = Model.predict(Model.get_mfcc(audio_test))[1]
except ValueError:
print clip["name"]
continue
llhd_ls.append(llhd)
clip["llhd"] = llhd
new_ls.append(clip)
z_score = stats.zscore(llhd_ls)
for i in xrange(len(llhd_ls)):
|
with open(list_fn, "w") as fh:
fh.write(to_json(new_ls, indent=2))
stat_dict[spk_name]["clip_num"]=len(clip_ls)
stat_dict[spk_name]["zpos_num"]=sum(z_score>0)
stat_dict[spk_name]["total_duration"]=sum([get_sec(clp["duration"]) for clp in new_ls])
stat_dict[spk_name]["clean_duration"]=sum([get_sec(clp["duration"]) for clp in new_ls if clp["zscore"]>-0.00001])
with open(stat_fn, "w") as fh:
fh.write(to_json(stat_dict, indent=2))
os.remove(audio_merge)
return llhd_ls
def merge_clips(spk_dir, clip_ls):
# Write the list of clips into a file for merging training data
temp_fl = os.path.join(spk_dir,"temp.txt")
count = 0
with open(temp_fl, "w") as fh:
for clip in clip_ls:
if count>100:
break
fh.write("file "+clip["name"]+"\n")
count+=1
# Merge all the data into one audio
audio_merge = os.path.join(spk_dir,"merged_gross.wav")
commands.getstatusoutput("ffmpeg -f concat -i "+temp_fl.replace(" ", "\ ")+" -c copy -y "+audio_merge)
os.remove(temp_fl)
return audio_merge
def from_jsonfile(filename):
with open(filename) as fh:
return json.load(fh)
def get_sec(time_str):
h, m, s = time_str.split(':')
return int(h) * 3600 + int(m) * 60 + float(s)
def to_json(result, **kwargs):
'''Return a JSON representation of the aligned transcript'''
options = {
'sort_keys': True,
'indent': 4,
'separators': (',', ': '),
}
options.update(kwargs)
return json.dumps(result, **options)
parser = argparse.ArgumentParser(
description='Detect outliers in a training dataset of one speaker.')
parser.add_argument(
'-i', '--input', dest='input_dir', type=str,
help='directory of audio clips')
parser.add_argument(
'-s', '--spk', dest='spk_name', type=str,
help='the name of the speaker')
args = parser.parse_args()
audio_dir = args.input_dir
spk_name = args.spk_name
with warnings.catch_warnings():
warnings.simplefilter("ignore")
outlier_detect(audio_dir, spk_name)
| new_ls[i]["zscore"] = z_score[i] | conditional_block |
activity.js | (function () {
'use strict';
var module = angular.module('fim.base');
module.config(function($routeProvider) {
$routeProvider
.when('/activity/:engine/:section/:period', {
templateUrl: 'partials/activity.html',
controller: 'ActivityController'
});
});
module.controller('ActivityController', function($scope, $location, $routeParams, nxt, requests, $q, $sce,
ActivityProvider, BlocksProvider, ForgersProvider, StatisticsProvider, AllAssetsProvider, BlockStateProvider,
$timeout, dateParser, dateFilter, $rootScope) {
$rootScope.paramEngine = $routeParams.engine;
$scope.paramEngine = $routeParams.engine;
$scope.paramSection = $routeParams.section;
$scope.paramPeriod = $routeParams.period;
$scope.paramTimestamp = 0;
$scope.statistics = {};
$scope.blockstate = {};
$scope.breadcrumb = [];
$scope.filter = {};
if ($scope.paramEngine == 'nxt') { var api = nxt.nxt(); }
else if ($scope.paramEngine == 'fim') { var api = nxt.fim(); }
else {
$location.path('/activity/fim/activity/latest');
return;
}
if (['activity', 'blockchain', 'forgers', 'assets'].indexOf($scope.paramSection) == -1) {
$location.path('/activity/'+$scope.paramEngine+'/activity/latest');
return;
}
/* Breadcrumbs */
$scope.breadcrumb.push({
label: 'translate.home',
href: "#/home/"+$scope.paramEngine+"/activity/latest",
translate: true
});
$scope.breadcrumb.push({
label: 'translate.explorer',
href: '#/activity/fim/activity/latest',
translate: true
});
$scope.breadcrumb.push({
label: api.engine.symbol,
active: true
});
$scope.breadcrumb.push({
label: 'translate.'+$scope.paramSection,
translate: true
});
if (['activity','blockchain'].indexOf($scope.paramSection) != -1) {
if ($scope.paramPeriod == 'latest') {
$scope.breadcrumb.push({
label: 'translate.latest',
translate: true
});
}
else {
$scope.breadcrumb.push({
label: $scope.paramPeriod,
period: true
});
}
}
/* Date picker */
$scope.dt = null;
$scope.format = 'dd-MMMM-yyyy';
if ($scope.paramPeriod != 'latest') {
var d = dateParser.parse($scope.paramPeriod, $scope.format);
if (!d) {
$location.path('/activity/'+$scope.paramEngine+'/'+$scope.paramSection+'/latest');
return;
}
$scope.dt = $scope.paramPeriod;
/* Timestamp is for 00:00 hour on selected day */
d = new Date(d.getFullYear(), d.getMonth(), d.getDate()+1, 0, 0, 0);
$scope.paramTimestamp = nxt.util.convertToEpochTimestamp(d.getTime());
}
$scope.symbol = api.engine.symbol;
$scope.blockstate['TYPE_FIM'] = new BlockStateProvider(nxt.fim(), $scope);
$scope.blockstate['TYPE_FIM'].load();
if ($rootScope.enableDualEngines) {
$scope.blockstate['TYPE_NXT'] = new BlockStateProvider(nxt.nxt(), $scope);
$scope.blockstate['TYPE_NXT'].load();
}
switch ($scope.paramSection) {
case 'activity':
$scope.showFilter = true;
$scope.showTransactionFilter = true;
$scope.provider = new ActivityProvider(api, $scope, $scope.paramTimestamp, null, $scope.filter);
$scope.provider.reload();
break;
case 'blockchain':
$scope.showFilter = true;
$scope.provider = new BlocksProvider(api, $scope, $scope.paramTimestamp);
$scope.provider.reload();
break;
case 'forgers':
$scope.showFilter = false;
$scope.provider = new ForgersProvider(api, $scope);
$scope.provider.reload();
break;
case 'assets':
$scope.showFilter = false;
$scope.provider = new AllAssetsProvider(api, $scope, 10);
$scope.provider.reload();
break;
default:
throw new Error('Not reached');
}
$scope.minDate = new Date(Date.UTC(2013, 10, 24, 12, 0, 0, 0));
$scope.maxDate = new Date();
$scope.dateOptions = {
formatYear: 'yy',
startingDay: 1
};
$scope.openDatePicker = function($event) {
$event.preventDefault();
$event.stopPropagation();
$scope.opened = true;
};
var stopWatching = false;
$scope.$watch('dt', function (newValue, oldValue) {
if (newValue && newValue !== oldValue && typeof oldValue != 'string' && !stopWatching) |
});
if ($scope.showTransactionFilter) {
$scope.filter.all = true;
$scope.filter.payments = true;
$scope.filter.messages = true;
$scope.filter.aliases = true;
$scope.filter.namespacedAliases = true;
$scope.filter.polls = true;
$scope.filter.accountInfo = true;
$scope.filter.announceHub = true;
$scope.filter.goodsStore = true;
$scope.filter.balanceLeasing = true;
$scope.filter.trades = true;
$scope.filter.assetIssued = true;
$scope.filter.assetTransfer = true;
$scope.filter.assetOrder = true;
$scope.filter.currencyIssued = true;
$scope.filter.currencyTransfer = true;
$scope.filter.currencyOther = true;
$scope.filterAllChanged = function () {
$scope.$evalAsync(function () {
var on = $scope.filter.all;
$scope.filter.payments = on;
$scope.filter.messages = on;
$scope.filter.aliases = on;
$scope.filter.namespacedAliases = on;
$scope.filter.polls = on;
$scope.filter.accountInfo = on;
$scope.filter.announceHub = on;
$scope.filter.goodsStore = on;
$scope.filter.balanceLeasing = on;
$scope.filter.trades = on;
$scope.filter.assetIssued = on;
$scope.filter.assetTransfer = on;
$scope.filter.assetOrder = on;
$scope.filter.currencyIssued = on;
$scope.filter.currencyTransfer = on;
$scope.filter.currencyOther = on;
$scope.filterChanged();
});
}
$scope.filterChanged = function () {
$scope.provider.applyFilter($scope.filter);
}
}
$scope.loadStatistics = function (engine, collapse_var) {
$scope[collapse_var] = !$scope[collapse_var];
if (!$scope[collapse_var]) {
if (!$scope.statistics[engine]) {
var api = nxt.get(engine);
$scope.statistics[engine] = new StatisticsProvider(api, $scope);
}
$scope.statistics[engine].load();
}
}
});
})(); | {
stopWatching = true;
var formatted = dateFilter(newValue, $scope.format);
$location.path('/activity/'+$scope.paramEngine+'/'+$scope.paramSection+'/'+formatted);
} | conditional_block |
activity.js | (function () {
'use strict';
var module = angular.module('fim.base');
module.config(function($routeProvider) {
$routeProvider
.when('/activity/:engine/:section/:period', {
templateUrl: 'partials/activity.html',
controller: 'ActivityController'
});
});
module.controller('ActivityController', function($scope, $location, $routeParams, nxt, requests, $q, $sce,
ActivityProvider, BlocksProvider, ForgersProvider, StatisticsProvider, AllAssetsProvider, BlockStateProvider,
$timeout, dateParser, dateFilter, $rootScope) {
$rootScope.paramEngine = $routeParams.engine;
$scope.paramEngine = $routeParams.engine;
$scope.paramSection = $routeParams.section;
$scope.paramPeriod = $routeParams.period;
$scope.paramTimestamp = 0;
$scope.statistics = {};
$scope.blockstate = {};
$scope.breadcrumb = [];
$scope.filter = {};
if ($scope.paramEngine == 'nxt') { var api = nxt.nxt(); }
else if ($scope.paramEngine == 'fim') { var api = nxt.fim(); }
else {
$location.path('/activity/fim/activity/latest');
return;
}
if (['activity', 'blockchain', 'forgers', 'assets'].indexOf($scope.paramSection) == -1) {
$location.path('/activity/'+$scope.paramEngine+'/activity/latest');
return;
}
/* Breadcrumbs */
$scope.breadcrumb.push({
label: 'translate.home',
href: "#/home/"+$scope.paramEngine+"/activity/latest",
translate: true
});
$scope.breadcrumb.push({
label: 'translate.explorer',
href: '#/activity/fim/activity/latest',
translate: true
});
$scope.breadcrumb.push({
label: api.engine.symbol,
active: true
});
$scope.breadcrumb.push({
label: 'translate.'+$scope.paramSection,
translate: true
});
if (['activity','blockchain'].indexOf($scope.paramSection) != -1) {
if ($scope.paramPeriod == 'latest') {
$scope.breadcrumb.push({
label: 'translate.latest',
translate: true
});
}
else {
$scope.breadcrumb.push({
label: $scope.paramPeriod,
period: true
});
}
}
/* Date picker */
$scope.dt = null;
$scope.format = 'dd-MMMM-yyyy';
if ($scope.paramPeriod != 'latest') {
var d = dateParser.parse($scope.paramPeriod, $scope.format);
if (!d) {
$location.path('/activity/'+$scope.paramEngine+'/'+$scope.paramSection+'/latest');
return;
}
$scope.dt = $scope.paramPeriod;
/* Timestamp is for 00:00 hour on selected day */
d = new Date(d.getFullYear(), d.getMonth(), d.getDate()+1, 0, 0, 0);
$scope.paramTimestamp = nxt.util.convertToEpochTimestamp(d.getTime());
}
$scope.symbol = api.engine.symbol;
$scope.blockstate['TYPE_FIM'] = new BlockStateProvider(nxt.fim(), $scope);
$scope.blockstate['TYPE_FIM'].load();
if ($rootScope.enableDualEngines) {
$scope.blockstate['TYPE_NXT'] = new BlockStateProvider(nxt.nxt(), $scope);
$scope.blockstate['TYPE_NXT'].load();
}
switch ($scope.paramSection) {
case 'activity':
$scope.showFilter = true;
$scope.showTransactionFilter = true;
$scope.provider = new ActivityProvider(api, $scope, $scope.paramTimestamp, null, $scope.filter);
$scope.provider.reload();
break;
case 'blockchain':
$scope.showFilter = true;
$scope.provider = new BlocksProvider(api, $scope, $scope.paramTimestamp);
$scope.provider.reload();
break;
case 'forgers':
$scope.showFilter = false;
$scope.provider = new ForgersProvider(api, $scope);
$scope.provider.reload();
break;
case 'assets':
$scope.showFilter = false;
$scope.provider = new AllAssetsProvider(api, $scope, 10);
$scope.provider.reload();
break;
default:
throw new Error('Not reached');
}
$scope.minDate = new Date(Date.UTC(2013, 10, 24, 12, 0, 0, 0));
$scope.maxDate = new Date();
$scope.dateOptions = {
formatYear: 'yy',
startingDay: 1
};
$scope.openDatePicker = function($event) {
$event.preventDefault();
$event.stopPropagation();
$scope.opened = true;
};
var stopWatching = false;
$scope.$watch('dt', function (newValue, oldValue) {
if (newValue && newValue !== oldValue && typeof oldValue != 'string' && !stopWatching) {
stopWatching = true;
var formatted = dateFilter(newValue, $scope.format);
$location.path('/activity/'+$scope.paramEngine+'/'+$scope.paramSection+'/'+formatted);
}
});
if ($scope.showTransactionFilter) {
$scope.filter.all = true;
$scope.filter.payments = true;
$scope.filter.messages = true;
$scope.filter.aliases = true;
$scope.filter.namespacedAliases = true;
$scope.filter.polls = true;
$scope.filter.accountInfo = true;
$scope.filter.announceHub = true;
$scope.filter.goodsStore = true;
$scope.filter.balanceLeasing = true;
$scope.filter.trades = true;
$scope.filter.assetIssued = true;
$scope.filter.assetTransfer = true;
$scope.filter.assetOrder = true;
$scope.filter.currencyIssued = true;
$scope.filter.currencyTransfer = true;
$scope.filter.currencyOther = true;
$scope.filterAllChanged = function () {
$scope.$evalAsync(function () {
var on = $scope.filter.all;
$scope.filter.payments = on;
$scope.filter.messages = on;
$scope.filter.aliases = on;
$scope.filter.namespacedAliases = on;
$scope.filter.polls = on;
$scope.filter.accountInfo = on;
$scope.filter.announceHub = on;
$scope.filter.goodsStore = on;
$scope.filter.balanceLeasing = on;
$scope.filter.trades = on;
$scope.filter.assetIssued = on;
$scope.filter.assetTransfer = on;
$scope.filter.assetOrder = on;
$scope.filter.currencyIssued = on;
$scope.filter.currencyTransfer = on;
$scope.filter.currencyOther = on;
$scope.filterChanged(); |
$scope.filterChanged = function () {
$scope.provider.applyFilter($scope.filter);
}
}
$scope.loadStatistics = function (engine, collapse_var) {
$scope[collapse_var] = !$scope[collapse_var];
if (!$scope[collapse_var]) {
if (!$scope.statistics[engine]) {
var api = nxt.get(engine);
$scope.statistics[engine] = new StatisticsProvider(api, $scope);
}
$scope.statistics[engine].load();
}
}
});
})(); | });
} | random_line_split |
set_watch_mode.rs | use std::io;
use super::super::{WriteTo, WatchMode, WatchModeReader, WriteResult, Reader, ReaderStatus, MessageInner, Message};
#[derive(Debug, Eq, PartialEq, Clone)]
pub struct SetWatchMode {
mode: WatchMode,
}
#[derive(Debug)]
pub struct SetWatchModeReader {
inner: WatchModeReader,
}
impl SetWatchMode {
pub fn new(mode: WatchMode) -> Self {
SetWatchMode { mode }
}
pub fn mode(&self) -> WatchMode {
self.mode
}
pub fn reader() -> SetWatchModeReader |
}
impl MessageInner for SetWatchMode {
#[inline]
fn wrap(self) -> Message {
Message::SetWatchMode(self)
}
}
impl Reader<SetWatchMode> for SetWatchModeReader {
fn resume<R>(&mut self, input: &mut R) -> io::Result<ReaderStatus<SetWatchMode>> where R: io::Read {
let status = self.inner.resume(input)?;
Ok(status.map(|mode| SetWatchMode::new(mode)))
}
fn rewind(&mut self) {
self.inner.rewind();
}
}
impl WriteTo for SetWatchMode {
fn write_to<W: io::Write>(&self, target: &mut W) -> WriteResult {
self.mode.write_to(target)
}
}
#[cfg(test)]
mod test {
use super::*;
use super::super::super::{MessageType, WatchMode};
#[test]
fn test_reader_with_tagged() {
let input = vec![
/* type */ MessageType::SetWatchMode.into(),
/* mode = tagged */ 2,
/* tag */ 0, 0, 0, 0, 0, 0, 255, 255
];
test_reader! {
Message::reader(),
input,
ReaderStatus::Pending,
ReaderStatus::Pending,
ReaderStatus::Pending,
ReaderStatus::Complete(Message::SetWatchMode(SetWatchMode::new(WatchMode::Tagged(65535))))
};
}
#[test]
fn test_reader() {
let input = vec![
/* type */ MessageType::SetWatchMode.into(),
/* mode = all */ 1
];
test_reader! {
Message::reader(),
input,
ReaderStatus::Pending,
ReaderStatus::Pending,
ReaderStatus::Complete(Message::SetWatchMode(SetWatchMode::new(WatchMode::All)))
};
}
} | {
SetWatchModeReader { inner: WatchMode::reader() }
} | identifier_body |
set_watch_mode.rs | use std::io;
use super::super::{WriteTo, WatchMode, WatchModeReader, WriteResult, Reader, ReaderStatus, MessageInner, Message};
#[derive(Debug, Eq, PartialEq, Clone)]
pub struct SetWatchMode {
mode: WatchMode,
}
#[derive(Debug)]
pub struct SetWatchModeReader {
inner: WatchModeReader,
}
impl SetWatchMode {
pub fn | (mode: WatchMode) -> Self {
SetWatchMode { mode }
}
pub fn mode(&self) -> WatchMode {
self.mode
}
pub fn reader() -> SetWatchModeReader {
SetWatchModeReader { inner: WatchMode::reader() }
}
}
impl MessageInner for SetWatchMode {
#[inline]
fn wrap(self) -> Message {
Message::SetWatchMode(self)
}
}
impl Reader<SetWatchMode> for SetWatchModeReader {
fn resume<R>(&mut self, input: &mut R) -> io::Result<ReaderStatus<SetWatchMode>> where R: io::Read {
let status = self.inner.resume(input)?;
Ok(status.map(|mode| SetWatchMode::new(mode)))
}
fn rewind(&mut self) {
self.inner.rewind();
}
}
impl WriteTo for SetWatchMode {
fn write_to<W: io::Write>(&self, target: &mut W) -> WriteResult {
self.mode.write_to(target)
}
}
#[cfg(test)]
mod test {
use super::*;
use super::super::super::{MessageType, WatchMode};
#[test]
fn test_reader_with_tagged() {
let input = vec![
/* type */ MessageType::SetWatchMode.into(),
/* mode = tagged */ 2,
/* tag */ 0, 0, 0, 0, 0, 0, 255, 255
];
test_reader! {
Message::reader(),
input,
ReaderStatus::Pending,
ReaderStatus::Pending,
ReaderStatus::Pending,
ReaderStatus::Complete(Message::SetWatchMode(SetWatchMode::new(WatchMode::Tagged(65535))))
};
}
#[test]
fn test_reader() {
let input = vec![
/* type */ MessageType::SetWatchMode.into(),
/* mode = all */ 1
];
test_reader! {
Message::reader(),
input,
ReaderStatus::Pending,
ReaderStatus::Pending,
ReaderStatus::Complete(Message::SetWatchMode(SetWatchMode::new(WatchMode::All)))
};
}
} | new | identifier_name |
set_watch_mode.rs | use std::io;
use super::super::{WriteTo, WatchMode, WatchModeReader, WriteResult, Reader, ReaderStatus, MessageInner, Message};
#[derive(Debug, Eq, PartialEq, Clone)]
pub struct SetWatchMode {
mode: WatchMode,
}
#[derive(Debug)]
pub struct SetWatchModeReader {
inner: WatchModeReader,
}
impl SetWatchMode {
pub fn new(mode: WatchMode) -> Self {
SetWatchMode { mode }
}
pub fn mode(&self) -> WatchMode {
self.mode
}
pub fn reader() -> SetWatchModeReader {
SetWatchModeReader { inner: WatchMode::reader() }
}
}
impl MessageInner for SetWatchMode {
#[inline]
fn wrap(self) -> Message {
Message::SetWatchMode(self)
}
}
impl Reader<SetWatchMode> for SetWatchModeReader {
fn resume<R>(&mut self, input: &mut R) -> io::Result<ReaderStatus<SetWatchMode>> where R: io::Read {
let status = self.inner.resume(input)?;
Ok(status.map(|mode| SetWatchMode::new(mode)))
}
fn rewind(&mut self) {
self.inner.rewind();
}
}
impl WriteTo for SetWatchMode {
fn write_to<W: io::Write>(&self, target: &mut W) -> WriteResult {
self.mode.write_to(target)
}
}
#[cfg(test)]
mod test {
use super::*;
use super::super::super::{MessageType, WatchMode};
#[test]
fn test_reader_with_tagged() {
let input = vec![
/* type */ MessageType::SetWatchMode.into(),
/* mode = tagged */ 2,
/* tag */ 0, 0, 0, 0, 0, 0, 255, 255
];
test_reader! {
Message::reader(),
input,
ReaderStatus::Pending,
ReaderStatus::Pending,
ReaderStatus::Pending,
ReaderStatus::Complete(Message::SetWatchMode(SetWatchMode::new(WatchMode::Tagged(65535))))
};
}
#[test]
fn test_reader() {
let input = vec![
/* type */ MessageType::SetWatchMode.into(),
/* mode = all */ 1
];
test_reader! {
Message::reader(),
input, | ReaderStatus::Pending,
ReaderStatus::Complete(Message::SetWatchMode(SetWatchMode::new(WatchMode::All)))
};
}
} | ReaderStatus::Pending, | random_line_split |
SchedulerProfiling.js | /**
* Copyright (c) Facebook, Inc. and its affiliates.
*
* This source code is licensed under the MIT license found in the
* LICENSE file in the root directory of this source tree.
*
* @flow
*/
import type {PriorityLevel} from './SchedulerPriorities';
import {enableProfiling} from './SchedulerFeatureFlags';
import {NoPriority} from './SchedulerPriorities';
let runIdCounter: number = 0;
let mainThreadIdCounter: number = 0;
const profilingStateSize = 4;
export const sharedProfilingBuffer = enableProfiling
? // $FlowFixMe Flow doesn't know about SharedArrayBuffer
typeof SharedArrayBuffer === 'function'
? new SharedArrayBuffer(profilingStateSize * Int32Array.BYTES_PER_ELEMENT)
: // $FlowFixMe Flow doesn't know about ArrayBuffer
typeof ArrayBuffer === 'function'
? new ArrayBuffer(profilingStateSize * Int32Array.BYTES_PER_ELEMENT)
: null // Don't crash the init path on IE9
: null;
const profilingState =
enableProfiling && sharedProfilingBuffer !== null
? new Int32Array(sharedProfilingBuffer)
: []; // We can't read this but it helps save bytes for null checks
const PRIORITY = 0;
const CURRENT_TASK_ID = 1;
const CURRENT_RUN_ID = 2;
const QUEUE_SIZE = 3;
if (enableProfiling) {
profilingState[PRIORITY] = NoPriority;
// This is maintained with a counter, because the size of the priority queue
// array might include canceled tasks.
profilingState[QUEUE_SIZE] = 0;
profilingState[CURRENT_TASK_ID] = 0;
}
// Bytes per element is 4
const INITIAL_EVENT_LOG_SIZE = 131072;
const MAX_EVENT_LOG_SIZE = 524288; // Equivalent to 2 megabytes
| let eventLog = null;
let eventLogIndex = 0;
const TaskStartEvent = 1;
const TaskCompleteEvent = 2;
const TaskErrorEvent = 3;
const TaskCancelEvent = 4;
const TaskRunEvent = 5;
const TaskYieldEvent = 6;
const SchedulerSuspendEvent = 7;
const SchedulerResumeEvent = 8;
function logEvent(entries) {
if (eventLog !== null) {
const offset = eventLogIndex;
eventLogIndex += entries.length;
if (eventLogIndex + 1 > eventLogSize) {
eventLogSize *= 2;
if (eventLogSize > MAX_EVENT_LOG_SIZE) {
// Using console['error'] to evade Babel and ESLint
console['error'](
"Scheduler Profiling: Event log exceeded maximum size. Don't " +
'forget to call `stopLoggingProfilingEvents()`.',
);
stopLoggingProfilingEvents();
return;
}
const newEventLog = new Int32Array(eventLogSize * 4);
newEventLog.set(eventLog);
eventLogBuffer = newEventLog.buffer;
eventLog = newEventLog;
}
eventLog.set(entries, offset);
}
}
export function startLoggingProfilingEvents(): void {
eventLogSize = INITIAL_EVENT_LOG_SIZE;
eventLogBuffer = new ArrayBuffer(eventLogSize * 4);
eventLog = new Int32Array(eventLogBuffer);
eventLogIndex = 0;
}
export function stopLoggingProfilingEvents(): ArrayBuffer | null {
const buffer = eventLogBuffer;
eventLogSize = 0;
eventLogBuffer = null;
eventLog = null;
eventLogIndex = 0;
return buffer;
}
export function markTaskStart(
task: {
id: number,
priorityLevel: PriorityLevel,
...
},
ms: number,
) {
if (enableProfiling) {
profilingState[QUEUE_SIZE]++;
if (eventLog !== null) {
// performance.now returns a float, representing milliseconds. When the
// event is logged, it's coerced to an int. Convert to microseconds to
// maintain extra degrees of precision.
logEvent([TaskStartEvent, ms * 1000, task.id, task.priorityLevel]);
}
}
}
export function markTaskCompleted(
task: {
id: number,
priorityLevel: PriorityLevel,
...
},
ms: number,
) {
if (enableProfiling) {
profilingState[PRIORITY] = NoPriority;
profilingState[CURRENT_TASK_ID] = 0;
profilingState[QUEUE_SIZE]--;
if (eventLog !== null) {
logEvent([TaskCompleteEvent, ms * 1000, task.id]);
}
}
}
export function markTaskCanceled(
task: {
id: number,
priorityLevel: PriorityLevel,
...
},
ms: number,
) {
if (enableProfiling) {
profilingState[QUEUE_SIZE]--;
if (eventLog !== null) {
logEvent([TaskCancelEvent, ms * 1000, task.id]);
}
}
}
export function markTaskErrored(
task: {
id: number,
priorityLevel: PriorityLevel,
...
},
ms: number,
) {
if (enableProfiling) {
profilingState[PRIORITY] = NoPriority;
profilingState[CURRENT_TASK_ID] = 0;
profilingState[QUEUE_SIZE]--;
if (eventLog !== null) {
logEvent([TaskErrorEvent, ms * 1000, task.id]);
}
}
}
export function markTaskRun(
task: {
id: number,
priorityLevel: PriorityLevel,
...
},
ms: number,
) {
if (enableProfiling) {
runIdCounter++;
profilingState[PRIORITY] = task.priorityLevel;
profilingState[CURRENT_TASK_ID] = task.id;
profilingState[CURRENT_RUN_ID] = runIdCounter;
if (eventLog !== null) {
logEvent([TaskRunEvent, ms * 1000, task.id, runIdCounter]);
}
}
}
export function markTaskYield(task: {id: number, ...}, ms: number) {
if (enableProfiling) {
profilingState[PRIORITY] = NoPriority;
profilingState[CURRENT_TASK_ID] = 0;
profilingState[CURRENT_RUN_ID] = 0;
if (eventLog !== null) {
logEvent([TaskYieldEvent, ms * 1000, task.id, runIdCounter]);
}
}
}
export function markSchedulerSuspended(ms: number) {
if (enableProfiling) {
mainThreadIdCounter++;
if (eventLog !== null) {
logEvent([SchedulerSuspendEvent, ms * 1000, mainThreadIdCounter]);
}
}
}
export function markSchedulerUnsuspended(ms: number) {
if (enableProfiling) {
if (eventLog !== null) {
logEvent([SchedulerResumeEvent, ms * 1000, mainThreadIdCounter]);
}
}
} | let eventLogSize = 0;
let eventLogBuffer = null; | random_line_split |
SchedulerProfiling.js | /**
* Copyright (c) Facebook, Inc. and its affiliates.
*
* This source code is licensed under the MIT license found in the
* LICENSE file in the root directory of this source tree.
*
* @flow
*/
import type {PriorityLevel} from './SchedulerPriorities';
import {enableProfiling} from './SchedulerFeatureFlags';
import {NoPriority} from './SchedulerPriorities';
let runIdCounter: number = 0;
let mainThreadIdCounter: number = 0;
const profilingStateSize = 4;
export const sharedProfilingBuffer = enableProfiling
? // $FlowFixMe Flow doesn't know about SharedArrayBuffer
typeof SharedArrayBuffer === 'function'
? new SharedArrayBuffer(profilingStateSize * Int32Array.BYTES_PER_ELEMENT)
: // $FlowFixMe Flow doesn't know about ArrayBuffer
typeof ArrayBuffer === 'function'
? new ArrayBuffer(profilingStateSize * Int32Array.BYTES_PER_ELEMENT)
: null // Don't crash the init path on IE9
: null;
const profilingState =
enableProfiling && sharedProfilingBuffer !== null
? new Int32Array(sharedProfilingBuffer)
: []; // We can't read this but it helps save bytes for null checks
const PRIORITY = 0;
const CURRENT_TASK_ID = 1;
const CURRENT_RUN_ID = 2;
const QUEUE_SIZE = 3;
if (enableProfiling) {
profilingState[PRIORITY] = NoPriority;
// This is maintained with a counter, because the size of the priority queue
// array might include canceled tasks.
profilingState[QUEUE_SIZE] = 0;
profilingState[CURRENT_TASK_ID] = 0;
}
// Bytes per element is 4
const INITIAL_EVENT_LOG_SIZE = 131072;
const MAX_EVENT_LOG_SIZE = 524288; // Equivalent to 2 megabytes
let eventLogSize = 0;
let eventLogBuffer = null;
let eventLog = null;
let eventLogIndex = 0;
const TaskStartEvent = 1;
const TaskCompleteEvent = 2;
const TaskErrorEvent = 3;
const TaskCancelEvent = 4;
const TaskRunEvent = 5;
const TaskYieldEvent = 6;
const SchedulerSuspendEvent = 7;
const SchedulerResumeEvent = 8;
function logEvent(entries) {
if (eventLog !== null) {
const offset = eventLogIndex;
eventLogIndex += entries.length;
if (eventLogIndex + 1 > eventLogSize) {
eventLogSize *= 2;
if (eventLogSize > MAX_EVENT_LOG_SIZE) {
// Using console['error'] to evade Babel and ESLint
console['error'](
"Scheduler Profiling: Event log exceeded maximum size. Don't " +
'forget to call `stopLoggingProfilingEvents()`.',
);
stopLoggingProfilingEvents();
return;
}
const newEventLog = new Int32Array(eventLogSize * 4);
newEventLog.set(eventLog);
eventLogBuffer = newEventLog.buffer;
eventLog = newEventLog;
}
eventLog.set(entries, offset);
}
}
export function startLoggingProfilingEvents(): void {
eventLogSize = INITIAL_EVENT_LOG_SIZE;
eventLogBuffer = new ArrayBuffer(eventLogSize * 4);
eventLog = new Int32Array(eventLogBuffer);
eventLogIndex = 0;
}
export function stopLoggingProfilingEvents(): ArrayBuffer | null {
const buffer = eventLogBuffer;
eventLogSize = 0;
eventLogBuffer = null;
eventLog = null;
eventLogIndex = 0;
return buffer;
}
export function markTaskStart(
task: {
id: number,
priorityLevel: PriorityLevel,
...
},
ms: number,
) {
if (enableProfiling) {
profilingState[QUEUE_SIZE]++;
if (eventLog !== null) {
// performance.now returns a float, representing milliseconds. When the
// event is logged, it's coerced to an int. Convert to microseconds to
// maintain extra degrees of precision.
logEvent([TaskStartEvent, ms * 1000, task.id, task.priorityLevel]);
}
}
}
export function markTaskCompleted(
task: {
id: number,
priorityLevel: PriorityLevel,
...
},
ms: number,
) {
if (enableProfiling) {
profilingState[PRIORITY] = NoPriority;
profilingState[CURRENT_TASK_ID] = 0;
profilingState | --;
if (eventLog !== null) {
logEvent([TaskCompleteEvent, ms * 1000, task.id]);
}
}
}
export function markTaskCanceled(
task: {
id: number,
priorityLevel: PriorityLevel,
...
},
ms: number,
) {
if (enableProfiling) {
profilingState[QUEUE_SIZE]--;
if (eventLog !== null) {
logEvent([TaskCancelEvent, ms * 1000, task.id]);
}
}
}
export function markTaskErrored(
task: {
id: number,
priorityLevel: PriorityLevel,
...
},
ms: number,
) {
if (enableProfiling) {
profilingState[PRIORITY] = NoPriority;
profilingState[CURRENT_TASK_ID] = 0;
profilingState[QUEUE_SIZE]--;
if (eventLog !== null) {
logEvent([TaskErrorEvent, ms * 1000, task.id]);
}
}
}
export function markTaskRun(
task: {
id: number,
priorityLevel: PriorityLevel,
...
},
ms: number,
) {
if (enableProfiling) {
runIdCounter++;
profilingState[PRIORITY] = task.priorityLevel;
profilingState[CURRENT_TASK_ID] = task.id;
profilingState[CURRENT_RUN_ID] = runIdCounter;
if (eventLog !== null) {
logEvent([TaskRunEvent, ms * 1000, task.id, runIdCounter]);
}
}
}
export function markTaskYield(task: {id: number, ...}, ms: number) {
if (enableProfiling) {
profilingState[PRIORITY] = NoPriority;
profilingState[CURRENT_TASK_ID] = 0;
profilingState[CURRENT_RUN_ID] = 0;
if (eventLog !== null) {
logEvent([TaskYieldEvent, ms * 1000, task.id, runIdCounter]);
}
}
}
export function markSchedulerSuspended(ms: number) {
if (enableProfiling) {
mainThreadIdCounter++;
if (eventLog !== null) {
logEvent([SchedulerSuspendEvent, ms * 1000, mainThreadIdCounter]);
}
}
}
export function markSchedulerUnsuspended(ms: number) {
if (enableProfiling) {
if (eventLog !== null) {
logEvent([SchedulerResumeEvent, ms * 1000, mainThreadIdCounter]);
}
}
}
| [QUEUE_SIZE] | identifier_name |
SchedulerProfiling.js | /**
* Copyright (c) Facebook, Inc. and its affiliates.
*
* This source code is licensed under the MIT license found in the
* LICENSE file in the root directory of this source tree.
*
* @flow
*/
import type {PriorityLevel} from './SchedulerPriorities';
import {enableProfiling} from './SchedulerFeatureFlags';
import {NoPriority} from './SchedulerPriorities';
let runIdCounter: number = 0;
let mainThreadIdCounter: number = 0;
const profilingStateSize = 4;
export const sharedProfilingBuffer = enableProfiling
? // $FlowFixMe Flow doesn't know about SharedArrayBuffer
typeof SharedArrayBuffer === 'function'
? new SharedArrayBuffer(profilingStateSize * Int32Array.BYTES_PER_ELEMENT)
: // $FlowFixMe Flow doesn't know about ArrayBuffer
typeof ArrayBuffer === 'function'
? new ArrayBuffer(profilingStateSize * Int32Array.BYTES_PER_ELEMENT)
: null // Don't crash the init path on IE9
: null;
const profilingState =
enableProfiling && sharedProfilingBuffer !== null
? new Int32Array(sharedProfilingBuffer)
: []; // We can't read this but it helps save bytes for null checks
const PRIORITY = 0;
const CURRENT_TASK_ID = 1;
const CURRENT_RUN_ID = 2;
const QUEUE_SIZE = 3;
if (enableProfiling) {
profilingState[PRIORITY] = NoPriority;
// This is maintained with a counter, because the size of the priority queue
// array might include canceled tasks.
profilingState[QUEUE_SIZE] = 0;
profilingState[CURRENT_TASK_ID] = 0;
}
// Bytes per element is 4
const INITIAL_EVENT_LOG_SIZE = 131072;
const MAX_EVENT_LOG_SIZE = 524288; // Equivalent to 2 megabytes
let eventLogSize = 0;
let eventLogBuffer = null;
let eventLog = null;
let eventLogIndex = 0;
const TaskStartEvent = 1;
const TaskCompleteEvent = 2;
const TaskErrorEvent = 3;
const TaskCancelEvent = 4;
const TaskRunEvent = 5;
const TaskYieldEvent = 6;
const SchedulerSuspendEvent = 7;
const SchedulerResumeEvent = 8;
function logEvent(entries) {
if (eventLog !== null) {
const offset = eventLogIndex;
eventLogIndex += entries.length;
if (eventLogIndex + 1 > eventLogSize) {
eventLogSize *= 2;
if (eventLogSize > MAX_EVENT_LOG_SIZE) {
// Using console['error'] to evade Babel and ESLint
console['error'](
"Scheduler Profiling: Event log exceeded maximum size. Don't " +
'forget to call `stopLoggingProfilingEvents()`.',
);
stopLoggingProfilingEvents();
return;
}
const newEventLog = new Int32Array(eventLogSize * 4);
newEventLog.set(eventLog);
eventLogBuffer = newEventLog.buffer;
eventLog = newEventLog;
}
eventLog.set(entries, offset);
}
}
export function startLoggingProfilingEvents(): void |
export function stopLoggingProfilingEvents(): ArrayBuffer | null {
const buffer = eventLogBuffer;
eventLogSize = 0;
eventLogBuffer = null;
eventLog = null;
eventLogIndex = 0;
return buffer;
}
export function markTaskStart(
task: {
id: number,
priorityLevel: PriorityLevel,
...
},
ms: number,
) {
if (enableProfiling) {
profilingState[QUEUE_SIZE]++;
if (eventLog !== null) {
// performance.now returns a float, representing milliseconds. When the
// event is logged, it's coerced to an int. Convert to microseconds to
// maintain extra degrees of precision.
logEvent([TaskStartEvent, ms * 1000, task.id, task.priorityLevel]);
}
}
}
export function markTaskCompleted(
task: {
id: number,
priorityLevel: PriorityLevel,
...
},
ms: number,
) {
if (enableProfiling) {
profilingState[PRIORITY] = NoPriority;
profilingState[CURRENT_TASK_ID] = 0;
profilingState[QUEUE_SIZE]--;
if (eventLog !== null) {
logEvent([TaskCompleteEvent, ms * 1000, task.id]);
}
}
}
export function markTaskCanceled(
task: {
id: number,
priorityLevel: PriorityLevel,
...
},
ms: number,
) {
if (enableProfiling) {
profilingState[QUEUE_SIZE]--;
if (eventLog !== null) {
logEvent([TaskCancelEvent, ms * 1000, task.id]);
}
}
}
export function markTaskErrored(
task: {
id: number,
priorityLevel: PriorityLevel,
...
},
ms: number,
) {
if (enableProfiling) {
profilingState[PRIORITY] = NoPriority;
profilingState[CURRENT_TASK_ID] = 0;
profilingState[QUEUE_SIZE]--;
if (eventLog !== null) {
logEvent([TaskErrorEvent, ms * 1000, task.id]);
}
}
}
export function markTaskRun(
task: {
id: number,
priorityLevel: PriorityLevel,
...
},
ms: number,
) {
if (enableProfiling) {
runIdCounter++;
profilingState[PRIORITY] = task.priorityLevel;
profilingState[CURRENT_TASK_ID] = task.id;
profilingState[CURRENT_RUN_ID] = runIdCounter;
if (eventLog !== null) {
logEvent([TaskRunEvent, ms * 1000, task.id, runIdCounter]);
}
}
}
export function markTaskYield(task: {id: number, ...}, ms: number) {
if (enableProfiling) {
profilingState[PRIORITY] = NoPriority;
profilingState[CURRENT_TASK_ID] = 0;
profilingState[CURRENT_RUN_ID] = 0;
if (eventLog !== null) {
logEvent([TaskYieldEvent, ms * 1000, task.id, runIdCounter]);
}
}
}
export function markSchedulerSuspended(ms: number) {
if (enableProfiling) {
mainThreadIdCounter++;
if (eventLog !== null) {
logEvent([SchedulerSuspendEvent, ms * 1000, mainThreadIdCounter]);
}
}
}
export function markSchedulerUnsuspended(ms: number) {
if (enableProfiling) {
if (eventLog !== null) {
logEvent([SchedulerResumeEvent, ms * 1000, mainThreadIdCounter]);
}
}
}
| {
eventLogSize = INITIAL_EVENT_LOG_SIZE;
eventLogBuffer = new ArrayBuffer(eventLogSize * 4);
eventLog = new Int32Array(eventLogBuffer);
eventLogIndex = 0;
} | identifier_body |
SchedulerProfiling.js | /**
* Copyright (c) Facebook, Inc. and its affiliates.
*
* This source code is licensed under the MIT license found in the
* LICENSE file in the root directory of this source tree.
*
* @flow
*/
import type {PriorityLevel} from './SchedulerPriorities';
import {enableProfiling} from './SchedulerFeatureFlags';
import {NoPriority} from './SchedulerPriorities';
let runIdCounter: number = 0;
let mainThreadIdCounter: number = 0;
const profilingStateSize = 4;
export const sharedProfilingBuffer = enableProfiling
? // $FlowFixMe Flow doesn't know about SharedArrayBuffer
typeof SharedArrayBuffer === 'function'
? new SharedArrayBuffer(profilingStateSize * Int32Array.BYTES_PER_ELEMENT)
: // $FlowFixMe Flow doesn't know about ArrayBuffer
typeof ArrayBuffer === 'function'
? new ArrayBuffer(profilingStateSize * Int32Array.BYTES_PER_ELEMENT)
: null // Don't crash the init path on IE9
: null;
const profilingState =
enableProfiling && sharedProfilingBuffer !== null
? new Int32Array(sharedProfilingBuffer)
: []; // We can't read this but it helps save bytes for null checks
const PRIORITY = 0;
const CURRENT_TASK_ID = 1;
const CURRENT_RUN_ID = 2;
const QUEUE_SIZE = 3;
if (enableProfiling) {
profilingState[PRIORITY] = NoPriority;
// This is maintained with a counter, because the size of the priority queue
// array might include canceled tasks.
profilingState[QUEUE_SIZE] = 0;
profilingState[CURRENT_TASK_ID] = 0;
}
// Bytes per element is 4
const INITIAL_EVENT_LOG_SIZE = 131072;
const MAX_EVENT_LOG_SIZE = 524288; // Equivalent to 2 megabytes
let eventLogSize = 0;
let eventLogBuffer = null;
let eventLog = null;
let eventLogIndex = 0;
const TaskStartEvent = 1;
const TaskCompleteEvent = 2;
const TaskErrorEvent = 3;
const TaskCancelEvent = 4;
const TaskRunEvent = 5;
const TaskYieldEvent = 6;
const SchedulerSuspendEvent = 7;
const SchedulerResumeEvent = 8;
function logEvent(entries) {
if (eventLog !== null) {
const offset = eventLogIndex;
eventLogIndex += entries.length;
if (eventLogIndex + 1 > eventLogSize) {
eventLogSize *= 2;
if (eventLogSize > MAX_EVENT_LOG_SIZE) {
// Using console['error'] to evade Babel and ESLint
console['error'](
"Scheduler Profiling: Event log exceeded maximum size. Don't " +
'forget to call `stopLoggingProfilingEvents()`.',
);
stopLoggingProfilingEvents();
return;
}
const newEventLog = new Int32Array(eventLogSize * 4);
newEventLog.set(eventLog);
eventLogBuffer = newEventLog.buffer;
eventLog = newEventLog;
}
eventLog.set(entries, offset);
}
}
export function startLoggingProfilingEvents(): void {
eventLogSize = INITIAL_EVENT_LOG_SIZE;
eventLogBuffer = new ArrayBuffer(eventLogSize * 4);
eventLog = new Int32Array(eventLogBuffer);
eventLogIndex = 0;
}
export function stopLoggingProfilingEvents(): ArrayBuffer | null {
const buffer = eventLogBuffer;
eventLogSize = 0;
eventLogBuffer = null;
eventLog = null;
eventLogIndex = 0;
return buffer;
}
export function markTaskStart(
task: {
id: number,
priorityLevel: PriorityLevel,
...
},
ms: number,
) {
if (enableProfiling) {
profilingState[QUEUE_SIZE]++;
if (eventLog !== null) {
// performance.now returns a float, representing milliseconds. When the
// event is logged, it's coerced to an int. Convert to microseconds to
// maintain extra degrees of precision.
logEvent([TaskStartEvent, ms * 1000, task.id, task.priorityLevel]);
}
}
}
export function markTaskCompleted(
task: {
id: number,
priorityLevel: PriorityLevel,
...
},
ms: number,
) {
if (enableProfiling) {
profilingState[PRIORITY] = NoPriority;
profilingState[CURRENT_TASK_ID] = 0;
profilingState[QUEUE_SIZE]--;
if (eventLog !== null) {
logEvent([TaskCompleteEvent, ms * 1000, task.id]);
}
}
}
export function markTaskCanceled(
task: {
id: number,
priorityLevel: PriorityLevel,
...
},
ms: number,
) {
if (enableProfiling) {
profilingState[QUEUE_SIZE]--;
if (eventLog !== null) {
logEvent([TaskCancelEvent, ms * 1000, task.id]);
}
}
}
export function markTaskErrored(
task: {
id: number,
priorityLevel: PriorityLevel,
...
},
ms: number,
) {
if (enableProfiling) {
profilingState[PRIORITY] = NoPriority;
profilingState[CURRENT_TASK_ID] = 0;
profilingState[QUEUE_SIZE]--;
if (eventLog !== null) {
logEvent([TaskErrorEvent, ms * 1000, task.id]);
}
}
}
export function markTaskRun(
task: {
id: number,
priorityLevel: PriorityLevel,
...
},
ms: number,
) {
if (enableProfiling) {
runIdCounter++;
profilingState[PRIORITY] = task.priorityLevel;
profilingState[CURRENT_TASK_ID] = task.id;
profilingState[CURRENT_RUN_ID] = runIdCounter;
if (eventLog !== null) {
logEvent([TaskRunEvent, ms * 1000, task.id, runIdCounter]);
}
}
}
export function markTaskYield(task: {id: number, ...}, ms: number) {
if (enableProfiling) {
profilingState[PRIORITY] = NoPriority;
profilingState[CURRENT_TASK_ID] = 0;
profilingState[CURRENT_RUN_ID] = 0;
if (eventLog !== null) {
logEvent([TaskYieldEvent, ms * 1000, task.id, runIdCounter]);
}
}
}
export function markSchedulerSuspended(ms: number) {
if (enableProfiling) {
mainThreadIdCounter++;
if (eventLog !== null) |
}
}
export function markSchedulerUnsuspended(ms: number) {
if (enableProfiling) {
if (eventLog !== null) {
logEvent([SchedulerResumeEvent, ms * 1000, mainThreadIdCounter]);
}
}
}
| {
logEvent([SchedulerSuspendEvent, ms * 1000, mainThreadIdCounter]);
} | conditional_block |
opportunity_kraken.py | from exchanges import helpers
from exchanges import kraken
from decimal import Decimal
### Kraken opportunities
#### ARBITRAGE OPPORTUNITY 1
def | ():
sellLTCbuyEUR = kraken.get_current_bid_LTCEUR()
sellEURbuyXBT = kraken.get_current_ask_XBTEUR()
sellXBTbuyLTC = kraken.get_current_ask_XBTLTC()
opport = 1-((sellLTCbuyEUR/sellEURbuyBTX)*sellXBTbuyLTC)
return Decimal(opport)
def opportunity_2():
sellEURbuyLTC = kraken.get_current_ask_LTCEUR()
sellLTCbuyXBT = kraken.get_current_ask_XBTLTC()
sellXBTbuyEUR = kraken.get_current_bid_XBTEUR()
opport = 1-(((1/sellEURbuyLTC)/sellLTCbuyXBT)*sellXBTbuyEUR)
return Decimal(opport) | opportunity_1 | identifier_name |
opportunity_kraken.py | from exchanges import helpers
from exchanges import kraken
from decimal import Decimal
### Kraken opportunities
#### ARBITRAGE OPPORTUNITY 1 | opport = 1-((sellLTCbuyEUR/sellEURbuyBTX)*sellXBTbuyLTC)
return Decimal(opport)
def opportunity_2():
sellEURbuyLTC = kraken.get_current_ask_LTCEUR()
sellLTCbuyXBT = kraken.get_current_ask_XBTLTC()
sellXBTbuyEUR = kraken.get_current_bid_XBTEUR()
opport = 1-(((1/sellEURbuyLTC)/sellLTCbuyXBT)*sellXBTbuyEUR)
return Decimal(opport) | def opportunity_1():
sellLTCbuyEUR = kraken.get_current_bid_LTCEUR()
sellEURbuyXBT = kraken.get_current_ask_XBTEUR()
sellXBTbuyLTC = kraken.get_current_ask_XBTLTC() | random_line_split |
opportunity_kraken.py | from exchanges import helpers
from exchanges import kraken
from decimal import Decimal
### Kraken opportunities
#### ARBITRAGE OPPORTUNITY 1
def opportunity_1():
sellLTCbuyEUR = kraken.get_current_bid_LTCEUR()
sellEURbuyXBT = kraken.get_current_ask_XBTEUR()
sellXBTbuyLTC = kraken.get_current_ask_XBTLTC()
opport = 1-((sellLTCbuyEUR/sellEURbuyBTX)*sellXBTbuyLTC)
return Decimal(opport)
def opportunity_2():
| sellEURbuyLTC = kraken.get_current_ask_LTCEUR()
sellLTCbuyXBT = kraken.get_current_ask_XBTLTC()
sellXBTbuyEUR = kraken.get_current_bid_XBTEUR()
opport = 1-(((1/sellEURbuyLTC)/sellLTCbuyXBT)*sellXBTbuyEUR)
return Decimal(opport) | identifier_body |
|
lib.fluidContent.ts | # Default configuration for content elements which are using FLUIDTEMPLATE directly
lib.fluidContent >
lib.fluidContent = FLUIDTEMPLATE
lib.fluidContent {
templateName = Default
templateRootPaths {
0 = EXT:fluid_styled_content/Resources/Private/Templates/
10 = {$styles.templates.templateRootPath}
}
partialRootPaths {
0 = EXT:fluid_styled_content/Resources/Private/Partials/
10 = {$styles.templates.partialRootPath}
}
layoutRootPaths {
0 = EXT:fluid_styled_content/Resources/Private/Layouts/
10 = {$styles.templates.layoutRootPath} | media {
popup {
bodyTag = <body style="margin:0; background:#fff;">
wrap = <a href="javascript:close();"> | </a>
width = {$styles.content.textmedia.linkWrap.width}
height = {$styles.content.textmedia.linkWrap.height}
JSwindow = 1
JSwindow {
newWindow = {$styles.content.textmedia.linkWrap.newWindow}
if.isFalse = {$styles.content.textmedia.linkWrap.lightboxEnabled}
}
directImageLink = {$styles.content.textmedia.linkWrap.lightboxEnabled}
linkParams.ATagParams.dataWrap = class="{$styles.content.textmedia.linkWrap.lightboxCssClass}" rel="{$styles.content.textmedia.linkWrap.lightboxRelAttribute}"
}
}
}
} | }
settings {
defaultHeaderType = {$styles.content.defaultHeaderType}
| random_line_split |
dev-server.ts | import express, { Router } from 'express';
import { Builder, logConfig, Options } from '@storybook/core-common';
import { getMiddleware } from './utils/middleware';
import { getServerAddresses } from './utils/server-address';
import { getServer } from './utils/server-init';
import { useStatics } from './utils/server-statics';
import * as managerBuilder from './manager/builder';
import { openInBrowser } from './utils/open-in-browser';
import { getPreviewBuilder } from './utils/get-preview-builder';
// @ts-ignore
export const router: Router = new Router();
export async function storybookDevServer(options: Options) |
const { port, host } = options;
const proto = options.https ? 'https' : 'http';
const { address, networkAddress } = getServerAddresses(port, host, proto);
await new Promise<void>((resolve, reject) => {
// FIXME: Following line doesn't match TypeScript signature at all 🤔
// @ts-ignore
server.listen({ port, host }, (error: Error) => (error ? reject(error) : resolve()));
});
const previewBuilder: Builder<unknown, unknown> = await getPreviewBuilder(options.configDir);
if (options.debugWebpack) {
logConfig('Preview webpack config', await previewBuilder.getConfig(options));
logConfig('Manager webpack config', await managerBuilder.getConfig(options));
}
const preview = options.ignorePreview
? Promise.resolve()
: previewBuilder.start({
startTime,
options,
router,
});
const manager = managerBuilder.start({
startTime,
options,
router,
});
const [previewResult, managerResult] = await Promise.all([
preview,
manager
// TODO #13083 Restore this when compiling the preview is fast enough
// .then((result) => {
// if (!options.ci && !options.smokeTest) openInBrowser(address);
// return result;
// })
.catch(previewBuilder.bail),
]);
// TODO #13083 Remove this when compiling the preview is fast enough
if (!options.ci && !options.smokeTest) openInBrowser(host ? networkAddress : address);
return { previewResult, managerResult, address, networkAddress };
}
| {
const startTime = process.hrtime();
const app = express();
const server = await getServer(app, options);
if (typeof options.extendServer === 'function') {
options.extendServer(server);
}
app.use((req, res, next) => {
res.header('Access-Control-Allow-Origin', '*');
res.header('Access-Control-Allow-Headers', 'Origin, X-Requested-With, Content-Type, Accept');
next();
});
// User's own static files
await useStatics(router, options);
getMiddleware(options.configDir)(router);
app.use(router); | identifier_body |
dev-server.ts | import express, { Router } from 'express';
import { Builder, logConfig, Options } from '@storybook/core-common';
import { getMiddleware } from './utils/middleware';
import { getServerAddresses } from './utils/server-address';
import { getServer } from './utils/server-init';
import { useStatics } from './utils/server-statics';
import * as managerBuilder from './manager/builder';
import { openInBrowser } from './utils/open-in-browser';
import { getPreviewBuilder } from './utils/get-preview-builder';
// @ts-ignore
export const router: Router = new Router();
export async function storybookDevServer(options: Options) {
const startTime = process.hrtime();
const app = express();
const server = await getServer(app, options);
if (typeof options.extendServer === 'function') {
options.extendServer(server);
}
app.use((req, res, next) => {
res.header('Access-Control-Allow-Origin', '*');
res.header('Access-Control-Allow-Headers', 'Origin, X-Requested-With, Content-Type, Accept');
next();
});
// User's own static files
await useStatics(router, options);
getMiddleware(options.configDir)(router);
app.use(router);
const { port, host } = options;
const proto = options.https ? 'https' : 'http';
const { address, networkAddress } = getServerAddresses(port, host, proto);
await new Promise<void>((resolve, reject) => {
// FIXME: Following line doesn't match TypeScript signature at all 🤔
// @ts-ignore
server.listen({ port, host }, (error: Error) => (error ? reject(error) : resolve()));
});
const previewBuilder: Builder<unknown, unknown> = await getPreviewBuilder(options.configDir);
if (options.debugWebpack) {
logConfig('Preview webpack config', await previewBuilder.getConfig(options));
logConfig('Manager webpack config', await managerBuilder.getConfig(options));
}
const preview = options.ignorePreview
? Promise.resolve()
: previewBuilder.start({
startTime,
options,
router,
});
const manager = managerBuilder.start({
startTime,
options,
router,
});
const [previewResult, managerResult] = await Promise.all([
preview,
manager
// TODO #13083 Restore this when compiling the preview is fast enough
// .then((result) => {
// if (!options.ci && !options.smokeTest) openInBrowser(address);
// return result;
// })
.catch(previewBuilder.bail),
]);
// TODO #13083 Remove this when compiling the preview is fast enough
if (!options.ci && !options.smokeTest) openInBrowser(host ? networkAddress : address); | } |
return { previewResult, managerResult, address, networkAddress }; | random_line_split |
dev-server.ts | import express, { Router } from 'express';
import { Builder, logConfig, Options } from '@storybook/core-common';
import { getMiddleware } from './utils/middleware';
import { getServerAddresses } from './utils/server-address';
import { getServer } from './utils/server-init';
import { useStatics } from './utils/server-statics';
import * as managerBuilder from './manager/builder';
import { openInBrowser } from './utils/open-in-browser';
import { getPreviewBuilder } from './utils/get-preview-builder';
// @ts-ignore
export const router: Router = new Router();
export async function | (options: Options) {
const startTime = process.hrtime();
const app = express();
const server = await getServer(app, options);
if (typeof options.extendServer === 'function') {
options.extendServer(server);
}
app.use((req, res, next) => {
res.header('Access-Control-Allow-Origin', '*');
res.header('Access-Control-Allow-Headers', 'Origin, X-Requested-With, Content-Type, Accept');
next();
});
// User's own static files
await useStatics(router, options);
getMiddleware(options.configDir)(router);
app.use(router);
const { port, host } = options;
const proto = options.https ? 'https' : 'http';
const { address, networkAddress } = getServerAddresses(port, host, proto);
await new Promise<void>((resolve, reject) => {
// FIXME: Following line doesn't match TypeScript signature at all 🤔
// @ts-ignore
server.listen({ port, host }, (error: Error) => (error ? reject(error) : resolve()));
});
const previewBuilder: Builder<unknown, unknown> = await getPreviewBuilder(options.configDir);
if (options.debugWebpack) {
logConfig('Preview webpack config', await previewBuilder.getConfig(options));
logConfig('Manager webpack config', await managerBuilder.getConfig(options));
}
const preview = options.ignorePreview
? Promise.resolve()
: previewBuilder.start({
startTime,
options,
router,
});
const manager = managerBuilder.start({
startTime,
options,
router,
});
const [previewResult, managerResult] = await Promise.all([
preview,
manager
// TODO #13083 Restore this when compiling the preview is fast enough
// .then((result) => {
// if (!options.ci && !options.smokeTest) openInBrowser(address);
// return result;
// })
.catch(previewBuilder.bail),
]);
// TODO #13083 Remove this when compiling the preview is fast enough
if (!options.ci && !options.smokeTest) openInBrowser(host ? networkAddress : address);
return { previewResult, managerResult, address, networkAddress };
}
| storybookDevServer | identifier_name |
dev-server.ts | import express, { Router } from 'express';
import { Builder, logConfig, Options } from '@storybook/core-common';
import { getMiddleware } from './utils/middleware';
import { getServerAddresses } from './utils/server-address';
import { getServer } from './utils/server-init';
import { useStatics } from './utils/server-statics';
import * as managerBuilder from './manager/builder';
import { openInBrowser } from './utils/open-in-browser';
import { getPreviewBuilder } from './utils/get-preview-builder';
// @ts-ignore
export const router: Router = new Router();
export async function storybookDevServer(options: Options) {
const startTime = process.hrtime();
const app = express();
const server = await getServer(app, options);
if (typeof options.extendServer === 'function') {
options.extendServer(server);
}
app.use((req, res, next) => {
res.header('Access-Control-Allow-Origin', '*');
res.header('Access-Control-Allow-Headers', 'Origin, X-Requested-With, Content-Type, Accept');
next();
});
// User's own static files
await useStatics(router, options);
getMiddleware(options.configDir)(router);
app.use(router);
const { port, host } = options;
const proto = options.https ? 'https' : 'http';
const { address, networkAddress } = getServerAddresses(port, host, proto);
await new Promise<void>((resolve, reject) => {
// FIXME: Following line doesn't match TypeScript signature at all 🤔
// @ts-ignore
server.listen({ port, host }, (error: Error) => (error ? reject(error) : resolve()));
});
const previewBuilder: Builder<unknown, unknown> = await getPreviewBuilder(options.configDir);
if (options.debugWebpack) {
| const preview = options.ignorePreview
? Promise.resolve()
: previewBuilder.start({
startTime,
options,
router,
});
const manager = managerBuilder.start({
startTime,
options,
router,
});
const [previewResult, managerResult] = await Promise.all([
preview,
manager
// TODO #13083 Restore this when compiling the preview is fast enough
// .then((result) => {
// if (!options.ci && !options.smokeTest) openInBrowser(address);
// return result;
// })
.catch(previewBuilder.bail),
]);
// TODO #13083 Remove this when compiling the preview is fast enough
if (!options.ci && !options.smokeTest) openInBrowser(host ? networkAddress : address);
return { previewResult, managerResult, address, networkAddress };
}
| logConfig('Preview webpack config', await previewBuilder.getConfig(options));
logConfig('Manager webpack config', await managerBuilder.getConfig(options));
}
| conditional_block |
005_tle_error.py | class Solution(object):
def longestPalindrome(self, s):
max_len = 0
max_str = ''
if len(s) <= 2:
return s
for i, ch in enumerate(s):
delta = 1
count = 0
# center is ch
while (i - delta) >= 0 and (i + delta) < len(s):
if s[i-delta] != s[i+delta]:
break
count += 1
delta += 1
if count * 2 + 1 > max_len:
max_len = count * 2 + 1 | # center is ch right
delta = 0.5
count = 0
j = i + 0.5
while (j - delta) >= 0 and (j + delta) < len(s):
if s[int(j - delta)] != s[int(j + delta)]:
break
count += 1
delta += 1
if count * 2 > max_len:
max_len = count * 2
max_str = s[i-count+1:i+count+1]
return max_str
def test(self):
assert self.longestPalindrome('a') == 'a'
assert self.longestPalindrome('abcba') == 'abcba'
assert self.longestPalindrome('eabcbae') == 'eabcbae'
assert self.longestPalindrome('abba') == 'abba'
assert self.longestPalindrome('abbc') == 'bb'
assert self.longestPalindrome('dbabba') == 'abba'
assert self.longestPalindrome('decababace') == 'ecababace'
assert self.longestPalindrome('decababaceehgagbgnag') == 'ecababace'
if __name__ == '__main__':
s = Solution()
s.test() | max_str = s[i-count:i+1+count]
| random_line_split |
005_tle_error.py | class Solution(object):
def longestPalindrome(self, s):
max_len = 0
max_str = ''
if len(s) <= 2:
return s
for i, ch in enumerate(s):
delta = 1
count = 0
# center is ch
while (i - delta) >= 0 and (i + delta) < len(s):
if s[i-delta] != s[i+delta]:
break
count += 1
delta += 1
if count * 2 + 1 > max_len:
max_len = count * 2 + 1
max_str = s[i-count:i+1+count]
# center is ch right
delta = 0.5
count = 0
j = i + 0.5
while (j - delta) >= 0 and (j + delta) < len(s):
if s[int(j - delta)] != s[int(j + delta)]:
break
count += 1
delta += 1
if count * 2 > max_len:
max_len = count * 2
max_str = s[i-count+1:i+count+1]
return max_str
def test(self):
assert self.longestPalindrome('a') == 'a'
assert self.longestPalindrome('abcba') == 'abcba'
assert self.longestPalindrome('eabcbae') == 'eabcbae'
assert self.longestPalindrome('abba') == 'abba'
assert self.longestPalindrome('abbc') == 'bb'
assert self.longestPalindrome('dbabba') == 'abba'
assert self.longestPalindrome('decababace') == 'ecababace'
assert self.longestPalindrome('decababaceehgagbgnag') == 'ecababace'
if __name__ == '__main__':
| s = Solution()
s.test() | conditional_block |
|
005_tle_error.py | class Solution(object):
def longestPalindrome(self, s):
max_len = 0
max_str = ''
if len(s) <= 2:
return s
for i, ch in enumerate(s):
delta = 1
count = 0
# center is ch
while (i - delta) >= 0 and (i + delta) < len(s):
if s[i-delta] != s[i+delta]:
break
count += 1
delta += 1
if count * 2 + 1 > max_len:
max_len = count * 2 + 1
max_str = s[i-count:i+1+count]
# center is ch right
delta = 0.5
count = 0
j = i + 0.5
while (j - delta) >= 0 and (j + delta) < len(s):
if s[int(j - delta)] != s[int(j + delta)]:
break
count += 1
delta += 1
if count * 2 > max_len:
max_len = count * 2
max_str = s[i-count+1:i+count+1]
return max_str
def | (self):
assert self.longestPalindrome('a') == 'a'
assert self.longestPalindrome('abcba') == 'abcba'
assert self.longestPalindrome('eabcbae') == 'eabcbae'
assert self.longestPalindrome('abba') == 'abba'
assert self.longestPalindrome('abbc') == 'bb'
assert self.longestPalindrome('dbabba') == 'abba'
assert self.longestPalindrome('decababace') == 'ecababace'
assert self.longestPalindrome('decababaceehgagbgnag') == 'ecababace'
if __name__ == '__main__':
s = Solution()
s.test()
| test | identifier_name |
005_tle_error.py | class Solution(object):
def longestPalindrome(self, s):
max_len = 0
max_str = ''
if len(s) <= 2:
return s
for i, ch in enumerate(s):
delta = 1
count = 0
# center is ch
while (i - delta) >= 0 and (i + delta) < len(s):
if s[i-delta] != s[i+delta]:
break
count += 1
delta += 1
if count * 2 + 1 > max_len:
max_len = count * 2 + 1
max_str = s[i-count:i+1+count]
# center is ch right
delta = 0.5
count = 0
j = i + 0.5
while (j - delta) >= 0 and (j + delta) < len(s):
if s[int(j - delta)] != s[int(j + delta)]:
break
count += 1
delta += 1
if count * 2 > max_len:
max_len = count * 2
max_str = s[i-count+1:i+count+1]
return max_str
def test(self):
|
if __name__ == '__main__':
s = Solution()
s.test()
| assert self.longestPalindrome('a') == 'a'
assert self.longestPalindrome('abcba') == 'abcba'
assert self.longestPalindrome('eabcbae') == 'eabcbae'
assert self.longestPalindrome('abba') == 'abba'
assert self.longestPalindrome('abbc') == 'bb'
assert self.longestPalindrome('dbabba') == 'abba'
assert self.longestPalindrome('decababace') == 'ecababace'
assert self.longestPalindrome('decababaceehgagbgnag') == 'ecababace' | identifier_body |
settings.py | """
Django settings for paulpruitt_net project.
For more information on this file, see
https://docs.djangoproject.com/en/1.6/topics/settings/
For the full list of settings and their values, see
https://docs.djangoproject.com/en/1.6/ref/settings/
"""
# Build paths inside the project like this: os.path.join(BASE_DIR, ...)
import os
from secrets import SECRET_KEY, DB_USER, DB_PASSWORD
BASE_DIR = os.path.dirname(os.path.realpath(__file__))
# Quick-start development settings - unsuitable for production
# See https://docs.djangoproject.com/en/1.6/howto/deployment/checklist/
# SECURITY WARNING: don't run with debug turned on in production!
DEBUG = True
TEMPLATE_DEBUG = True
ALLOWED_HOSTS = []
# Application definition
INSTALLED_APPS = (
'django.contrib.admin',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.sessions',
'django.contrib.messages',
'django.contrib.staticfiles',
'taggit',
'pblog'
)
MIDDLEWARE_CLASSES = (
'django.contrib.sessions.middleware.SessionMiddleware',
'django.middleware.common.CommonMiddleware',
'django.middleware.csrf.CsrfViewMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'django.middleware.clickjacking.XFrameOptionsMiddleware',
)
ROOT_URLCONF = 'paulpruitt_net.urls'
WSGI_APPLICATION = 'paulpruitt_net.wsgi.application'
TEMPLATE_DIRS = (
os.path.join(BASE_DIR, 'templates'),
)
# Database
# https://docs.djangoproject.com/en/1.6/ref/settings/#databases
DATABASES = {
'default': {
'ENGINE' : 'django.db.backends.postgresql_psycopg2',
'NAME' : 'site',
'USER' : DB_USER,
'PASSWORD': DB_PASSWORD,
'HOST' : '127.0.0.1',
'PORT' : '',
} | # Internationalization
# https://docs.djangoproject.com/en/1.6/topics/i18n/
LANGUAGE_CODE = 'en-us'
TIME_ZONE = 'UTC'
USE_I18N = True
USE_L10N = True
USE_TZ = True
# Static files (CSS, JavaScript, Images)
# https://docs.djangoproject.com/en/1.6/howto/static-files/
STATIC_URL = '/static/'
STATIC_ROOT = '/srv/www/site/static'
STATICFILES_DIRS = (
os.path.join(BASE_DIR, "static"),
)
# Do not allow IFrames
X_FRAME_OPTIONS = 'DENY' | }
| random_line_split |
violations.py | ends=True)
offset = self.definition.start # type: ignore
lines_stripped = list(
reversed(list(dropwhile(is_blank, reversed(lines))))
)
numbers_width = len(str(offset + len(lines_stripped)))
line_format = f'{{:{numbers_width}}}:{{}}'
for n, line in enumerate(lines_stripped):
if line:
line = ' ' + line
source += line_format.format(n + offset, line)
if n > 5:
source += ' ...\n'
break
return source
def __str__(self) -> str:
if self.explanation:
self.explanation = '\n'.join(
l for l in self.explanation.split('\n') if not is_blank(l)
)
template = '{filename}:{line} {definition}:\n {message}'
if self.source and self.explain:
template += '\n\n{explanation}\n\n{lines}\n'
elif self.source and not self.explain:
template += '\n\n{lines}\n'
elif self.explain and not self.source:
template += '\n\n{explanation}\n\n'
return template.format(
**{
name: getattr(self, name)
for name in [
'filename',
'line',
'definition',
'message',
'explanation',
'lines',
]
}
)
def __repr__(self) -> str:
return str(self)
def __lt__(self, other: 'Error') -> bool:
return (self.filename, self.line) < (other.filename, other.line)
class ErrorRegistry:
"""A registry of all error codes, divided to groups."""
groups = [] # type: ignore
class ErrorGroup:
"""A group of similarly themed errors."""
def __init__(self, prefix: str, name: str) -> None:
"""Initialize the object.
`Prefix` should be the common prefix for errors in this group,
e.g., "D1".
`name` is the name of the group (its subject).
"""
self.prefix = prefix
self.name = name
self.errors = [] # type: List[ErrorParams]
def create_error(
self,
error_code: str,
error_desc: str,
error_context: Optional[str] = None,
) -> Callable[[Iterable[str]], Error]:
"""Create an error, register it to this group and return it."""
# TODO: check prefix
error_params = ErrorParams(error_code, error_desc, error_context)
factory = partial(Error, *error_params)
self.errors.append(error_params)
return factory
@classmethod
def create_group(cls, prefix: str, name: str) -> ErrorGroup:
"""Create a new error group and return it."""
group = cls.ErrorGroup(prefix, name)
cls.groups.append(group)
return group
@classmethod
def get_error_codes(cls) -> Iterable[str]:
|
@classmethod
def to_rst(cls) -> str:
"""Output the registry as reStructuredText, for documentation."""
max_len = max(
len(error.short_desc)
for group in cls.groups
for error in group.errors
)
sep_line = '+' + 6 * '-' + '+' + '-' * (max_len + 2) + '+\n'
blank_line = '|' + (max_len + 9) * ' ' + '|\n'
table = ''
for group in cls.groups:
table += sep_line
table += blank_line
table += '|' + f'**{group.name}**'.center(max_len + 9) + '|\n'
table += blank_line
for error in group.errors:
table += sep_line
table += (
'|'
+ error.code.center(6)
+ '| '
+ error.short_desc.ljust(max_len + 1)
+ '|\n'
)
table += sep_line
return table
D1xx = ErrorRegistry.create_group('D1', 'Missing Docstrings')
D100 = D1xx.create_error(
'D100',
'Missing docstring in public module',
)
D101 = D1xx.create_error(
'D101',
'Missing docstring in public class',
)
D102 = D1xx.create_error(
'D102',
'Missing docstring in public method',
)
D103 = D1xx.create_error(
'D103',
'Missing docstring in public function',
)
D104 = D1xx.create_error(
'D104',
'Missing docstring in public package',
)
D105 = D1xx.create_error(
'D105',
'Missing docstring in magic method',
)
D106 = D1xx.create_error(
'D106',
'Missing docstring in public nested class',
)
D107 = D1xx.create_error(
'D107',
'Missing docstring in __init__',
)
D2xx = ErrorRegistry.create_group('D2', 'Whitespace Issues')
D200 = D2xx.create_error(
'D200',
'One-line docstring should fit on one line ' 'with quotes',
'found {0}',
)
D201 = D2xx.create_error(
'D201',
'No blank lines allowed before function docstring',
'found {0}',
)
D202 = D2xx.create_error(
'D202',
'No blank lines allowed after function docstring',
'found {0}',
)
D203 = D2xx.create_error(
'D203',
'1 blank line required before class docstring',
'found {0}',
)
D204 = D2xx.create_error(
'D204',
'1 blank line required after class docstring',
'found {0}',
)
D205 = D2xx.create_error(
'D205',
'1 blank line required between summary line and description',
'found {0}',
)
D206 = D2xx.create_error(
'D206',
'Docstring should be indented with spaces, not tabs',
)
D207 = D2xx.create_error(
'D207',
'Docstring is under-indented',
)
D208 = D2xx.create_error(
'D208',
'Docstring is over-indented',
)
D209 = D2xx.create_error(
'D209',
'Multi-line docstring closing quotes should be on a separate line',
)
D210 = D2xx.create_error(
'D210',
'No whitespaces allowed surrounding docstring text',
)
D211 = D2xx.create_error(
'D211',
'No blank lines allowed before class docstring',
'found {0}',
)
D212 = D2xx.create_error(
'D212',
'Multi-line docstring summary should start at the first line',
)
D213 = D2xx.create_error(
'D213',
'Multi-line docstring summary should start at the second line',
)
D214 = D2xx.create_error(
'D214',
'Section is over-indented',
'{0!r}',
)
D215 = D2xx.create_error(
'D215',
'Section underline is over-indented',
'in section {0!r}',
)
D3xx = ErrorRegistry.create_group('D3', 'Quotes Issues')
D300 = D3xx.create_error(
'D300',
'Use """triple double quotes"""',
'found {0}-quotes',
)
D301 = D3xx.create_error(
'D301',
'Use r""" if any backslashes in a docstring',
)
D302 = D3xx.create_error(
'D302',
'Deprecated: Use u""" for Unicode docstrings',
)
D4xx = ErrorRegistry.create_group('D4', 'Docstring Content Issues')
D400 = D4xx.create_error(
'D400',
'First line should end with a period',
'not {0!r}',
)
D401 = D4xx.create_error(
'D401',
'First line should be in imperative mood',
"perhaps '{0}', not '{1}'",
)
D401b = D4xx.create_error(
'D401',
'First line should be in imperative mood; try rephrasing',
"found '{0}'",
)
D402 = D4xx.create_error(
'D402',
'First line should not be the function\'s "signature"',
)
D403 = D4xx.create_error(
'D403',
'First word of the first line should be properly capitalized',
'{0!r}, not {1!r}',
)
D404 = D4xx.create_error(
| """Yield all registered codes."""
for group in cls.groups:
for error in group.errors:
yield error.code | identifier_body |
violations.py | ends=True)
offset = self.definition.start # type: ignore
lines_stripped = list(
reversed(list(dropwhile(is_blank, reversed(lines))))
)
numbers_width = len(str(offset + len(lines_stripped)))
line_format = f'{{:{numbers_width}}}:{{}}'
for n, line in enumerate(lines_stripped):
if line:
line = ' ' + line
source += line_format.format(n + offset, line)
if n > 5:
source += ' ...\n'
break
return source
def __str__(self) -> str:
if self.explanation:
self.explanation = '\n'.join(
l for l in self.explanation.split('\n') if not is_blank(l)
)
template = '{filename}:{line} {definition}:\n {message}'
if self.source and self.explain:
template += '\n\n{explanation}\n\n{lines}\n'
elif self.source and not self.explain:
template += '\n\n{lines}\n'
elif self.explain and not self.source:
template += '\n\n{explanation}\n\n'
return template.format(
**{
name: getattr(self, name)
for name in [
'filename',
'line',
'definition',
'message',
'explanation',
'lines',
]
}
)
def __repr__(self) -> str:
return str(self)
def __lt__(self, other: 'Error') -> bool:
return (self.filename, self.line) < (other.filename, other.line)
class ErrorRegistry:
"""A registry of all error codes, divided to groups."""
groups = [] # type: ignore
class ErrorGroup:
"""A group of similarly themed errors."""
def __init__(self, prefix: str, name: str) -> None:
"""Initialize the object.
`Prefix` should be the common prefix for errors in this group,
e.g., "D1".
`name` is the name of the group (its subject).
"""
self.prefix = prefix
self.name = name
self.errors = [] # type: List[ErrorParams]
def create_error(
self,
error_code: str,
error_desc: str,
error_context: Optional[str] = None,
) -> Callable[[Iterable[str]], Error]:
"""Create an error, register it to this group and return it."""
# TODO: check prefix
error_params = ErrorParams(error_code, error_desc, error_context)
factory = partial(Error, *error_params)
self.errors.append(error_params)
return factory
@classmethod
def create_group(cls, prefix: str, name: str) -> ErrorGroup:
"""Create a new error group and return it."""
group = cls.ErrorGroup(prefix, name)
cls.groups.append(group)
return group
@classmethod
def get_error_codes(cls) -> Iterable[str]:
"""Yield all registered codes."""
for group in cls.groups:
for error in group.errors:
yield error.code
@classmethod
def to_rst(cls) -> str:
"""Output the registry as reStructuredText, for documentation."""
max_len = max(
len(error.short_desc)
for group in cls.groups
for error in group.errors
)
sep_line = '+' + 6 * '-' + '+' + '-' * (max_len + 2) + '+\n'
blank_line = '|' + (max_len + 9) * ' ' + '|\n'
table = ''
for group in cls.groups:
table += sep_line
table += blank_line
table += '|' + f'**{group.name}**'.center(max_len + 9) + '|\n'
table += blank_line
for error in group.errors:
table += sep_line
table += (
'|'
+ error.code.center(6)
+ '| '
+ error.short_desc.ljust(max_len + 1)
+ '|\n'
)
table += sep_line
return table
D1xx = ErrorRegistry.create_group('D1', 'Missing Docstrings')
D100 = D1xx.create_error(
'D100',
'Missing docstring in public module',
)
D101 = D1xx.create_error(
'D101',
'Missing docstring in public class',
)
D102 = D1xx.create_error(
'D102',
'Missing docstring in public method',
)
D103 = D1xx.create_error(
'D103',
'Missing docstring in public function',
)
D104 = D1xx.create_error(
'D104',
'Missing docstring in public package',
)
D105 = D1xx.create_error(
'D105',
'Missing docstring in magic method',
)
D106 = D1xx.create_error(
'D106',
'Missing docstring in public nested class',
)
D107 = D1xx.create_error(
'D107',
'Missing docstring in __init__',
)
D2xx = ErrorRegistry.create_group('D2', 'Whitespace Issues')
D200 = D2xx.create_error(
'D200',
'One-line docstring should fit on one line ' 'with quotes',
'found {0}',
)
D201 = D2xx.create_error(
'D201',
'No blank lines allowed before function docstring',
'found {0}',
)
D202 = D2xx.create_error(
'D202',
'No blank lines allowed after function docstring',
'found {0}',
)
D203 = D2xx.create_error(
'D203',
'1 blank line required before class docstring',
'found {0}',
)
D204 = D2xx.create_error(
'D204',
'1 blank line required after class docstring',
'found {0}',
)
D205 = D2xx.create_error(
'D205',
'1 blank line required between summary line and description',
'found {0}',
)
D206 = D2xx.create_error(
'D206',
'Docstring should be indented with spaces, not tabs',
)
D207 = D2xx.create_error(
'D207',
'Docstring is under-indented',
)
D208 = D2xx.create_error(
'D208',
'Docstring is over-indented',
)
D209 = D2xx.create_error(
'D209',
'Multi-line docstring closing quotes should be on a separate line',
)
D210 = D2xx.create_error(
'D210',
'No whitespaces allowed surrounding docstring text',
)
D211 = D2xx.create_error(
'D211',
'No blank lines allowed before class docstring',
'found {0}',
)
D212 = D2xx.create_error(
'D212',
'Multi-line docstring summary should start at the first line',
)
D213 = D2xx.create_error(
'D213',
'Multi-line docstring summary should start at the second line',
)
D214 = D2xx.create_error(
'D214',
'Section is over-indented',
'{0!r}',
) | 'D215',
'Section underline is over-indented',
'in section {0!r}',
)
D3xx = ErrorRegistry.create_group('D3', 'Quotes Issues')
D300 = D3xx.create_error(
'D300',
'Use """triple double quotes"""',
'found {0}-quotes',
)
D301 = D3xx.create_error(
'D301',
'Use r""" if any backslashes in a docstring',
)
D302 = D3xx.create_error(
'D302',
'Deprecated: Use u""" for Unicode docstrings',
)
D4xx = ErrorRegistry.create_group('D4', 'Docstring Content Issues')
D400 = D4xx.create_error(
'D400',
'First line should end with a period',
'not {0!r}',
)
D401 = D4xx.create_error(
'D401',
'First line should be in imperative mood',
"perhaps '{0}', not '{1}'",
)
D401b = D4xx.create_error(
'D401',
'First line should be in imperative mood; try rephrasing',
"found '{0}'",
)
D402 = D4xx.create_error(
'D402',
'First line should not be the function\'s "signature"',
)
D403 = D4xx.create_error(
'D403',
'First word of the first line should be properly capitalized',
'{0!r}, not {1!r}',
)
D404 = D4xx.create_error(
| D215 = D2xx.create_error( | random_line_split |
violations.py | ends=True)
offset = self.definition.start # type: ignore
lines_stripped = list(
reversed(list(dropwhile(is_blank, reversed(lines))))
)
numbers_width = len(str(offset + len(lines_stripped)))
line_format = f'{{:{numbers_width}}}:{{}}'
for n, line in enumerate(lines_stripped):
if line:
line = ' ' + line
source += line_format.format(n + offset, line)
if n > 5:
source += ' ...\n'
break
return source
def __str__(self) -> str:
if self.explanation:
self.explanation = '\n'.join(
l for l in self.explanation.split('\n') if not is_blank(l)
)
template = '{filename}:{line} {definition}:\n {message}'
if self.source and self.explain:
template += '\n\n{explanation}\n\n{lines}\n'
elif self.source and not self.explain:
template += '\n\n{lines}\n'
elif self.explain and not self.source:
template += '\n\n{explanation}\n\n'
return template.format(
**{
name: getattr(self, name)
for name in [
'filename',
'line',
'definition',
'message',
'explanation',
'lines',
]
}
)
def __repr__(self) -> str:
return str(self)
def __lt__(self, other: 'Error') -> bool:
return (self.filename, self.line) < (other.filename, other.line)
class ErrorRegistry:
"""A registry of all error codes, divided to groups."""
groups = [] # type: ignore
class ErrorGroup:
"""A group of similarly themed errors."""
def __init__(self, prefix: str, name: str) -> None:
"""Initialize the object.
`Prefix` should be the common prefix for errors in this group,
e.g., "D1".
`name` is the name of the group (its subject).
"""
self.prefix = prefix
self.name = name
self.errors = [] # type: List[ErrorParams]
def create_error(
self,
error_code: str,
error_desc: str,
error_context: Optional[str] = None,
) -> Callable[[Iterable[str]], Error]:
"""Create an error, register it to this group and return it."""
# TODO: check prefix
error_params = ErrorParams(error_code, error_desc, error_context)
factory = partial(Error, *error_params)
self.errors.append(error_params)
return factory
@classmethod
def create_group(cls, prefix: str, name: str) -> ErrorGroup:
"""Create a new error group and return it."""
group = cls.ErrorGroup(prefix, name)
cls.groups.append(group)
return group
@classmethod
def get_error_codes(cls) -> Iterable[str]:
"""Yield all registered codes."""
for group in cls.groups:
|
@classmethod
def to_rst(cls) -> str:
"""Output the registry as reStructuredText, for documentation."""
max_len = max(
len(error.short_desc)
for group in cls.groups
for error in group.errors
)
sep_line = '+' + 6 * '-' + '+' + '-' * (max_len + 2) + '+\n'
blank_line = '|' + (max_len + 9) * ' ' + '|\n'
table = ''
for group in cls.groups:
table += sep_line
table += blank_line
table += '|' + f'**{group.name}**'.center(max_len + 9) + '|\n'
table += blank_line
for error in group.errors:
table += sep_line
table += (
'|'
+ error.code.center(6)
+ '| '
+ error.short_desc.ljust(max_len + 1)
+ '|\n'
)
table += sep_line
return table
D1xx = ErrorRegistry.create_group('D1', 'Missing Docstrings')
D100 = D1xx.create_error(
'D100',
'Missing docstring in public module',
)
D101 = D1xx.create_error(
'D101',
'Missing docstring in public class',
)
D102 = D1xx.create_error(
'D102',
'Missing docstring in public method',
)
D103 = D1xx.create_error(
'D103',
'Missing docstring in public function',
)
D104 = D1xx.create_error(
'D104',
'Missing docstring in public package',
)
D105 = D1xx.create_error(
'D105',
'Missing docstring in magic method',
)
D106 = D1xx.create_error(
'D106',
'Missing docstring in public nested class',
)
D107 = D1xx.create_error(
'D107',
'Missing docstring in __init__',
)
D2xx = ErrorRegistry.create_group('D2', 'Whitespace Issues')
D200 = D2xx.create_error(
'D200',
'One-line docstring should fit on one line ' 'with quotes',
'found {0}',
)
D201 = D2xx.create_error(
'D201',
'No blank lines allowed before function docstring',
'found {0}',
)
D202 = D2xx.create_error(
'D202',
'No blank lines allowed after function docstring',
'found {0}',
)
D203 = D2xx.create_error(
'D203',
'1 blank line required before class docstring',
'found {0}',
)
D204 = D2xx.create_error(
'D204',
'1 blank line required after class docstring',
'found {0}',
)
D205 = D2xx.create_error(
'D205',
'1 blank line required between summary line and description',
'found {0}',
)
D206 = D2xx.create_error(
'D206',
'Docstring should be indented with spaces, not tabs',
)
D207 = D2xx.create_error(
'D207',
'Docstring is under-indented',
)
D208 = D2xx.create_error(
'D208',
'Docstring is over-indented',
)
D209 = D2xx.create_error(
'D209',
'Multi-line docstring closing quotes should be on a separate line',
)
D210 = D2xx.create_error(
'D210',
'No whitespaces allowed surrounding docstring text',
)
D211 = D2xx.create_error(
'D211',
'No blank lines allowed before class docstring',
'found {0}',
)
D212 = D2xx.create_error(
'D212',
'Multi-line docstring summary should start at the first line',
)
D213 = D2xx.create_error(
'D213',
'Multi-line docstring summary should start at the second line',
)
D214 = D2xx.create_error(
'D214',
'Section is over-indented',
'{0!r}',
)
D215 = D2xx.create_error(
'D215',
'Section underline is over-indented',
'in section {0!r}',
)
D3xx = ErrorRegistry.create_group('D3', 'Quotes Issues')
D300 = D3xx.create_error(
'D300',
'Use """triple double quotes"""',
'found {0}-quotes',
)
D301 = D3xx.create_error(
'D301',
'Use r""" if any backslashes in a docstring',
)
D302 = D3xx.create_error(
'D302',
'Deprecated: Use u""" for Unicode docstrings',
)
D4xx = ErrorRegistry.create_group('D4', 'Docstring Content Issues')
D400 = D4xx.create_error(
'D400',
'First line should end with a period',
'not {0!r}',
)
D401 = D4xx.create_error(
'D401',
'First line should be in imperative mood',
"perhaps '{0}', not '{1}'",
)
D401b = D4xx.create_error(
'D401',
'First line should be in imperative mood; try rephrasing',
"found '{0}'",
)
D402 = D4xx.create_error(
'D402',
'First line should not be the function\'s "signature"',
)
D403 = D4xx.create_error(
'D403',
'First word of the first line should be properly capitalized',
'{0!r}, not {1!r}',
)
D404 = D4xx.create_error(
| for error in group.errors:
yield error.code | conditional_block |
violations.py | (cls, prefix: str, name: str) -> ErrorGroup:
"""Create a new error group and return it."""
group = cls.ErrorGroup(prefix, name)
cls.groups.append(group)
return group
@classmethod
def get_error_codes(cls) -> Iterable[str]:
"""Yield all registered codes."""
for group in cls.groups:
for error in group.errors:
yield error.code
@classmethod
def to_rst(cls) -> str:
"""Output the registry as reStructuredText, for documentation."""
max_len = max(
len(error.short_desc)
for group in cls.groups
for error in group.errors
)
sep_line = '+' + 6 * '-' + '+' + '-' * (max_len + 2) + '+\n'
blank_line = '|' + (max_len + 9) * ' ' + '|\n'
table = ''
for group in cls.groups:
table += sep_line
table += blank_line
table += '|' + f'**{group.name}**'.center(max_len + 9) + '|\n'
table += blank_line
for error in group.errors:
table += sep_line
table += (
'|'
+ error.code.center(6)
+ '| '
+ error.short_desc.ljust(max_len + 1)
+ '|\n'
)
table += sep_line
return table
D1xx = ErrorRegistry.create_group('D1', 'Missing Docstrings')
D100 = D1xx.create_error(
'D100',
'Missing docstring in public module',
)
D101 = D1xx.create_error(
'D101',
'Missing docstring in public class',
)
D102 = D1xx.create_error(
'D102',
'Missing docstring in public method',
)
D103 = D1xx.create_error(
'D103',
'Missing docstring in public function',
)
D104 = D1xx.create_error(
'D104',
'Missing docstring in public package',
)
D105 = D1xx.create_error(
'D105',
'Missing docstring in magic method',
)
D106 = D1xx.create_error(
'D106',
'Missing docstring in public nested class',
)
D107 = D1xx.create_error(
'D107',
'Missing docstring in __init__',
)
D2xx = ErrorRegistry.create_group('D2', 'Whitespace Issues')
D200 = D2xx.create_error(
'D200',
'One-line docstring should fit on one line ' 'with quotes',
'found {0}',
)
D201 = D2xx.create_error(
'D201',
'No blank lines allowed before function docstring',
'found {0}',
)
D202 = D2xx.create_error(
'D202',
'No blank lines allowed after function docstring',
'found {0}',
)
D203 = D2xx.create_error(
'D203',
'1 blank line required before class docstring',
'found {0}',
)
D204 = D2xx.create_error(
'D204',
'1 blank line required after class docstring',
'found {0}',
)
D205 = D2xx.create_error(
'D205',
'1 blank line required between summary line and description',
'found {0}',
)
D206 = D2xx.create_error(
'D206',
'Docstring should be indented with spaces, not tabs',
)
D207 = D2xx.create_error(
'D207',
'Docstring is under-indented',
)
D208 = D2xx.create_error(
'D208',
'Docstring is over-indented',
)
D209 = D2xx.create_error(
'D209',
'Multi-line docstring closing quotes should be on a separate line',
)
D210 = D2xx.create_error(
'D210',
'No whitespaces allowed surrounding docstring text',
)
D211 = D2xx.create_error(
'D211',
'No blank lines allowed before class docstring',
'found {0}',
)
D212 = D2xx.create_error(
'D212',
'Multi-line docstring summary should start at the first line',
)
D213 = D2xx.create_error(
'D213',
'Multi-line docstring summary should start at the second line',
)
D214 = D2xx.create_error(
'D214',
'Section is over-indented',
'{0!r}',
)
D215 = D2xx.create_error(
'D215',
'Section underline is over-indented',
'in section {0!r}',
)
D3xx = ErrorRegistry.create_group('D3', 'Quotes Issues')
D300 = D3xx.create_error(
'D300',
'Use """triple double quotes"""',
'found {0}-quotes',
)
D301 = D3xx.create_error(
'D301',
'Use r""" if any backslashes in a docstring',
)
D302 = D3xx.create_error(
'D302',
'Deprecated: Use u""" for Unicode docstrings',
)
D4xx = ErrorRegistry.create_group('D4', 'Docstring Content Issues')
D400 = D4xx.create_error(
'D400',
'First line should end with a period',
'not {0!r}',
)
D401 = D4xx.create_error(
'D401',
'First line should be in imperative mood',
"perhaps '{0}', not '{1}'",
)
D401b = D4xx.create_error(
'D401',
'First line should be in imperative mood; try rephrasing',
"found '{0}'",
)
D402 = D4xx.create_error(
'D402',
'First line should not be the function\'s "signature"',
)
D403 = D4xx.create_error(
'D403',
'First word of the first line should be properly capitalized',
'{0!r}, not {1!r}',
)
D404 = D4xx.create_error(
'D404',
'First word of the docstring should not be `This`',
)
D405 = D4xx.create_error(
'D405',
'Section name should be properly capitalized',
'{0!r}, not {1!r}',
)
D406 = D4xx.create_error(
'D406',
'Section name should end with a newline',
'{0!r}, not {1!r}',
)
D407 = D4xx.create_error(
'D407',
'Missing dashed underline after section',
'{0!r}',
)
D408 = D4xx.create_error(
'D408',
'Section underline should be in the line following the section\'s name',
'{0!r}',
)
D409 = D4xx.create_error(
'D409',
'Section underline should match the length of its name',
'Expected {0!r} dashes in section {1!r}, got {2!r}',
)
D410 = D4xx.create_error(
'D410',
'Missing blank line after section',
'{0!r}',
)
D411 = D4xx.create_error(
'D411',
'Missing blank line before section',
'{0!r}',
)
D412 = D4xx.create_error(
'D412',
'No blank lines allowed between a section header and its content',
'{0!r}',
)
D413 = D4xx.create_error(
'D413',
'Missing blank line after last section',
'{0!r}',
)
D414 = D4xx.create_error(
'D414',
'Section has no content',
'{0!r}',
)
D415 = D4xx.create_error(
'D415',
(
'First line should end with a period, question '
'mark, or exclamation point'
),
'not {0!r}',
)
D416 = D4xx.create_error(
'D416',
'Section name should end with a colon',
'{0!r}, not {1!r}',
)
D417 = D4xx.create_error(
'D417',
'Missing argument descriptions in the docstring',
'argument(s) {0} are missing descriptions in {1!r} docstring',
)
D418 = D4xx.create_error(
'D418',
'Function/ Method decorated with @overload shouldn\'t contain a docstring',
)
class AttrDict(dict):
def | __getattr__ | identifier_name |
|
VoiceNetwork.tsx | import React, { useMemo, useState } from 'react';
import { TMemo } from '@shared/components/TMemo';
import { useRTCRoomClientContext } from '@rtc/RoomContext';
import { useAsyncTimeout } from '@shared/hooks/useAsyncTimeout';
import _get from 'lodash/get';
import filesize from 'filesize';
import { useTranslation } from '@shared/i18n';
/**
* 获取传输速度字符串
* @param bitrate 比特率
*/
function getStreamRate(bitrate: number): string {
return filesize(Number(bitrate) / 8, { bits: true }) + '/s';
}
/**
* 语音网络状态显示
*/
export const VoiceNetwork: React.FC = TMemo(() => {
const { client } = useRTCRoomClientContext();
const [remoteStats, setRemoteStats] = useState<any>({});
const { t } = useTranslation();
useAsyncTimeout(async () => {
if (client) {
const sendRemoteStats = await cl | () => {
return {
// 即生产者的传输通道在远程接收到的信息
upstream: getStreamRate(
_get(remoteStats, ['sendRemoteStats', 0, 'recvBitrate'], 0)
),
// 即消费者的传输通道在远程发送的信息
downstream: getStreamRate(
_get(remoteStats, ['recvRemoteStats', 0, 'sendBitrate'], 0)
),
};
}, [remoteStats]);
return (
<pre>
{t('上传')}: {bitrate.upstream} {t('下载')}: {bitrate.downstream}
</pre>
);
});
VoiceNetwork.displayName = 'VoiceNetwork';
| ient
.getSendTransportRemoteStats()
.catch(() => {});
const recvRemoteStats = await client
.getRecvTransportRemoteStats()
.catch(() => {});
setRemoteStats({
sendRemoteStats,
recvRemoteStats,
});
}
}, 2000);
const bitrate = useMemo( | conditional_block |
VoiceNetwork.tsx | import React, { useMemo, useState } from 'react';
import { TMemo } from '@shared/components/TMemo';
import { useRTCRoomClientContext } from '@rtc/RoomContext';
import { useAsyncTimeout } from '@shared/hooks/useAsyncTimeout';
import _get from 'lodash/get';
import filesize from 'filesize';
import { useTranslation } from '@shared/i18n';
/**
* 获取传输速度字符串
* @param bitrate 比特率
*/
function getStreamRate(bitrate: number): string {
return filesize(Numb | port const VoiceNetwork: React.FC = TMemo(() => {
const { client } = useRTCRoomClientContext();
const [remoteStats, setRemoteStats] = useState<any>({});
const { t } = useTranslation();
useAsyncTimeout(async () => {
if (client) {
const sendRemoteStats = await client
.getSendTransportRemoteStats()
.catch(() => {});
const recvRemoteStats = await client
.getRecvTransportRemoteStats()
.catch(() => {});
setRemoteStats({
sendRemoteStats,
recvRemoteStats,
});
}
}, 2000);
const bitrate = useMemo(() => {
return {
// 即生产者的传输通道在远程接收到的信息
upstream: getStreamRate(
_get(remoteStats, ['sendRemoteStats', 0, 'recvBitrate'], 0)
),
// 即消费者的传输通道在远程发送的信息
downstream: getStreamRate(
_get(remoteStats, ['recvRemoteStats', 0, 'sendBitrate'], 0)
),
};
}, [remoteStats]);
return (
<pre>
{t('上传')}: {bitrate.upstream} {t('下载')}: {bitrate.downstream}
</pre>
);
});
VoiceNetwork.displayName = 'VoiceNetwork';
| er(bitrate) / 8, { bits: true }) + '/s';
}
/**
* 语音网络状态显示
*/
ex | identifier_body |
VoiceNetwork.tsx | import React, { useMemo, useState } from 'react';
import { TMemo } from '@shared/components/TMemo';
import { useRTCRoomClientContext } from '@rtc/RoomContext';
import { useAsyncTimeout } from '@shared/hooks/useAsyncTimeout';
import _get from 'lodash/get';
import filesize from 'filesize';
import { useTranslation } from '@shared/i18n';
/**
* 获取传输速度字符串
* @param bitrate 比特率
*/
function getStreamRate(bitrate: number): string {
return filesize(Number(bitrate) / 8, { bits: true }) + '/s';
}
/**
* 语音网络状态显示
*/
export const VoiceNetwork: React.FC = TMemo(() => {
const { client } = useRTCRoomClientContext();
const [remoteStats, setRemoteStats] = useState<any>({});
const { t } = useTranslation();
useAsyncTimeout(async () => {
if (client) {
const sendRemoteStats = await client
.getSendTransportRemoteStats()
.catch(() => {});
const recvRemoteStats = await client
.getRecvTransportRemoteStats()
.catch(() => {});
setRemoteStats({
sendRemoteStats,
recvRemoteStats,
});
}
}, 2000);
const bitrate = useMemo(() => {
return {
// 即生产者的传输通道在远程接收到的信息
upstream: getStreamRate(
_get(remoteStats, ['sendRemoteStats', 0, 'recvBitrate'], 0)
),
// 即消费者的传输通道在远程发送的信息
downstream: getStreamRate( |
return (
<pre>
{t('上传')}: {bitrate.upstream} {t('下载')}: {bitrate.downstream}
</pre>
);
});
VoiceNetwork.displayName = 'VoiceNetwork'; | _get(remoteStats, ['recvRemoteStats', 0, 'sendBitrate'], 0)
),
};
}, [remoteStats]); | random_line_split |
VoiceNetwork.tsx | import React, { useMemo, useState } from 'react';
import { TMemo } from '@shared/components/TMemo';
import { useRTCRoomClientContext } from '@rtc/RoomContext';
import { useAsyncTimeout } from '@shared/hooks/useAsyncTimeout';
import _get from 'lodash/get';
import filesize from 'filesize';
import { useTranslation } from '@shared/i18n';
/**
* 获取传输速度字符串
* @param bitrate 比特率
*/
function getStreamRate(bitrate: n | g {
return filesize(Number(bitrate) / 8, { bits: true }) + '/s';
}
/**
* 语音网络状态显示
*/
export const VoiceNetwork: React.FC = TMemo(() => {
const { client } = useRTCRoomClientContext();
const [remoteStats, setRemoteStats] = useState<any>({});
const { t } = useTranslation();
useAsyncTimeout(async () => {
if (client) {
const sendRemoteStats = await client
.getSendTransportRemoteStats()
.catch(() => {});
const recvRemoteStats = await client
.getRecvTransportRemoteStats()
.catch(() => {});
setRemoteStats({
sendRemoteStats,
recvRemoteStats,
});
}
}, 2000);
const bitrate = useMemo(() => {
return {
// 即生产者的传输通道在远程接收到的信息
upstream: getStreamRate(
_get(remoteStats, ['sendRemoteStats', 0, 'recvBitrate'], 0)
),
// 即消费者的传输通道在远程发送的信息
downstream: getStreamRate(
_get(remoteStats, ['recvRemoteStats', 0, 'sendBitrate'], 0)
),
};
}, [remoteStats]);
return (
<pre>
{t('上传')}: {bitrate.upstream} {t('下载')}: {bitrate.downstream}
</pre>
);
});
VoiceNetwork.displayName = 'VoiceNetwork';
| umber): strin | identifier_name |
rayon.rs | use crate::{ProgressBar, ProgressBarIter};
use rayon::iter::{
plumbing::{Consumer, Folder, Producer, ProducerCallback, UnindexedConsumer},
IndexedParallelIterator, ParallelIterator,
};
use std::convert::TryFrom;
/// Wraps a Rayon parallel iterator.
///
/// See [`ProgressIterator`](trait.ProgressIterator.html) for method
/// documentation.
pub trait ParallelProgressIterator
where
Self: Sized + ParallelIterator,
{
/// Wrap an iterator with a custom progress bar.
fn progress_with(self, progress: ProgressBar) -> ProgressBarIter<Self>;
/// Wrap an iterator with an explicit element count.
fn progress_count(self, len: u64) -> ProgressBarIter<Self> {
self.progress_with(ProgressBar::new(len))
}
fn progress(self) -> ProgressBarIter<Self>
where
Self: IndexedParallelIterator,
{
let len = u64::try_from(self.len()).unwrap();
self.progress_count(len)
}
/// Wrap an iterator with a progress bar and style it.
fn progress_with_style(self, style: crate::ProgressStyle) -> ProgressBarIter<Self>
where
Self: IndexedParallelIterator,
{
let len = u64::try_from(self.len()).unwrap();
let bar = ProgressBar::new(len).with_style(style);
self.progress_with(bar)
}
}
impl<S: Send, T: ParallelIterator<Item = S>> ParallelProgressIterator for T {
fn progress_with(self, progress: ProgressBar) -> ProgressBarIter<Self> {
ProgressBarIter { it: self, progress }
}
}
impl<S: Send, T: IndexedParallelIterator<Item = S>> IndexedParallelIterator for ProgressBarIter<T> {
fn len(&self) -> usize |
fn drive<C: Consumer<Self::Item>>(self, consumer: C) -> <C as Consumer<Self::Item>>::Result {
let consumer = ProgressConsumer::new(consumer, self.progress);
self.it.drive(consumer)
}
fn with_producer<CB: ProducerCallback<Self::Item>>(
self,
callback: CB,
) -> <CB as ProducerCallback<Self::Item>>::Output {
return self.it.with_producer(Callback {
callback,
progress: self.progress,
});
struct Callback<CB> {
callback: CB,
progress: ProgressBar,
}
impl<T, CB: ProducerCallback<T>> ProducerCallback<T> for Callback<CB> {
type Output = CB::Output;
fn callback<P>(self, base: P) -> CB::Output
where
P: Producer<Item = T>,
{
let producer = ProgressProducer {
base,
progress: self.progress,
};
self.callback.callback(producer)
}
}
}
}
struct ProgressProducer<T> {
base: T,
progress: ProgressBar,
}
impl<T, P: Producer<Item = T>> Producer for ProgressProducer<P> {
type Item = T;
type IntoIter = ProgressBarIter<P::IntoIter>;
fn into_iter(self) -> Self::IntoIter {
ProgressBarIter {
it: self.base.into_iter(),
progress: self.progress,
}
}
fn min_len(&self) -> usize {
self.base.min_len()
}
fn max_len(&self) -> usize {
self.base.max_len()
}
fn split_at(self, index: usize) -> (Self, Self) {
let (left, right) = self.base.split_at(index);
(
ProgressProducer {
base: left,
progress: self.progress.clone(),
},
ProgressProducer {
base: right,
progress: self.progress,
},
)
}
}
struct ProgressConsumer<C> {
base: C,
progress: ProgressBar,
}
impl<C> ProgressConsumer<C> {
fn new(base: C, progress: ProgressBar) -> Self {
ProgressConsumer { base, progress }
}
}
impl<T, C: Consumer<T>> Consumer<T> for ProgressConsumer<C> {
type Folder = ProgressFolder<C::Folder>;
type Reducer = C::Reducer;
type Result = C::Result;
fn split_at(self, index: usize) -> (Self, Self, Self::Reducer) {
let (left, right, reducer) = self.base.split_at(index);
(
ProgressConsumer::new(left, self.progress.clone()),
ProgressConsumer::new(right, self.progress),
reducer,
)
}
fn into_folder(self) -> Self::Folder {
ProgressFolder {
base: self.base.into_folder(),
progress: self.progress,
}
}
fn full(&self) -> bool {
self.base.full()
}
}
impl<T, C: UnindexedConsumer<T>> UnindexedConsumer<T> for ProgressConsumer<C> {
fn split_off_left(&self) -> Self {
ProgressConsumer::new(self.base.split_off_left(), self.progress.clone())
}
fn to_reducer(&self) -> Self::Reducer {
self.base.to_reducer()
}
}
struct ProgressFolder<C> {
base: C,
progress: ProgressBar,
}
impl<T, C: Folder<T>> Folder<T> for ProgressFolder<C> {
type Result = C::Result;
fn consume(self, item: T) -> Self {
self.progress.inc(1);
ProgressFolder {
base: self.base.consume(item),
progress: self.progress,
}
}
fn complete(self) -> C::Result {
self.base.complete()
}
fn full(&self) -> bool {
self.base.full()
}
}
impl<S: Send, T: ParallelIterator<Item = S>> ParallelIterator for ProgressBarIter<T> {
type Item = S;
fn drive_unindexed<C: UnindexedConsumer<Self::Item>>(self, consumer: C) -> C::Result {
let consumer1 = ProgressConsumer::new(consumer, self.progress.clone());
self.it.drive_unindexed(consumer1)
}
}
#[cfg(test)]
mod test {
use crate::ProgressStyle;
use crate::{ParallelProgressIterator, ProgressBar, ProgressBarIter};
use rayon::iter::{IntoParallelRefIterator, ParallelIterator};
#[test]
fn it_can_wrap_a_parallel_iterator() {
let v = vec![1, 2, 3];
fn wrap<'a, T: ParallelIterator<Item = &'a i32>>(it: ProgressBarIter<T>) {
assert_eq!(it.map(|x| x * 2).collect::<Vec<_>>(), vec![2, 4, 6]);
}
wrap(v.par_iter().progress_count(3));
wrap({
let pb = ProgressBar::new(v.len() as u64);
v.par_iter().progress_with(pb)
});
wrap({
let style = ProgressStyle::default_bar().template("{wide_bar:.red} {percent}/100%");
v.par_iter().progress_with_style(style)
});
}
}
| {
self.it.len()
} | identifier_body |
rayon.rs | use crate::{ProgressBar, ProgressBarIter};
use rayon::iter::{
plumbing::{Consumer, Folder, Producer, ProducerCallback, UnindexedConsumer},
IndexedParallelIterator, ParallelIterator,
};
use std::convert::TryFrom;
/// Wraps a Rayon parallel iterator.
///
/// See [`ProgressIterator`](trait.ProgressIterator.html) for method
/// documentation.
pub trait ParallelProgressIterator
where
Self: Sized + ParallelIterator,
{
/// Wrap an iterator with a custom progress bar.
fn progress_with(self, progress: ProgressBar) -> ProgressBarIter<Self>;
/// Wrap an iterator with an explicit element count.
fn progress_count(self, len: u64) -> ProgressBarIter<Self> {
self.progress_with(ProgressBar::new(len))
}
fn progress(self) -> ProgressBarIter<Self>
where
Self: IndexedParallelIterator,
{
let len = u64::try_from(self.len()).unwrap();
self.progress_count(len)
}
/// Wrap an iterator with a progress bar and style it.
fn progress_with_style(self, style: crate::ProgressStyle) -> ProgressBarIter<Self>
where
Self: IndexedParallelIterator,
{
let len = u64::try_from(self.len()).unwrap();
let bar = ProgressBar::new(len).with_style(style);
self.progress_with(bar)
}
}
impl<S: Send, T: ParallelIterator<Item = S>> ParallelProgressIterator for T {
fn progress_with(self, progress: ProgressBar) -> ProgressBarIter<Self> {
ProgressBarIter { it: self, progress }
}
}
impl<S: Send, T: IndexedParallelIterator<Item = S>> IndexedParallelIterator for ProgressBarIter<T> {
fn len(&self) -> usize {
self.it.len()
}
fn drive<C: Consumer<Self::Item>>(self, consumer: C) -> <C as Consumer<Self::Item>>::Result {
let consumer = ProgressConsumer::new(consumer, self.progress);
self.it.drive(consumer)
}
fn with_producer<CB: ProducerCallback<Self::Item>>(
self,
callback: CB,
) -> <CB as ProducerCallback<Self::Item>>::Output {
return self.it.with_producer(Callback {
callback,
progress: self.progress,
});
struct Callback<CB> {
callback: CB,
progress: ProgressBar,
}
impl<T, CB: ProducerCallback<T>> ProducerCallback<T> for Callback<CB> {
type Output = CB::Output;
fn callback<P>(self, base: P) -> CB::Output
where
P: Producer<Item = T>,
{
let producer = ProgressProducer {
base,
progress: self.progress,
};
self.callback.callback(producer)
}
}
}
}
struct ProgressProducer<T> {
base: T,
progress: ProgressBar,
}
impl<T, P: Producer<Item = T>> Producer for ProgressProducer<P> {
type Item = T;
type IntoIter = ProgressBarIter<P::IntoIter>;
fn into_iter(self) -> Self::IntoIter {
ProgressBarIter {
it: self.base.into_iter(),
progress: self.progress,
}
}
fn min_len(&self) -> usize {
self.base.min_len()
}
fn max_len(&self) -> usize {
self.base.max_len()
}
fn split_at(self, index: usize) -> (Self, Self) {
let (left, right) = self.base.split_at(index);
(
ProgressProducer {
base: left,
progress: self.progress.clone(),
},
ProgressProducer {
base: right,
progress: self.progress,
},
)
}
}
struct ProgressConsumer<C> {
base: C,
progress: ProgressBar,
}
impl<C> ProgressConsumer<C> {
fn new(base: C, progress: ProgressBar) -> Self {
ProgressConsumer { base, progress }
}
}
impl<T, C: Consumer<T>> Consumer<T> for ProgressConsumer<C> {
type Folder = ProgressFolder<C::Folder>;
type Reducer = C::Reducer;
type Result = C::Result;
fn split_at(self, index: usize) -> (Self, Self, Self::Reducer) {
let (left, right, reducer) = self.base.split_at(index);
(
ProgressConsumer::new(left, self.progress.clone()),
ProgressConsumer::new(right, self.progress),
reducer,
)
}
fn into_folder(self) -> Self::Folder {
ProgressFolder {
base: self.base.into_folder(),
progress: self.progress,
}
}
fn full(&self) -> bool {
self.base.full()
}
}
impl<T, C: UnindexedConsumer<T>> UnindexedConsumer<T> for ProgressConsumer<C> {
fn split_off_left(&self) -> Self {
ProgressConsumer::new(self.base.split_off_left(), self.progress.clone())
}
fn to_reducer(&self) -> Self::Reducer {
self.base.to_reducer()
}
}
struct ProgressFolder<C> {
base: C,
progress: ProgressBar,
}
impl<T, C: Folder<T>> Folder<T> for ProgressFolder<C> {
type Result = C::Result;
fn | (self, item: T) -> Self {
self.progress.inc(1);
ProgressFolder {
base: self.base.consume(item),
progress: self.progress,
}
}
fn complete(self) -> C::Result {
self.base.complete()
}
fn full(&self) -> bool {
self.base.full()
}
}
impl<S: Send, T: ParallelIterator<Item = S>> ParallelIterator for ProgressBarIter<T> {
type Item = S;
fn drive_unindexed<C: UnindexedConsumer<Self::Item>>(self, consumer: C) -> C::Result {
let consumer1 = ProgressConsumer::new(consumer, self.progress.clone());
self.it.drive_unindexed(consumer1)
}
}
#[cfg(test)]
mod test {
use crate::ProgressStyle;
use crate::{ParallelProgressIterator, ProgressBar, ProgressBarIter};
use rayon::iter::{IntoParallelRefIterator, ParallelIterator};
#[test]
fn it_can_wrap_a_parallel_iterator() {
let v = vec![1, 2, 3];
fn wrap<'a, T: ParallelIterator<Item = &'a i32>>(it: ProgressBarIter<T>) {
assert_eq!(it.map(|x| x * 2).collect::<Vec<_>>(), vec![2, 4, 6]);
}
wrap(v.par_iter().progress_count(3));
wrap({
let pb = ProgressBar::new(v.len() as u64);
v.par_iter().progress_with(pb)
});
wrap({
let style = ProgressStyle::default_bar().template("{wide_bar:.red} {percent}/100%");
v.par_iter().progress_with_style(style)
});
}
}
| consume | identifier_name |
rayon.rs | use crate::{ProgressBar, ProgressBarIter};
use rayon::iter::{
plumbing::{Consumer, Folder, Producer, ProducerCallback, UnindexedConsumer},
IndexedParallelIterator, ParallelIterator,
};
use std::convert::TryFrom;
/// Wraps a Rayon parallel iterator.
///
/// See [`ProgressIterator`](trait.ProgressIterator.html) for method
/// documentation.
pub trait ParallelProgressIterator
where
Self: Sized + ParallelIterator,
{
/// Wrap an iterator with a custom progress bar.
fn progress_with(self, progress: ProgressBar) -> ProgressBarIter<Self>;
/// Wrap an iterator with an explicit element count.
fn progress_count(self, len: u64) -> ProgressBarIter<Self> {
self.progress_with(ProgressBar::new(len))
}
fn progress(self) -> ProgressBarIter<Self>
where
Self: IndexedParallelIterator,
{
let len = u64::try_from(self.len()).unwrap();
self.progress_count(len)
}
/// Wrap an iterator with a progress bar and style it.
fn progress_with_style(self, style: crate::ProgressStyle) -> ProgressBarIter<Self>
where
Self: IndexedParallelIterator,
{
let len = u64::try_from(self.len()).unwrap();
let bar = ProgressBar::new(len).with_style(style);
self.progress_with(bar)
}
}
impl<S: Send, T: ParallelIterator<Item = S>> ParallelProgressIterator for T {
fn progress_with(self, progress: ProgressBar) -> ProgressBarIter<Self> {
ProgressBarIter { it: self, progress }
}
}
impl<S: Send, T: IndexedParallelIterator<Item = S>> IndexedParallelIterator for ProgressBarIter<T> {
fn len(&self) -> usize {
self.it.len()
}
fn drive<C: Consumer<Self::Item>>(self, consumer: C) -> <C as Consumer<Self::Item>>::Result {
let consumer = ProgressConsumer::new(consumer, self.progress);
self.it.drive(consumer)
}
fn with_producer<CB: ProducerCallback<Self::Item>>(
self,
callback: CB,
) -> <CB as ProducerCallback<Self::Item>>::Output {
return self.it.with_producer(Callback {
callback,
progress: self.progress,
});
struct Callback<CB> {
callback: CB,
progress: ProgressBar,
}
impl<T, CB: ProducerCallback<T>> ProducerCallback<T> for Callback<CB> {
type Output = CB::Output;
fn callback<P>(self, base: P) -> CB::Output
where
P: Producer<Item = T>,
{
let producer = ProgressProducer {
base,
progress: self.progress,
};
self.callback.callback(producer)
}
}
}
}
struct ProgressProducer<T> {
base: T,
progress: ProgressBar,
}
impl<T, P: Producer<Item = T>> Producer for ProgressProducer<P> {
type Item = T;
type IntoIter = ProgressBarIter<P::IntoIter>;
fn into_iter(self) -> Self::IntoIter {
ProgressBarIter {
it: self.base.into_iter(),
progress: self.progress,
}
}
fn min_len(&self) -> usize {
self.base.min_len()
}
fn max_len(&self) -> usize {
self.base.max_len()
}
fn split_at(self, index: usize) -> (Self, Self) {
let (left, right) = self.base.split_at(index);
(
ProgressProducer {
base: left,
progress: self.progress.clone(),
},
ProgressProducer {
base: right,
progress: self.progress,
},
)
}
}
struct ProgressConsumer<C> {
base: C,
progress: ProgressBar,
}
impl<C> ProgressConsumer<C> { | }
impl<T, C: Consumer<T>> Consumer<T> for ProgressConsumer<C> {
type Folder = ProgressFolder<C::Folder>;
type Reducer = C::Reducer;
type Result = C::Result;
fn split_at(self, index: usize) -> (Self, Self, Self::Reducer) {
let (left, right, reducer) = self.base.split_at(index);
(
ProgressConsumer::new(left, self.progress.clone()),
ProgressConsumer::new(right, self.progress),
reducer,
)
}
fn into_folder(self) -> Self::Folder {
ProgressFolder {
base: self.base.into_folder(),
progress: self.progress,
}
}
fn full(&self) -> bool {
self.base.full()
}
}
impl<T, C: UnindexedConsumer<T>> UnindexedConsumer<T> for ProgressConsumer<C> {
fn split_off_left(&self) -> Self {
ProgressConsumer::new(self.base.split_off_left(), self.progress.clone())
}
fn to_reducer(&self) -> Self::Reducer {
self.base.to_reducer()
}
}
struct ProgressFolder<C> {
base: C,
progress: ProgressBar,
}
impl<T, C: Folder<T>> Folder<T> for ProgressFolder<C> {
type Result = C::Result;
fn consume(self, item: T) -> Self {
self.progress.inc(1);
ProgressFolder {
base: self.base.consume(item),
progress: self.progress,
}
}
fn complete(self) -> C::Result {
self.base.complete()
}
fn full(&self) -> bool {
self.base.full()
}
}
impl<S: Send, T: ParallelIterator<Item = S>> ParallelIterator for ProgressBarIter<T> {
type Item = S;
fn drive_unindexed<C: UnindexedConsumer<Self::Item>>(self, consumer: C) -> C::Result {
let consumer1 = ProgressConsumer::new(consumer, self.progress.clone());
self.it.drive_unindexed(consumer1)
}
}
#[cfg(test)]
mod test {
use crate::ProgressStyle;
use crate::{ParallelProgressIterator, ProgressBar, ProgressBarIter};
use rayon::iter::{IntoParallelRefIterator, ParallelIterator};
#[test]
fn it_can_wrap_a_parallel_iterator() {
let v = vec![1, 2, 3];
fn wrap<'a, T: ParallelIterator<Item = &'a i32>>(it: ProgressBarIter<T>) {
assert_eq!(it.map(|x| x * 2).collect::<Vec<_>>(), vec![2, 4, 6]);
}
wrap(v.par_iter().progress_count(3));
wrap({
let pb = ProgressBar::new(v.len() as u64);
v.par_iter().progress_with(pb)
});
wrap({
let style = ProgressStyle::default_bar().template("{wide_bar:.red} {percent}/100%");
v.par_iter().progress_with_style(style)
});
}
} | fn new(base: C, progress: ProgressBar) -> Self {
ProgressConsumer { base, progress }
} | random_line_split |
class_physics_component.js | var class_physics_component =
[
[ "createXmlNode", "class_physics_component.html#a5a2e3761a13d45a4dd38fe3b69253332", null ],
[ "destroyDispatcher", "class_physics_component.html#a3c17f238e0ea725fc91a151591bf9510", null ],
[ "getPosition", "class_physics_component.html#aeee07d4204bae0ff7747f5c0009907a1", null ],
[ "setPosition", "class_physics_component.html#a12d373e7cba22ea2d5925063664fc7e2", null ],
[ "PhysicsSystem", "class_physics_component.html#a6fb7520528fab4a670001f041b872bf2", null ], | [ "myPhysicsSystem", "class_physics_component.html#a975c62b57bcba88f3738edfe308da17b", null ]
]; | random_line_split |
|
view.tree.js | class TreeView {
constructor($dom, store, adapter) {
this.store = store;
this.adapter = adapter;
this.$view = $dom.find('.octotree_treeview');
this.$tree = this.$view
.find('.octotree_view_body')
.on('click.jstree', '.jstree-open>a', ({target}) => {
setTimeout(() => {
this.$jstree.close_node(target)
}, 0);
})
.on('click.jstree', '.jstree-closed>a', ({target}) => {
setTimeout(() => {
this.$jstree.open_node(target)
}, 0);
})
.on('click', this._onItemClick.bind(this))
.jstree({
core: {multiple: false, worker: false, themes: {responsive: false}},
plugins: ['wholerow']
});
}
get $jstree() {
return this.$tree.jstree(true);
}
show(repo, token) {
const $jstree = this.$jstree;
$jstree.settings.core.data = (node, cb) => {
const prMode = this.store.get(STORE.PR) && repo.pullNumber;
const loadAll = this.adapter.canLoadEntireTree(repo) && (prMode || this.store.get(STORE.LOADALL));
node = !loadAll && (node.id === '#' ? {path: ''} : node.original);
this.adapter.loadCodeTree({repo, token, node}, (err, treeData) => {
if (err) {
if (err.status === 206 && loadAll) { // The repo is too big to load all, need to retry
$jstree.refresh(true);
} else {
$(this).trigger(EVENT.FETCH_ERROR, [err]);
}
} else {
treeData = this._sort(treeData);
if (loadAll) {
treeData = this._collapse(treeData);
}
cb(treeData);
}
});
};
this.$tree.one('refresh.jstree', () => {
this.syncSelection(repo);
$(this).trigger(EVENT.VIEW_READY);
});
this._showHeader(repo);
$jstree.refresh(true);
}
_showHeader(repo) {
const adapter = this.adapter;
this.$view
.find('.octotree_view_header')
.html(
`<div class="octotree_header_repo">
<a href="/${repo.username}">${repo.username}</a>
/
<a data-pjax href="/${repo.username}/${repo.reponame}">${repo.reponame}</a>
</div>
<div class="octotree_header_branch">
${this._deXss(repo.branch.toString())}
</div>`
)
.on('click', 'a[data-pjax]', function(event) {
event.preventDefault();
// A.href always return absolute URL, don't want that
const href = $(this).attr('href');
const newTab = event.shiftKey || event.ctrlKey || event.metaKey;
newTab ? adapter.openInNewTab(href) : adapter.selectFile(href);
});
}
_deXss(str) {
return str && str.replace(/[<>'"&]/g, '-');
}
_sort(folder) {
folder.sort((a, b) => {
if (a.type === b.type) return a.text === b.text ? 0 : a.text < b.text ? -1 : 1;
return a.type === 'blob' ? 1 : -1;
});
folder.forEach((item) => {
if (item.type === 'tree' && item.children !== true && item.children.length > 0) {
this._sort(item.children);
}
});
return folder;
}
_collapse(folder) {
return folder.map((item) => {
if (item.type === 'tree') {
item.children = this._collapse(item.children);
if (item.children.length === 1 && item.children[0].type === 'tree') {
const onlyChild = item.children[0];
onlyChild.text = item.text + '/' + onlyChild.text;
return onlyChild;
}
}
return item;
});
}
_onItemClick(event) {
let $target = $(event.target);
let download = false;
// Handle middle click
if (event.which === 2) return;
// Handle icon click, fix #122
if ($target.is('i.jstree-icon')) {
$target = $target.parent();
download = true;
}
if (!$target.is('a.jstree-anchor')) return;
// Refocus after complete so that keyboard navigation works, fix #158
const refocusAfterCompletion = () => {
$(document).one('pjax:success page:load', () => {
this.$jstree.get_container().focus();
});
};
const adapter = this.adapter;
const newTab = event.shiftKey || event.ctrlKey || event.metaKey;
const href = $target.attr('href');
// The 2nd path is for submodule child links
const $icon = $target.children().length ? $target.children(':first') : $target.siblings(':first');
if ($icon.hasClass('commit')) {
refocusAfterCompletion();
newTab ? adapter.openInNewTab(href) : adapter.selectSubmodule(href);
} else if ($icon.hasClass('blob')) {
if (download) {
const downloadUrl = $target.attr('data-download-url');
const downloadFileName = $target.attr('data-download-filename');
adapter.downloadFile(downloadUrl, downloadFileName);
} else {
refocusAfterCompletion();
newTab ? adapter.openInNewTab(href) : adapter.selectFile(href);
}
}
}
syncSelection(repo) {
const $jstree = this.$jstree;
if (!$jstree) return;
// Convert /username/reponame/object_type/branch/path to path
const path = decodeURIComponent(location.pathname);
const match = path.match(/(?:[^\/]+\/){4}(.*)/);
if (!match) return;
const currentPath = match[1];
const loadAll = this.adapter.canLoadEntireTree(repo) && this.store.get(STORE.LOADALL);
selectPath(loadAll ? [currentPath] : breakPath(currentPath));
// Convert ['a/b'] to ['a', 'a/b']
function breakPath(fullPath) {
return fullPath.split('/').reduce((res, path, idx) => {
res.push(idx === 0 ? path : `${res[idx - 1]}/${path}`);
return res;
}, []); |
if ($jstree.get_node(nodeId)) {
$jstree.deselect_all();
$jstree.select_node(nodeId);
$jstree.open_node(nodeId, () => {
if (++index < paths.length) {
selectPath(paths, index);
}
});
}
}
}
} | }
function selectPath(paths, index = 0) {
const nodeId = NODE_PREFIX + paths[index]; | random_line_split |
view.tree.js | class TreeView {
constructor($dom, store, adapter) {
this.store = store;
this.adapter = adapter;
this.$view = $dom.find('.octotree_treeview');
this.$tree = this.$view
.find('.octotree_view_body')
.on('click.jstree', '.jstree-open>a', ({target}) => {
setTimeout(() => {
this.$jstree.close_node(target)
}, 0);
})
.on('click.jstree', '.jstree-closed>a', ({target}) => {
setTimeout(() => {
this.$jstree.open_node(target)
}, 0);
})
.on('click', this._onItemClick.bind(this))
.jstree({
core: {multiple: false, worker: false, themes: {responsive: false}},
plugins: ['wholerow']
});
}
get $jstree() {
return this.$tree.jstree(true);
}
show(repo, token) {
const $jstree = this.$jstree;
$jstree.settings.core.data = (node, cb) => {
const prMode = this.store.get(STORE.PR) && repo.pullNumber;
const loadAll = this.adapter.canLoadEntireTree(repo) && (prMode || this.store.get(STORE.LOADALL));
node = !loadAll && (node.id === '#' ? {path: ''} : node.original);
this.adapter.loadCodeTree({repo, token, node}, (err, treeData) => {
if (err) {
if (err.status === 206 && loadAll) { // The repo is too big to load all, need to retry
$jstree.refresh(true);
} else {
$(this).trigger(EVENT.FETCH_ERROR, [err]);
}
} else {
treeData = this._sort(treeData);
if (loadAll) {
treeData = this._collapse(treeData);
}
cb(treeData);
}
});
};
this.$tree.one('refresh.jstree', () => {
this.syncSelection(repo);
$(this).trigger(EVENT.VIEW_READY);
});
this._showHeader(repo);
$jstree.refresh(true);
}
_showHeader(repo) {
const adapter = this.adapter;
this.$view
.find('.octotree_view_header')
.html(
`<div class="octotree_header_repo">
<a href="/${repo.username}">${repo.username}</a>
/
<a data-pjax href="/${repo.username}/${repo.reponame}">${repo.reponame}</a>
</div>
<div class="octotree_header_branch">
${this._deXss(repo.branch.toString())}
</div>`
)
.on('click', 'a[data-pjax]', function(event) {
event.preventDefault();
// A.href always return absolute URL, don't want that
const href = $(this).attr('href');
const newTab = event.shiftKey || event.ctrlKey || event.metaKey;
newTab ? adapter.openInNewTab(href) : adapter.selectFile(href);
});
}
_deXss(str) {
return str && str.replace(/[<>'"&]/g, '-');
}
| (folder) {
folder.sort((a, b) => {
if (a.type === b.type) return a.text === b.text ? 0 : a.text < b.text ? -1 : 1;
return a.type === 'blob' ? 1 : -1;
});
folder.forEach((item) => {
if (item.type === 'tree' && item.children !== true && item.children.length > 0) {
this._sort(item.children);
}
});
return folder;
}
_collapse(folder) {
return folder.map((item) => {
if (item.type === 'tree') {
item.children = this._collapse(item.children);
if (item.children.length === 1 && item.children[0].type === 'tree') {
const onlyChild = item.children[0];
onlyChild.text = item.text + '/' + onlyChild.text;
return onlyChild;
}
}
return item;
});
}
_onItemClick(event) {
let $target = $(event.target);
let download = false;
// Handle middle click
if (event.which === 2) return;
// Handle icon click, fix #122
if ($target.is('i.jstree-icon')) {
$target = $target.parent();
download = true;
}
if (!$target.is('a.jstree-anchor')) return;
// Refocus after complete so that keyboard navigation works, fix #158
const refocusAfterCompletion = () => {
$(document).one('pjax:success page:load', () => {
this.$jstree.get_container().focus();
});
};
const adapter = this.adapter;
const newTab = event.shiftKey || event.ctrlKey || event.metaKey;
const href = $target.attr('href');
// The 2nd path is for submodule child links
const $icon = $target.children().length ? $target.children(':first') : $target.siblings(':first');
if ($icon.hasClass('commit')) {
refocusAfterCompletion();
newTab ? adapter.openInNewTab(href) : adapter.selectSubmodule(href);
} else if ($icon.hasClass('blob')) {
if (download) {
const downloadUrl = $target.attr('data-download-url');
const downloadFileName = $target.attr('data-download-filename');
adapter.downloadFile(downloadUrl, downloadFileName);
} else {
refocusAfterCompletion();
newTab ? adapter.openInNewTab(href) : adapter.selectFile(href);
}
}
}
syncSelection(repo) {
const $jstree = this.$jstree;
if (!$jstree) return;
// Convert /username/reponame/object_type/branch/path to path
const path = decodeURIComponent(location.pathname);
const match = path.match(/(?:[^\/]+\/){4}(.*)/);
if (!match) return;
const currentPath = match[1];
const loadAll = this.adapter.canLoadEntireTree(repo) && this.store.get(STORE.LOADALL);
selectPath(loadAll ? [currentPath] : breakPath(currentPath));
// Convert ['a/b'] to ['a', 'a/b']
function breakPath(fullPath) {
return fullPath.split('/').reduce((res, path, idx) => {
res.push(idx === 0 ? path : `${res[idx - 1]}/${path}`);
return res;
}, []);
}
function selectPath(paths, index = 0) {
const nodeId = NODE_PREFIX + paths[index];
if ($jstree.get_node(nodeId)) {
$jstree.deselect_all();
$jstree.select_node(nodeId);
$jstree.open_node(nodeId, () => {
if (++index < paths.length) {
selectPath(paths, index);
}
});
}
}
}
}
| _sort | identifier_name |
view.tree.js | class TreeView {
constructor($dom, store, adapter) {
this.store = store;
this.adapter = adapter;
this.$view = $dom.find('.octotree_treeview');
this.$tree = this.$view
.find('.octotree_view_body')
.on('click.jstree', '.jstree-open>a', ({target}) => {
setTimeout(() => {
this.$jstree.close_node(target)
}, 0);
})
.on('click.jstree', '.jstree-closed>a', ({target}) => {
setTimeout(() => {
this.$jstree.open_node(target)
}, 0);
})
.on('click', this._onItemClick.bind(this))
.jstree({
core: {multiple: false, worker: false, themes: {responsive: false}},
plugins: ['wholerow']
});
}
get $jstree() {
return this.$tree.jstree(true);
}
show(repo, token) {
const $jstree = this.$jstree;
$jstree.settings.core.data = (node, cb) => {
const prMode = this.store.get(STORE.PR) && repo.pullNumber;
const loadAll = this.adapter.canLoadEntireTree(repo) && (prMode || this.store.get(STORE.LOADALL));
node = !loadAll && (node.id === '#' ? {path: ''} : node.original);
this.adapter.loadCodeTree({repo, token, node}, (err, treeData) => {
if (err) {
if (err.status === 206 && loadAll) { // The repo is too big to load all, need to retry
$jstree.refresh(true);
} else {
$(this).trigger(EVENT.FETCH_ERROR, [err]);
}
} else {
treeData = this._sort(treeData);
if (loadAll) {
treeData = this._collapse(treeData);
}
cb(treeData);
}
});
};
this.$tree.one('refresh.jstree', () => {
this.syncSelection(repo);
$(this).trigger(EVENT.VIEW_READY);
});
this._showHeader(repo);
$jstree.refresh(true);
}
_showHeader(repo) {
const adapter = this.adapter;
this.$view
.find('.octotree_view_header')
.html(
`<div class="octotree_header_repo">
<a href="/${repo.username}">${repo.username}</a>
/
<a data-pjax href="/${repo.username}/${repo.reponame}">${repo.reponame}</a>
</div>
<div class="octotree_header_branch">
${this._deXss(repo.branch.toString())}
</div>`
)
.on('click', 'a[data-pjax]', function(event) {
event.preventDefault();
// A.href always return absolute URL, don't want that
const href = $(this).attr('href');
const newTab = event.shiftKey || event.ctrlKey || event.metaKey;
newTab ? adapter.openInNewTab(href) : adapter.selectFile(href);
});
}
_deXss(str) {
return str && str.replace(/[<>'"&]/g, '-');
}
_sort(folder) {
folder.sort((a, b) => {
if (a.type === b.type) return a.text === b.text ? 0 : a.text < b.text ? -1 : 1;
return a.type === 'blob' ? 1 : -1;
});
folder.forEach((item) => {
if (item.type === 'tree' && item.children !== true && item.children.length > 0) {
this._sort(item.children);
}
});
return folder;
}
_collapse(folder) {
return folder.map((item) => {
if (item.type === 'tree') {
item.children = this._collapse(item.children);
if (item.children.length === 1 && item.children[0].type === 'tree') {
const onlyChild = item.children[0];
onlyChild.text = item.text + '/' + onlyChild.text;
return onlyChild;
}
}
return item;
});
}
_onItemClick(event) {
let $target = $(event.target);
let download = false;
// Handle middle click
if (event.which === 2) return;
// Handle icon click, fix #122
if ($target.is('i.jstree-icon')) {
$target = $target.parent();
download = true;
}
if (!$target.is('a.jstree-anchor')) return;
// Refocus after complete so that keyboard navigation works, fix #158
const refocusAfterCompletion = () => {
$(document).one('pjax:success page:load', () => {
this.$jstree.get_container().focus();
});
};
const adapter = this.adapter;
const newTab = event.shiftKey || event.ctrlKey || event.metaKey;
const href = $target.attr('href');
// The 2nd path is for submodule child links
const $icon = $target.children().length ? $target.children(':first') : $target.siblings(':first');
if ($icon.hasClass('commit')) {
refocusAfterCompletion();
newTab ? adapter.openInNewTab(href) : adapter.selectSubmodule(href);
} else if ($icon.hasClass('blob')) {
if (download) {
const downloadUrl = $target.attr('data-download-url');
const downloadFileName = $target.attr('data-download-filename');
adapter.downloadFile(downloadUrl, downloadFileName);
} else {
refocusAfterCompletion();
newTab ? adapter.openInNewTab(href) : adapter.selectFile(href);
}
}
}
syncSelection(repo) {
const $jstree = this.$jstree;
if (!$jstree) return;
// Convert /username/reponame/object_type/branch/path to path
const path = decodeURIComponent(location.pathname);
const match = path.match(/(?:[^\/]+\/){4}(.*)/);
if (!match) return;
const currentPath = match[1];
const loadAll = this.adapter.canLoadEntireTree(repo) && this.store.get(STORE.LOADALL);
selectPath(loadAll ? [currentPath] : breakPath(currentPath));
// Convert ['a/b'] to ['a', 'a/b']
function breakPath(fullPath) {
return fullPath.split('/').reduce((res, path, idx) => {
res.push(idx === 0 ? path : `${res[idx - 1]}/${path}`);
return res;
}, []);
}
function selectPath(paths, index = 0) {
const nodeId = NODE_PREFIX + paths[index];
if ($jstree.get_node(nodeId)) |
}
}
}
| {
$jstree.deselect_all();
$jstree.select_node(nodeId);
$jstree.open_node(nodeId, () => {
if (++index < paths.length) {
selectPath(paths, index);
}
});
} | conditional_block |
view.tree.js | class TreeView {
constructor($dom, store, adapter) {
this.store = store;
this.adapter = adapter;
this.$view = $dom.find('.octotree_treeview');
this.$tree = this.$view
.find('.octotree_view_body')
.on('click.jstree', '.jstree-open>a', ({target}) => {
setTimeout(() => {
this.$jstree.close_node(target)
}, 0);
})
.on('click.jstree', '.jstree-closed>a', ({target}) => {
setTimeout(() => {
this.$jstree.open_node(target)
}, 0);
})
.on('click', this._onItemClick.bind(this))
.jstree({
core: {multiple: false, worker: false, themes: {responsive: false}},
plugins: ['wholerow']
});
}
get $jstree() {
return this.$tree.jstree(true);
}
show(repo, token) {
const $jstree = this.$jstree;
$jstree.settings.core.data = (node, cb) => {
const prMode = this.store.get(STORE.PR) && repo.pullNumber;
const loadAll = this.adapter.canLoadEntireTree(repo) && (prMode || this.store.get(STORE.LOADALL));
node = !loadAll && (node.id === '#' ? {path: ''} : node.original);
this.adapter.loadCodeTree({repo, token, node}, (err, treeData) => {
if (err) {
if (err.status === 206 && loadAll) { // The repo is too big to load all, need to retry
$jstree.refresh(true);
} else {
$(this).trigger(EVENT.FETCH_ERROR, [err]);
}
} else {
treeData = this._sort(treeData);
if (loadAll) {
treeData = this._collapse(treeData);
}
cb(treeData);
}
});
};
this.$tree.one('refresh.jstree', () => {
this.syncSelection(repo);
$(this).trigger(EVENT.VIEW_READY);
});
this._showHeader(repo);
$jstree.refresh(true);
}
_showHeader(repo) {
const adapter = this.adapter;
this.$view
.find('.octotree_view_header')
.html(
`<div class="octotree_header_repo">
<a href="/${repo.username}">${repo.username}</a>
/
<a data-pjax href="/${repo.username}/${repo.reponame}">${repo.reponame}</a>
</div>
<div class="octotree_header_branch">
${this._deXss(repo.branch.toString())}
</div>`
)
.on('click', 'a[data-pjax]', function(event) {
event.preventDefault();
// A.href always return absolute URL, don't want that
const href = $(this).attr('href');
const newTab = event.shiftKey || event.ctrlKey || event.metaKey;
newTab ? adapter.openInNewTab(href) : adapter.selectFile(href);
});
}
_deXss(str) {
return str && str.replace(/[<>'"&]/g, '-');
}
_sort(folder) |
_collapse(folder) {
return folder.map((item) => {
if (item.type === 'tree') {
item.children = this._collapse(item.children);
if (item.children.length === 1 && item.children[0].type === 'tree') {
const onlyChild = item.children[0];
onlyChild.text = item.text + '/' + onlyChild.text;
return onlyChild;
}
}
return item;
});
}
_onItemClick(event) {
let $target = $(event.target);
let download = false;
// Handle middle click
if (event.which === 2) return;
// Handle icon click, fix #122
if ($target.is('i.jstree-icon')) {
$target = $target.parent();
download = true;
}
if (!$target.is('a.jstree-anchor')) return;
// Refocus after complete so that keyboard navigation works, fix #158
const refocusAfterCompletion = () => {
$(document).one('pjax:success page:load', () => {
this.$jstree.get_container().focus();
});
};
const adapter = this.adapter;
const newTab = event.shiftKey || event.ctrlKey || event.metaKey;
const href = $target.attr('href');
// The 2nd path is for submodule child links
const $icon = $target.children().length ? $target.children(':first') : $target.siblings(':first');
if ($icon.hasClass('commit')) {
refocusAfterCompletion();
newTab ? adapter.openInNewTab(href) : adapter.selectSubmodule(href);
} else if ($icon.hasClass('blob')) {
if (download) {
const downloadUrl = $target.attr('data-download-url');
const downloadFileName = $target.attr('data-download-filename');
adapter.downloadFile(downloadUrl, downloadFileName);
} else {
refocusAfterCompletion();
newTab ? adapter.openInNewTab(href) : adapter.selectFile(href);
}
}
}
syncSelection(repo) {
const $jstree = this.$jstree;
if (!$jstree) return;
// Convert /username/reponame/object_type/branch/path to path
const path = decodeURIComponent(location.pathname);
const match = path.match(/(?:[^\/]+\/){4}(.*)/);
if (!match) return;
const currentPath = match[1];
const loadAll = this.adapter.canLoadEntireTree(repo) && this.store.get(STORE.LOADALL);
selectPath(loadAll ? [currentPath] : breakPath(currentPath));
// Convert ['a/b'] to ['a', 'a/b']
function breakPath(fullPath) {
return fullPath.split('/').reduce((res, path, idx) => {
res.push(idx === 0 ? path : `${res[idx - 1]}/${path}`);
return res;
}, []);
}
function selectPath(paths, index = 0) {
const nodeId = NODE_PREFIX + paths[index];
if ($jstree.get_node(nodeId)) {
$jstree.deselect_all();
$jstree.select_node(nodeId);
$jstree.open_node(nodeId, () => {
if (++index < paths.length) {
selectPath(paths, index);
}
});
}
}
}
}
| {
folder.sort((a, b) => {
if (a.type === b.type) return a.text === b.text ? 0 : a.text < b.text ? -1 : 1;
return a.type === 'blob' ? 1 : -1;
});
folder.forEach((item) => {
if (item.type === 'tree' && item.children !== true && item.children.length > 0) {
this._sort(item.children);
}
});
return folder;
} | identifier_body |
statistic_spec.ts | /**
* @license
* Copyright Google Inc. All Rights Reserved.
*
* Use of this source code is governed by an MIT-style license that can be
* found in the LICENSE file at https://angular.io/license
*/
import {AsyncTestCompleter, afterEach, beforeEach, ddescribe, describe, expect, iit, inject, it, xit} from '@angular/core/testing/testing_internal';
import {Statistic} from 'benchpress/src/statistic';
export function | () {
describe('statistic', () => {
it('should calculate the mean', () => {
expect(Statistic.calculateMean([])).toBeNaN();
expect(Statistic.calculateMean([1, 2, 3])).toBe(2.0);
});
it('should calculate the standard deviation', () => {
expect(Statistic.calculateStandardDeviation([], NaN)).toBeNaN();
expect(Statistic.calculateStandardDeviation([1], 1)).toBe(0.0);
expect(Statistic.calculateStandardDeviation([2, 4, 4, 4, 5, 5, 7, 9], 5)).toBe(2.0);
});
it('should calculate the coefficient of variation', () => {
expect(Statistic.calculateCoefficientOfVariation([], NaN)).toBeNaN();
expect(Statistic.calculateCoefficientOfVariation([1], 1)).toBe(0.0);
expect(Statistic.calculateCoefficientOfVariation([2, 4, 4, 4, 5, 5, 7, 9], 5)).toBe(40.0);
});
it('should calculate the regression slope', () => {
expect(Statistic.calculateRegressionSlope([], NaN, [], NaN)).toBeNaN();
expect(Statistic.calculateRegressionSlope([1], 1, [2], 2)).toBeNaN();
expect(Statistic.calculateRegressionSlope([1, 2], 1.5, [2, 4], 3)).toBe(2.0);
});
});
}
| main | identifier_name |
statistic_spec.ts | /**
* @license
* Copyright Google Inc. All Rights Reserved.
*
* Use of this source code is governed by an MIT-style license that can be
* found in the LICENSE file at https://angular.io/license
*/
import {AsyncTestCompleter, afterEach, beforeEach, ddescribe, describe, expect, iit, inject, it, xit} from '@angular/core/testing/testing_internal';
import {Statistic} from 'benchpress/src/statistic';
export function main() {
describe('statistic', () => {
it('should calculate the mean', () => {
expect(Statistic.calculateMean([])).toBeNaN();
expect(Statistic.calculateMean([1, 2, 3])).toBe(2.0);
});
it('should calculate the standard deviation', () => {
expect(Statistic.calculateStandardDeviation([], NaN)).toBeNaN(); | expect(Statistic.calculateStandardDeviation([2, 4, 4, 4, 5, 5, 7, 9], 5)).toBe(2.0);
});
it('should calculate the coefficient of variation', () => {
expect(Statistic.calculateCoefficientOfVariation([], NaN)).toBeNaN();
expect(Statistic.calculateCoefficientOfVariation([1], 1)).toBe(0.0);
expect(Statistic.calculateCoefficientOfVariation([2, 4, 4, 4, 5, 5, 7, 9], 5)).toBe(40.0);
});
it('should calculate the regression slope', () => {
expect(Statistic.calculateRegressionSlope([], NaN, [], NaN)).toBeNaN();
expect(Statistic.calculateRegressionSlope([1], 1, [2], 2)).toBeNaN();
expect(Statistic.calculateRegressionSlope([1, 2], 1.5, [2, 4], 3)).toBe(2.0);
});
});
} | expect(Statistic.calculateStandardDeviation([1], 1)).toBe(0.0); | random_line_split |
statistic_spec.ts | /**
* @license
* Copyright Google Inc. All Rights Reserved.
*
* Use of this source code is governed by an MIT-style license that can be
* found in the LICENSE file at https://angular.io/license
*/
import {AsyncTestCompleter, afterEach, beforeEach, ddescribe, describe, expect, iit, inject, it, xit} from '@angular/core/testing/testing_internal';
import {Statistic} from 'benchpress/src/statistic';
export function main() | it('should calculate the regression slope', () => {
expect(Statistic.calculateRegressionSlope([], NaN, [], NaN)).toBeNaN();
expect(Statistic.calculateRegressionSlope([1], 1, [2], 2)).toBeNaN();
expect(Statistic.calculateRegressionSlope([1, 2], 1.5, [2, 4], 3)).toBe(2.0);
});
});
}
| {
describe('statistic', () => {
it('should calculate the mean', () => {
expect(Statistic.calculateMean([])).toBeNaN();
expect(Statistic.calculateMean([1, 2, 3])).toBe(2.0);
});
it('should calculate the standard deviation', () => {
expect(Statistic.calculateStandardDeviation([], NaN)).toBeNaN();
expect(Statistic.calculateStandardDeviation([1], 1)).toBe(0.0);
expect(Statistic.calculateStandardDeviation([2, 4, 4, 4, 5, 5, 7, 9], 5)).toBe(2.0);
});
it('should calculate the coefficient of variation', () => {
expect(Statistic.calculateCoefficientOfVariation([], NaN)).toBeNaN();
expect(Statistic.calculateCoefficientOfVariation([1], 1)).toBe(0.0);
expect(Statistic.calculateCoefficientOfVariation([2, 4, 4, 4, 5, 5, 7, 9], 5)).toBe(40.0);
});
| identifier_body |
edit.state.d.ts | import { Command } from '../command/command';
/**
* Property that controls grid edit unit.
*
* * `'cell'` data is editable through the grid cells.
* * `'row'` data is editable through the grid rows.
* * `'null'` data is not editable.
*/
export declare type EditStateMode = null | 'cell' | 'row';
/**
* Indicates if q-grid is in `'edit'` or in a `'view'` mode.
*/
export declare type EditStateStatus = 'edit' | 'view' | 'startBatch' | 'endBatch';
/**
* Property that controls grid edit behavior.
*
* * `'batch'` batch update.
*/
export declare type EditStateMethod = null | 'batch';
/**
* A class represent options to control q-grid edit mode.
*/
export declare class EditState {
/**
* Property that controls grid edit unit.
*/
mode: EditStateMode;
/**
* Indicates if q-grid is in `'edit'` or in a `'view'` mode.
*/
status: EditStateStatus;
/**
* Property that controls grid edit behavior.
*/
method: EditStateMethod;
/**
* Allows to the grid user to control if cell or row can be edited or not.
*/
enter: Command;
/**
* Allows to the grid user to control if new cell value can be stored in data source or not.
*/
commit: Command;
/**
* Allows to the grid user to control if cell can exit edit mode.
*/
cancel: Command;
/**
* Allows to the grid user to control if cell can exit edit mode.
*/
reset: Command;
/**
* Allows to the grid user to manage clear action behavior in edit mode.
*/ | /**
* Object that contains `{columnKey: keyboardKeys}` map, that is used by q-grid to manage
* when cancel command should be execute on key down event.
*/
cancelShortcuts: { [key: string]: string };
/**
* Object that contains `{columnKey: keyboardKeys}` map, that is used by q-grid to manage
* when enter command should be execute on key down event.
*/
enterShortcuts: { [key: string]: string };
/**
* Object that contains `{columnKey: keyboardKeys}` map, that is used by q-grid to manage
* when commit command should be execute on key down event.
*/
commitShortcuts: { [key: string]: string };
} | clear: Command;
| random_line_split |
edit.state.d.ts | import { Command } from '../command/command';
/**
* Property that controls grid edit unit.
*
* * `'cell'` data is editable through the grid cells.
* * `'row'` data is editable through the grid rows.
* * `'null'` data is not editable.
*/
export declare type EditStateMode = null | 'cell' | 'row';
/**
* Indicates if q-grid is in `'edit'` or in a `'view'` mode.
*/
export declare type EditStateStatus = 'edit' | 'view' | 'startBatch' | 'endBatch';
/**
* Property that controls grid edit behavior.
*
* * `'batch'` batch update.
*/
export declare type EditStateMethod = null | 'batch';
/**
* A class represent options to control q-grid edit mode.
*/
export declare class | {
/**
* Property that controls grid edit unit.
*/
mode: EditStateMode;
/**
* Indicates if q-grid is in `'edit'` or in a `'view'` mode.
*/
status: EditStateStatus;
/**
* Property that controls grid edit behavior.
*/
method: EditStateMethod;
/**
* Allows to the grid user to control if cell or row can be edited or not.
*/
enter: Command;
/**
* Allows to the grid user to control if new cell value can be stored in data source or not.
*/
commit: Command;
/**
* Allows to the grid user to control if cell can exit edit mode.
*/
cancel: Command;
/**
* Allows to the grid user to control if cell can exit edit mode.
*/
reset: Command;
/**
* Allows to the grid user to manage clear action behavior in edit mode.
*/
clear: Command;
/**
* Object that contains `{columnKey: keyboardKeys}` map, that is used by q-grid to manage
* when cancel command should be execute on key down event.
*/
cancelShortcuts: { [key: string]: string };
/**
* Object that contains `{columnKey: keyboardKeys}` map, that is used by q-grid to manage
* when enter command should be execute on key down event.
*/
enterShortcuts: { [key: string]: string };
/**
* Object that contains `{columnKey: keyboardKeys}` map, that is used by q-grid to manage
* when commit command should be execute on key down event.
*/
commitShortcuts: { [key: string]: string };
}
| EditState | identifier_name |
loaders.py | """
Test cases for the template loaders
Note: This test requires setuptools!
"""
from django.conf import settings
if __name__ == '__main__':
settings.configure()
import unittest
import sys
import pkg_resources
import imp
import StringIO
import os.path
from django.template import TemplateDoesNotExist
from django.template.loaders.eggs import load_template_source as lts_egg
# Mock classes and objects for pkg_resources functions.
class MockProvider(pkg_resources.NullProvider):
def __init__(self, module):
pkg_resources.NullProvider.__init__(self, module)
self.module = module
def _has(self, path):
return path in self.module._resources
def _isdir(self,path):
return False
def get_resource_stream(self, manager, resource_name):
return self.module._resources[resource_name]
def _get(self, path):
return self.module._resources[path].read()
class MockLoader(object):
pass
def create_egg(name, resources):
"""
Creates a mock egg with a list of resources.
name: The name of the module.
resources: A dictionary of resources. Keys are the names and values the data.
"""
egg = imp.new_module(name)
egg.__loader__ = MockLoader()
egg._resources = resources
sys.modules[name] = egg
class EggLoader(unittest.TestCase):
def setUp(self):
pkg_resources._provider_factories[MockLoader] = MockProvider
self.empty_egg = create_egg("egg_empty", {})
self.egg_1 = create_egg("egg_1", {
os.path.normcase('templates/y.html') : StringIO.StringIO("y"),
os.path.normcase('templates/x.txt') : StringIO.StringIO("x"),
})
self._old_installed_apps = settings.INSTALLED_APPS
settings.INSTALLED_APPS = []
def tearDown(self):
settings.INSTALLED_APPS = self._old_installed_apps
def test_empty(self):
"Loading any template on an empty egg should fail"
settings.INSTALLED_APPS = ['egg_empty']
self.assertRaises(TemplateDoesNotExist, lts_egg, "not-existing.html")
def test_non_existing(self):
"Template loading fails if the template is not in the egg"
settings.INSTALLED_APPS = ['egg_1']
self.assertRaises(TemplateDoesNotExist, lts_egg, "not-existing.html")
def test_existing(self):
"A template can be loaded from an egg"
settings.INSTALLED_APPS = ['egg_1']
contents, template_name = lts_egg("y.html")
self.assertEqual(contents, "y")
self.assertEqual(template_name, "egg:egg_1:templates/y.html")
def test_not_installed(self): | unittest.main() | "Loading an existent template from an egg not included in INSTALLED_APPS should fail"
settings.INSTALLED_APPS = []
self.assertRaises(TemplateDoesNotExist, lts_egg, "y.html")
if __name__ == "__main__": | random_line_split |
loaders.py | """
Test cases for the template loaders
Note: This test requires setuptools!
"""
from django.conf import settings
if __name__ == '__main__':
settings.configure()
import unittest
import sys
import pkg_resources
import imp
import StringIO
import os.path
from django.template import TemplateDoesNotExist
from django.template.loaders.eggs import load_template_source as lts_egg
# Mock classes and objects for pkg_resources functions.
class MockProvider(pkg_resources.NullProvider):
def __init__(self, module):
pkg_resources.NullProvider.__init__(self, module)
self.module = module
def _has(self, path):
return path in self.module._resources
def _isdir(self,path):
return False
def get_resource_stream(self, manager, resource_name):
return self.module._resources[resource_name]
def _get(self, path):
return self.module._resources[path].read()
class MockLoader(object):
pass
def create_egg(name, resources):
"""
Creates a mock egg with a list of resources.
name: The name of the module.
resources: A dictionary of resources. Keys are the names and values the data.
"""
egg = imp.new_module(name)
egg.__loader__ = MockLoader()
egg._resources = resources
sys.modules[name] = egg
class EggLoader(unittest.TestCase):
def setUp(self):
pkg_resources._provider_factories[MockLoader] = MockProvider
self.empty_egg = create_egg("egg_empty", {})
self.egg_1 = create_egg("egg_1", {
os.path.normcase('templates/y.html') : StringIO.StringIO("y"),
os.path.normcase('templates/x.txt') : StringIO.StringIO("x"),
})
self._old_installed_apps = settings.INSTALLED_APPS
settings.INSTALLED_APPS = []
def | (self):
settings.INSTALLED_APPS = self._old_installed_apps
def test_empty(self):
"Loading any template on an empty egg should fail"
settings.INSTALLED_APPS = ['egg_empty']
self.assertRaises(TemplateDoesNotExist, lts_egg, "not-existing.html")
def test_non_existing(self):
"Template loading fails if the template is not in the egg"
settings.INSTALLED_APPS = ['egg_1']
self.assertRaises(TemplateDoesNotExist, lts_egg, "not-existing.html")
def test_existing(self):
"A template can be loaded from an egg"
settings.INSTALLED_APPS = ['egg_1']
contents, template_name = lts_egg("y.html")
self.assertEqual(contents, "y")
self.assertEqual(template_name, "egg:egg_1:templates/y.html")
def test_not_installed(self):
"Loading an existent template from an egg not included in INSTALLED_APPS should fail"
settings.INSTALLED_APPS = []
self.assertRaises(TemplateDoesNotExist, lts_egg, "y.html")
if __name__ == "__main__":
unittest.main()
| tearDown | identifier_name |
loaders.py | """
Test cases for the template loaders
Note: This test requires setuptools!
"""
from django.conf import settings
if __name__ == '__main__':
settings.configure()
import unittest
import sys
import pkg_resources
import imp
import StringIO
import os.path
from django.template import TemplateDoesNotExist
from django.template.loaders.eggs import load_template_source as lts_egg
# Mock classes and objects for pkg_resources functions.
class MockProvider(pkg_resources.NullProvider):
def __init__(self, module):
pkg_resources.NullProvider.__init__(self, module)
self.module = module
def _has(self, path):
return path in self.module._resources
def _isdir(self,path):
return False
def get_resource_stream(self, manager, resource_name):
return self.module._resources[resource_name]
def _get(self, path):
return self.module._resources[path].read()
class MockLoader(object):
pass
def create_egg(name, resources):
"""
Creates a mock egg with a list of resources.
name: The name of the module.
resources: A dictionary of resources. Keys are the names and values the data.
"""
egg = imp.new_module(name)
egg.__loader__ = MockLoader()
egg._resources = resources
sys.modules[name] = egg
class EggLoader(unittest.TestCase):
def setUp(self):
pkg_resources._provider_factories[MockLoader] = MockProvider
self.empty_egg = create_egg("egg_empty", {})
self.egg_1 = create_egg("egg_1", {
os.path.normcase('templates/y.html') : StringIO.StringIO("y"),
os.path.normcase('templates/x.txt') : StringIO.StringIO("x"),
})
self._old_installed_apps = settings.INSTALLED_APPS
settings.INSTALLED_APPS = []
def tearDown(self):
|
def test_empty(self):
"Loading any template on an empty egg should fail"
settings.INSTALLED_APPS = ['egg_empty']
self.assertRaises(TemplateDoesNotExist, lts_egg, "not-existing.html")
def test_non_existing(self):
"Template loading fails if the template is not in the egg"
settings.INSTALLED_APPS = ['egg_1']
self.assertRaises(TemplateDoesNotExist, lts_egg, "not-existing.html")
def test_existing(self):
"A template can be loaded from an egg"
settings.INSTALLED_APPS = ['egg_1']
contents, template_name = lts_egg("y.html")
self.assertEqual(contents, "y")
self.assertEqual(template_name, "egg:egg_1:templates/y.html")
def test_not_installed(self):
"Loading an existent template from an egg not included in INSTALLED_APPS should fail"
settings.INSTALLED_APPS = []
self.assertRaises(TemplateDoesNotExist, lts_egg, "y.html")
if __name__ == "__main__":
unittest.main()
| settings.INSTALLED_APPS = self._old_installed_apps | identifier_body |
loaders.py | """
Test cases for the template loaders
Note: This test requires setuptools!
"""
from django.conf import settings
if __name__ == '__main__':
settings.configure()
import unittest
import sys
import pkg_resources
import imp
import StringIO
import os.path
from django.template import TemplateDoesNotExist
from django.template.loaders.eggs import load_template_source as lts_egg
# Mock classes and objects for pkg_resources functions.
class MockProvider(pkg_resources.NullProvider):
def __init__(self, module):
pkg_resources.NullProvider.__init__(self, module)
self.module = module
def _has(self, path):
return path in self.module._resources
def _isdir(self,path):
return False
def get_resource_stream(self, manager, resource_name):
return self.module._resources[resource_name]
def _get(self, path):
return self.module._resources[path].read()
class MockLoader(object):
pass
def create_egg(name, resources):
"""
Creates a mock egg with a list of resources.
name: The name of the module.
resources: A dictionary of resources. Keys are the names and values the data.
"""
egg = imp.new_module(name)
egg.__loader__ = MockLoader()
egg._resources = resources
sys.modules[name] = egg
class EggLoader(unittest.TestCase):
def setUp(self):
pkg_resources._provider_factories[MockLoader] = MockProvider
self.empty_egg = create_egg("egg_empty", {})
self.egg_1 = create_egg("egg_1", {
os.path.normcase('templates/y.html') : StringIO.StringIO("y"),
os.path.normcase('templates/x.txt') : StringIO.StringIO("x"),
})
self._old_installed_apps = settings.INSTALLED_APPS
settings.INSTALLED_APPS = []
def tearDown(self):
settings.INSTALLED_APPS = self._old_installed_apps
def test_empty(self):
"Loading any template on an empty egg should fail"
settings.INSTALLED_APPS = ['egg_empty']
self.assertRaises(TemplateDoesNotExist, lts_egg, "not-existing.html")
def test_non_existing(self):
"Template loading fails if the template is not in the egg"
settings.INSTALLED_APPS = ['egg_1']
self.assertRaises(TemplateDoesNotExist, lts_egg, "not-existing.html")
def test_existing(self):
"A template can be loaded from an egg"
settings.INSTALLED_APPS = ['egg_1']
contents, template_name = lts_egg("y.html")
self.assertEqual(contents, "y")
self.assertEqual(template_name, "egg:egg_1:templates/y.html")
def test_not_installed(self):
"Loading an existent template from an egg not included in INSTALLED_APPS should fail"
settings.INSTALLED_APPS = []
self.assertRaises(TemplateDoesNotExist, lts_egg, "y.html")
if __name__ == "__main__":
| unittest.main() | conditional_block |
|
gui.js | var gui={};
(function() {
try {
var open = require("open");
var fs = require("fs");
} catch(e) {
var open = function(path) {
window.open("file://"+path);
}
}
var state = document.getElementById("state"),
statemsg = document.getElementById("statemsg"),
progress = document.getElementById("progress"),
pickdir = document.getElementById("pickdir"),
table = document.getElementById("collisions_table").tBodies[0];
pickdir.addEventListener("click", function(){
var fc = document.createElement("input");
fc.type = "file";
fc.value = "";
fc.nwdirectory = true;
fc.multiple = true;
state.classList.remove("hidden");
fc.onchange = function() {
analyze_dir(fc.value);
}
fc.click();
}, true);
gui.update_progress = function(rate) {
progress.value = rate;
};
gui.set_statemsg = function (msg) {
statemsg.innerHTML = msg;
};
gui.analyze_authorized = function (auth) {
pickdir.disabled = auth;
};
function readableSize (size) |
function insert_collision (idx, files, dist) {
var row = table.insertRow(idx);
row.dataset["dist"] = dist;
for (var i=0; i<2; i++) {
var cell = row.insertCell(i);
var pathElem = document.createTextNode(files[i].dirname+"/");
var fileNameElem = document.createElement("b");
var sizeElem = document.createElement("i");
var deleteBtn = document.createElement("button");
cell.dataset["filepath"] = files[i].filepath;
fileNameElem.addEventListener("click",function(e) {
var path = e.target.parentElement.dataset["filepath"];
open(path);
}, true);
fileNameElem.textContent = files[i].stats.name;
deleteBtn.textContent = "delete";
deleteBtn.addEventListener("click",function(e) {
var path = e.target.parentElement.dataset["filepath"];
if (confirm("Delete "+path+"?")) {
fs.unlink(path, function (err) {
if (err) {
alert("Unable to delete "+path);
} else {
var row = e.target.parentElement.parentElement;
row.parentElement.removeChild(row);
}
});
}
}, true);
sizeElem.textContent = readableSize(files[i].stats.size);
cell.appendChild(pathElem);
cell.appendChild(fileNameElem);
cell.appendChild(sizeElem);
cell.appendChild(deleteBtn);
}
cell = row.insertCell(2);
cell.textContent = dist;
};
gui.display_collision = function (files, dist) {
for (var idx=0; idx < table.rows.length; idx++) { //May not be necessary to do a dichotomy
if (table.rows[idx].dataset["dist"] >= dist) break;
}
insert_collision(idx, files, dist);
};
gui.init_display_collisions = function() {
table.parentElement.classList.remove("hidden");
table.innerHTML = "";
};
gui.all_collisions_displayed = function (ndoublets) {
gui.set_statemsg(ndoublets + " collisions found");
gui.update_progress(1);
};
})();
| {
if (size > 1e9) return ((size/1e8|0)/10) + " Gb";
else if (size > 1e6) return ((size/1e5|0)/10) + " Mb";
else if (size > 1e3) return ((size/1e2|0)/10) + " Kb";
else return size+" bytes";
} | identifier_body |
gui.js | var gui={};
(function() {
try {
var open = require("open");
var fs = require("fs");
} catch(e) {
var open = function(path) {
window.open("file://"+path);
}
}
var state = document.getElementById("state"),
statemsg = document.getElementById("statemsg"),
progress = document.getElementById("progress"),
pickdir = document.getElementById("pickdir"),
table = document.getElementById("collisions_table").tBodies[0];
pickdir.addEventListener("click", function(){
var fc = document.createElement("input");
fc.type = "file";
fc.value = "";
fc.nwdirectory = true;
fc.multiple = true;
state.classList.remove("hidden");
fc.onchange = function() {
analyze_dir(fc.value);
}
fc.click();
}, true);
gui.update_progress = function(rate) {
progress.value = rate;
};
gui.set_statemsg = function (msg) {
statemsg.innerHTML = msg;
};
gui.analyze_authorized = function (auth) {
pickdir.disabled = auth;
};
function readableSize (size) {
if (size > 1e9) return ((size/1e8|0)/10) + " Gb";
else if (size > 1e6) return ((size/1e5|0)/10) + " Mb";
else if (size > 1e3) return ((size/1e2|0)/10) + " Kb";
else return size+" bytes";
}
function insert_collision (idx, files, dist) {
var row = table.insertRow(idx);
row.dataset["dist"] = dist;
for (var i=0; i<2; i++) {
var cell = row.insertCell(i);
var pathElem = document.createTextNode(files[i].dirname+"/");
var fileNameElem = document.createElement("b");
var sizeElem = document.createElement("i");
var deleteBtn = document.createElement("button"); | fileNameElem.textContent = files[i].stats.name;
deleteBtn.textContent = "delete";
deleteBtn.addEventListener("click",function(e) {
var path = e.target.parentElement.dataset["filepath"];
if (confirm("Delete "+path+"?")) {
fs.unlink(path, function (err) {
if (err) {
alert("Unable to delete "+path);
} else {
var row = e.target.parentElement.parentElement;
row.parentElement.removeChild(row);
}
});
}
}, true);
sizeElem.textContent = readableSize(files[i].stats.size);
cell.appendChild(pathElem);
cell.appendChild(fileNameElem);
cell.appendChild(sizeElem);
cell.appendChild(deleteBtn);
}
cell = row.insertCell(2);
cell.textContent = dist;
};
gui.display_collision = function (files, dist) {
for (var idx=0; idx < table.rows.length; idx++) { //May not be necessary to do a dichotomy
if (table.rows[idx].dataset["dist"] >= dist) break;
}
insert_collision(idx, files, dist);
};
gui.init_display_collisions = function() {
table.parentElement.classList.remove("hidden");
table.innerHTML = "";
};
gui.all_collisions_displayed = function (ndoublets) {
gui.set_statemsg(ndoublets + " collisions found");
gui.update_progress(1);
};
})(); | cell.dataset["filepath"] = files[i].filepath;
fileNameElem.addEventListener("click",function(e) {
var path = e.target.parentElement.dataset["filepath"];
open(path);
}, true); | random_line_split |
gui.js | var gui={};
(function() {
try {
var open = require("open");
var fs = require("fs");
} catch(e) {
var open = function(path) {
window.open("file://"+path);
}
}
var state = document.getElementById("state"),
statemsg = document.getElementById("statemsg"),
progress = document.getElementById("progress"),
pickdir = document.getElementById("pickdir"),
table = document.getElementById("collisions_table").tBodies[0];
pickdir.addEventListener("click", function(){
var fc = document.createElement("input");
fc.type = "file";
fc.value = "";
fc.nwdirectory = true;
fc.multiple = true;
state.classList.remove("hidden");
fc.onchange = function() {
analyze_dir(fc.value);
}
fc.click();
}, true);
gui.update_progress = function(rate) {
progress.value = rate;
};
gui.set_statemsg = function (msg) {
statemsg.innerHTML = msg;
};
gui.analyze_authorized = function (auth) {
pickdir.disabled = auth;
};
function | (size) {
if (size > 1e9) return ((size/1e8|0)/10) + " Gb";
else if (size > 1e6) return ((size/1e5|0)/10) + " Mb";
else if (size > 1e3) return ((size/1e2|0)/10) + " Kb";
else return size+" bytes";
}
function insert_collision (idx, files, dist) {
var row = table.insertRow(idx);
row.dataset["dist"] = dist;
for (var i=0; i<2; i++) {
var cell = row.insertCell(i);
var pathElem = document.createTextNode(files[i].dirname+"/");
var fileNameElem = document.createElement("b");
var sizeElem = document.createElement("i");
var deleteBtn = document.createElement("button");
cell.dataset["filepath"] = files[i].filepath;
fileNameElem.addEventListener("click",function(e) {
var path = e.target.parentElement.dataset["filepath"];
open(path);
}, true);
fileNameElem.textContent = files[i].stats.name;
deleteBtn.textContent = "delete";
deleteBtn.addEventListener("click",function(e) {
var path = e.target.parentElement.dataset["filepath"];
if (confirm("Delete "+path+"?")) {
fs.unlink(path, function (err) {
if (err) {
alert("Unable to delete "+path);
} else {
var row = e.target.parentElement.parentElement;
row.parentElement.removeChild(row);
}
});
}
}, true);
sizeElem.textContent = readableSize(files[i].stats.size);
cell.appendChild(pathElem);
cell.appendChild(fileNameElem);
cell.appendChild(sizeElem);
cell.appendChild(deleteBtn);
}
cell = row.insertCell(2);
cell.textContent = dist;
};
gui.display_collision = function (files, dist) {
for (var idx=0; idx < table.rows.length; idx++) { //May not be necessary to do a dichotomy
if (table.rows[idx].dataset["dist"] >= dist) break;
}
insert_collision(idx, files, dist);
};
gui.init_display_collisions = function() {
table.parentElement.classList.remove("hidden");
table.innerHTML = "";
};
gui.all_collisions_displayed = function (ndoublets) {
gui.set_statemsg(ndoublets + " collisions found");
gui.update_progress(1);
};
})();
| readableSize | identifier_name |
HumidTemp.js | const sensor = require('node-dht-sensor');
const logger = require('../logging/Logger');
/**
* Reads pin 4 of the raspberry PI to obtain temperature and humidity information.
* @return {Promise} A promise that will resolve with the results. In the
* case where there was an error reading, will return a zero filled object,
* with an additional error field.
* { temperature: Number,
* humidity: Number,
* error: Error|undefined }
*/
exports.getHumidityTemperature = function() {
return new Promise( (resolve, reject) => {
sensor.read(22, 4, (err, temperature, humidity) => {
if(err) |
resolve({
temperature: temperature * 1.8 + 32,
humidity: humidity});
});
});
}
| {
logger.error("Could not read from the DHT sensor. " + err);
return resolve({
temperature: 0,
humidity: 0,
error: err});
} | conditional_block |
HumidTemp.js | const sensor = require('node-dht-sensor');
const logger = require('../logging/Logger');
/**
* Reads pin 4 of the raspberry PI to obtain temperature and humidity information.
* @return {Promise} A promise that will resolve with the results. In the
* case where there was an error reading, will return a zero filled object,
* with an additional error field.
* { temperature: Number,
* humidity: Number,
* error: Error|undefined }
*/
exports.getHumidityTemperature = function() {
return new Promise( (resolve, reject) => {
sensor.read(22, 4, (err, temperature, humidity) => {
if(err) {
logger.error("Could not read from the DHT sensor. " + err);
return resolve({
temperature: 0,
humidity: 0,
error: err});
} | resolve({
temperature: temperature * 1.8 + 32,
humidity: humidity});
});
});
} | random_line_split |
|
issue-17441.rs | // Copyright 2014 The Rust Project Developers. See the COPYRIGHT
// file at the top-level directory of this distribution and at
// http://rust-lang.org/COPYRIGHT.
//
// Licensed under the Apache License, Version 2.0 <LICENSE-APACHE or
// http://www.apache.org/licenses/LICENSE-2.0> or the MIT license
// <LICENSE-MIT or http://opensource.org/licenses/MIT>, at your
// option. This file may not be copied, modified, or distributed
// except according to those terms.
fn | () {
let _foo = &[1_usize, 2] as [usize];
//~^ ERROR cast to unsized type: `&[usize; 2]` as `[usize]`
//~^^ HELP consider using an implicit coercion to `&[usize]` instead
// FIXME (#22405): Replace `Box::new` with `box` here when/if possible.
let _bar = Box::new(1_usize) as std::fmt::Debug;
//~^ ERROR cast to unsized type: `Box<usize>` as `core::fmt::Debug`
//~^^ HELP did you mean `Box<core::fmt::Debug>`?
let _baz = 1_usize as std::fmt::Debug;
//~^ ERROR cast to unsized type: `usize` as `core::fmt::Debug`
//~^^ HELP consider using a box or reference as appropriate
let _quux = [1_usize, 2] as [usize];
//~^ ERROR cast to unsized type: `[usize; 2]` as `[usize]`
//~^^ HELP consider using a box or reference as appropriate
}
| main | identifier_name |
issue-17441.rs | // Copyright 2014 The Rust Project Developers. See the COPYRIGHT
// file at the top-level directory of this distribution and at
// http://rust-lang.org/COPYRIGHT.
//
// Licensed under the Apache License, Version 2.0 <LICENSE-APACHE or
// http://www.apache.org/licenses/LICENSE-2.0> or the MIT license
// <LICENSE-MIT or http://opensource.org/licenses/MIT>, at your
// option. This file may not be copied, modified, or distributed
// except according to those terms.
fn main() | {
let _foo = &[1_usize, 2] as [usize];
//~^ ERROR cast to unsized type: `&[usize; 2]` as `[usize]`
//~^^ HELP consider using an implicit coercion to `&[usize]` instead
// FIXME (#22405): Replace `Box::new` with `box` here when/if possible.
let _bar = Box::new(1_usize) as std::fmt::Debug;
//~^ ERROR cast to unsized type: `Box<usize>` as `core::fmt::Debug`
//~^^ HELP did you mean `Box<core::fmt::Debug>`?
let _baz = 1_usize as std::fmt::Debug;
//~^ ERROR cast to unsized type: `usize` as `core::fmt::Debug`
//~^^ HELP consider using a box or reference as appropriate
let _quux = [1_usize, 2] as [usize];
//~^ ERROR cast to unsized type: `[usize; 2]` as `[usize]`
//~^^ HELP consider using a box or reference as appropriate
} | identifier_body |
|
issue-17441.rs | // Copyright 2014 The Rust Project Developers. See the COPYRIGHT
// file at the top-level directory of this distribution and at
// http://rust-lang.org/COPYRIGHT.
// |
fn main() {
let _foo = &[1_usize, 2] as [usize];
//~^ ERROR cast to unsized type: `&[usize; 2]` as `[usize]`
//~^^ HELP consider using an implicit coercion to `&[usize]` instead
// FIXME (#22405): Replace `Box::new` with `box` here when/if possible.
let _bar = Box::new(1_usize) as std::fmt::Debug;
//~^ ERROR cast to unsized type: `Box<usize>` as `core::fmt::Debug`
//~^^ HELP did you mean `Box<core::fmt::Debug>`?
let _baz = 1_usize as std::fmt::Debug;
//~^ ERROR cast to unsized type: `usize` as `core::fmt::Debug`
//~^^ HELP consider using a box or reference as appropriate
let _quux = [1_usize, 2] as [usize];
//~^ ERROR cast to unsized type: `[usize; 2]` as `[usize]`
//~^^ HELP consider using a box or reference as appropriate
} | // Licensed under the Apache License, Version 2.0 <LICENSE-APACHE or
// http://www.apache.org/licenses/LICENSE-2.0> or the MIT license
// <LICENSE-MIT or http://opensource.org/licenses/MIT>, at your
// option. This file may not be copied, modified, or distributed
// except according to those terms. | random_line_split |
main.ts | class | extends Rf.ETS.FrameWork.GameMain
{
private group: Rf.ETS.FrameWork.Group = null;
private sprite: Rf.ETS.FrameWork.Sprite = null;
private touchCharactor: Rf.ETS.FrameWork.Character = null;
private touchCharactorTouchPosX:number = 0;
private touchCharactorTouchPosY:number = 0;
/**
* 初期化イベント
* @method
* @name FrameWork.GameMain#onInitialize
*/
protected onInitialize():void
{
//サイズを640x640に変更
this.screenHeight = 640;
//fpsを10から30に変更
this.fps = 30;
}
/**
* リソース設定イベント
* @method
* @name FrameWork.GameMain#resourceLoad
*/
protected onResourceSetting():void
{
this.resourceManager.SetResourcePath("./assets/resources/");
this.resourceManager.AddResourceName("charaImage", "chara.png");
}
/**
* ロードイベント
* @method
* @name FrameWork.GameMain#onLoad
* @param {Object} parent - 親Group
*/
protected onLoad(parent: enchant.Group):void
{
//グループインスタンス作成
this.group = new Rf.ETS.FrameWork.Group(parent);
this.group.y = 100;
//画像無しスプライトインスタンス作成
let surface: Rf.ETS.FrameWork.NoImageSprite =
new Rf.ETS.FrameWork.NoImageSprite(100, 100, this.group);
surface.SetSurface("rgb(128,255,255)");
surface.opacity = 0.5;
//スプライトインスタンス作成
this.sprite = new Rf.ETS.FrameWork.Sprite(32, 32, this.group);
this.sprite.FileName = this.resourceManager.GetResourceName("charaImage");
this.sprite.originX = 16; //中心で回転するように設定
this.sprite.originY = 16; //中心で回転するように設定
this.sprite.frame = 26*2; //サンプル画像で正面画像を表示する
//タッチイベント用キャラクタ
this.touchCharactor = new Rf.ETS.FrameWork.Character(32,32,parent);
this.touchCharactor.FileName = this.resourceManager.GetResourceName("charaImage");
this.touchCharactor.charaIndex = 3;
this.touchCharactor.Dir = Rf.ETS.FrameWork.Direction.Up;
this.touchCharactor.x = 32;
this.touchCharactor.y = 32;
this.touchCharactor.originX = 16*2;
this.touchCharactor.originY = 16*2;
this.touchCharactor.scale(2.0,2.0);
this.touchCharactor.maxWaitCount = 6;
this.touchCharactor.addEventListener(enchant.Event.TOUCH_START,(e:enchant.Event)=>{
//タッチ開始時は前を向いて、アニメーションを停止させる
this.touchCharactor.Dir = Rf.ETS.FrameWork.Direction.Down;
this.touchCharactor.SuspendAnime();
this.touchCharactorTouchPosX = this.touchCharactor.x - e.x;
this.touchCharactorTouchPosY = this.touchCharactor.y - e.y;
});
this.touchCharactor.addEventListener(enchant.Event.TOUCH_MOVE,(e:enchant.Event)=>{
//タッチ中はその位置にキャラクタを移動させる
this.touchCharactor.x = e.x + this.touchCharactorTouchPosX;
this.touchCharactor.y = e.y + this.touchCharactorTouchPosY;
});
this.touchCharactor.addEventListener(enchant.Event.TOUCH_END,(e:enchant.Event)=>{
//タッチ終了時は後ろを向いて、アニメーションを再開させる
this.touchCharactor.Dir = Rf.ETS.FrameWork.Direction.Up;
this.touchCharactor.ResumeAnime();
});
}
/**
* 実行イベント
* @method
* @name FrameWork.GameMain#onRun
*/
protected onRun():void
{
//グループを右に移動する
this.group.x += 2;
if (this.group.x >= 200) {
this.group.x = 0;
}
//グループ内メンバのスプライトを回転させる
this.sprite.rotation += 5;
if (this.sprite.rotation >= 360) {
this.sprite.rotation = 0;
}
//タッチイベント用キャラクタのアニメーションを実行する
this.touchCharactor.Run();
}
}
//メインクラスのインスタンス作成
createMain(GameMain); | GameMain | identifier_name |
main.ts | class GameMain extends Rf.ETS.FrameWork.GameMain
{
private group: Rf.ETS.FrameWork.Group = null;
private sprite: Rf.ETS.FrameWork.Sprite = null;
private touchCharactor: Rf.ETS.FrameWork.Character = null;
private touchCharactorTouchPosX:number = 0;
private touchCharactorTouchPosY:number = 0;
/**
* 初期化イベント
* @method
* @name FrameWork.GameMain#onInitialize
*/
protected onInitialize():void
{
//サイズを640x640に変更
this.screenHeight = 640;
//fpsを10から30に変更
this.fps = 30;
}
/**
* リソース設定イベント
* @method
* @name FrameWork.GameMain#resourceLoad
*/
protected onResourceSetting():void
{
this.resourceManager.SetResourcePath("./assets/resources/");
this.resourceManager.AddResourceName("charaImage", "chara.png");
}
/**
* ロードイベント
* @method
* @name FrameWork.GameMain#onLoad
* @param {Object} parent - 親Group
*/
protected onLoad(parent: enchant.Group):void
{
//グループインスタンス作成
this.group = new Rf.ETS.FrameWork.Group(parent);
this.group.y = 100;
//画像無しスプライトインスタンス作成
let surface: Rf.ETS.FrameWork.NoImageSprite =
new Rf.ETS.FrameWork.NoImageSprite(100, 100, this.group);
surface.SetSurface("rgb(128,255,255)");
surface.opacity = 0.5;
//スプライトインスタンス作成
this.sprite = new Rf.ETS.FrameWork.Sprite(32, 32, this.group);
this.sprite.FileName = this.resourceManager.GetResourceName("charaImage");
this.sprite.originX = 16; //中心で回転するように設定
this.sprite.originY = 16; //中心で回転するように設定
this.sprite.frame = 26*2; //サンプル画像で正面画像を表示する
//タッチイベント用キャラクタ
| this.touchCharactor.y = 32;
this.touchCharactor.originX = 16*2;
this.touchCharactor.originY = 16*2;
this.touchCharactor.scale(2.0,2.0);
this.touchCharactor.maxWaitCount = 6;
this.touchCharactor.addEventListener(enchant.Event.TOUCH_START,(e:enchant.Event)=>{
//タッチ開始時は前を向いて、アニメーションを停止させる
this.touchCharactor.Dir = Rf.ETS.FrameWork.Direction.Down;
this.touchCharactor.SuspendAnime();
this.touchCharactorTouchPosX = this.touchCharactor.x - e.x;
this.touchCharactorTouchPosY = this.touchCharactor.y - e.y;
});
this.touchCharactor.addEventListener(enchant.Event.TOUCH_MOVE,(e:enchant.Event)=>{
//タッチ中はその位置にキャラクタを移動させる
this.touchCharactor.x = e.x + this.touchCharactorTouchPosX;
this.touchCharactor.y = e.y + this.touchCharactorTouchPosY;
});
this.touchCharactor.addEventListener(enchant.Event.TOUCH_END,(e:enchant.Event)=>{
//タッチ終了時は後ろを向いて、アニメーションを再開させる
this.touchCharactor.Dir = Rf.ETS.FrameWork.Direction.Up;
this.touchCharactor.ResumeAnime();
});
}
/**
* 実行イベント
* @method
* @name FrameWork.GameMain#onRun
*/
protected onRun():void
{
//グループを右に移動する
this.group.x += 2;
if (this.group.x >= 200) {
this.group.x = 0;
}
//グループ内メンバのスプライトを回転させる
this.sprite.rotation += 5;
if (this.sprite.rotation >= 360) {
this.sprite.rotation = 0;
}
//タッチイベント用キャラクタのアニメーションを実行する
this.touchCharactor.Run();
}
}
//メインクラスのインスタンス作成
createMain(GameMain); | this.touchCharactor = new Rf.ETS.FrameWork.Character(32,32,parent);
this.touchCharactor.FileName = this.resourceManager.GetResourceName("charaImage");
this.touchCharactor.charaIndex = 3;
this.touchCharactor.Dir = Rf.ETS.FrameWork.Direction.Up;
this.touchCharactor.x = 32;
| random_line_split |
main.ts | class GameMain extends Rf.ETS.FrameWork.GameMain
{
private group: Rf.ETS.FrameWork.Group = null;
private sprite: Rf.ETS.FrameWork.Sprite = null;
private touchCharactor: Rf.ETS.FrameWork.Character = null;
private touchCharactorTouchPosX:number = 0;
private touchCharactorTouchPosY:number = 0;
/**
* 初期化イベント
* @method
* @name FrameWork.GameMain#onInitialize
*/
protected onInitialize():void
{
//サイズを640x640に変更
this.screenHeight = 640;
//fpsを10から30に変更
this.fps = 30;
}
/**
* リソース設定イベント
* @method
* @name FrameWork.GameMain#resourceLoad
*/
protected onResourceSetting():void
{
this.resourceManager.SetResourcePath("./assets/resources/");
this.resourceManager.AddResourceName("charaImage", "chara.png");
}
/**
* ロードイベント
* @method
* @name FrameWork.GameMain#onLoad
* @param {Object} parent - 親Group
*/
protected onLoad(parent: enchant.Group):void
{
//グループインスタンス作成
this.group = new Rf.ETS.FrameWork.Group(parent);
this.group.y = 100;
//画像無しスプライトインスタンス作成
let surface: Rf.ETS.FrameWork.NoImageSprite =
new Rf.ETS.FrameWork.NoImageSprite(100, 100, this.group);
surface.SetSurface("rgb(128,255,255)");
surface.opacity = 0.5;
//スプライトインスタンス作成
this.sprite = new Rf.ETS.FrameWork.Sprite(32, 32, this.group);
this.sprite.FileName = this.resourceManager.GetResourceName("charaImage");
this.sprite.originX = 16; //中心で回転するように設定
this.sprite.originY = 16; //中心で回転するように設定
this.sprite.frame = 26*2; //サンプル画像で正面画像を表示する
//タッチイベント用キャラクタ
this.touchCharactor = new Rf.ETS.FrameWork.Character(32,32,parent);
this.touchCharactor.FileName = this.resourceManager.GetResourceName("charaImage");
this.touchCharactor.charaIndex = 3;
this.touchCharactor.Dir = Rf.ETS.FrameWork.Direction.Up;
this.touchCharactor.x = 32;
this.touchCharactor.y = 32;
this.touchCharactor.originX = 16*2;
this.touchCharactor.originY = 16*2;
this.touchCharactor.scale(2.0,2.0);
this.touchCharactor.maxWaitCount = 6;
this.touchCharactor.addEventListener(enchant.Event.TOUCH_START,(e:enchant.Event)=>{
//タッチ開始時は前を向いて、アニメーションを停止させる
this.touchCharactor.Dir = Rf.ETS.FrameWork.Direction.Down;
this.touchCharactor.SuspendAnime();
this.touchCharactorTouchPosX = this.touchCharactor.x - e.x;
this.touchCharactorTouchPosY = this.touchCharactor.y - e.y;
});
this.touchCharactor.addEventListener(enchant.Event.TOUCH_MOVE,(e:enchant.Event)=>{
//タッチ中はその位置にキャラクタを移動させる
this.touchCharactor.x = e.x + this.touchCharactorTouchPosX;
this.touchCharactor.y = e.y + this.touchCharactorTouchPosY;
});
this.touchCharactor.addEventListener(enchant.Event.TOUCH_END,(e:enchant.Event)=>{
//タッチ終了時は後ろを向いて、アニメーションを再開させる
this.touchCharactor.Dir = Rf.ETS.FrameWork.Direction.Up;
this.touchCharactor.ResumeAnime();
});
}
/**
* 実行イベント
* @method
* @name FrameWork.GameMain#onRun
*/
protected onRun():void
{
//グループを右に移動する
this.group.x += 2;
if (this.group.x >= 200) {
this.group.x = 0;
}
//グループ内メンバのスプライトを回転させる
this.sprite.rotation += 5;
if (this.sprite.rotation >= 360) {
this.sprite.rotation = 0;
}
//タッチイベント用キャラクタのアニメーションを実行する
this.touchCharactor.Run();
}
}
//メインクラスのインスタンス作成
createMain(GameMain); | conditional_block |
||
test_markup.py | # Check translations of pango markup
#
# This will look for translatable strings that appear to contain markup and
# check that the markup in the translation matches.
#
# Copyright (C) 2015 Red Hat, Inc.
#
# This copyrighted material is made available to anyone wishing to use,
# modify, copy, or redistribute it subject to the terms and conditions of
# the GNU General Public License v.2, or (at your option) any later version.
# This program is distributed in the hope that it will be useful, but WITHOUT
# ANY WARRANTY expressed or implied, including the implied warranties of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General
# Public License for more details. You should have received a copy of the
# GNU General Public License along with this program; if not, write to the
# Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA
# 02110-1301, USA. Any Red Hat trademarks that are incorporated in the
# source code or documentation are not subject to the GNU General Public
# License and may only be used or replicated with the express permission of
# Red Hat, Inc.
#
# Red Hat Author(s): David Shea <[email protected]>
try:
import polib
except ImportError:
print("You need to install the python-polib package to read translations")
raise
from pocketlint.pangocheck import is_markup, markup_match
import xml.etree.ElementTree as ET
def test_markup(mofile):
mo = polib.mofile(mofile)
for entry in mo.translated_entries():
if is_markup(entry.msgid):
# If this is a plural, check each of the plural translations
if entry.msgid_plural:
xlations = entry.msgstr_plural
else:
xlations = {None: entry.msgstr}
| for plural_id, msgstr in xlations.items():
# Check if the markup is valid at all
try:
# pylint: disable=unescaped-markup
ET.fromstring('<markup>%s</markup>' % msgstr)
except ET.ParseError:
if entry.msgid_plural:
raise AssertionError("Invalid markup translation for %d translation of msgid %s" %
(plural_id, entry.msgid))
else:
raise AssertionError("Invalid markup translation for msgid %s" % entry.msgid)
# Check if the markup has the same number and kind of tags
if not markup_match(entry.msgid, msgstr):
if entry.msgid_plural:
raise AssertionError("Markup does not match for %d translation of msgid %s" %
(plural_id, entry.msgid))
else:
raise AssertionError("Markup does not match for msgid %s" % entry.msgid) | random_line_split |
|
test_markup.py | # Check translations of pango markup
#
# This will look for translatable strings that appear to contain markup and
# check that the markup in the translation matches.
#
# Copyright (C) 2015 Red Hat, Inc.
#
# This copyrighted material is made available to anyone wishing to use,
# modify, copy, or redistribute it subject to the terms and conditions of
# the GNU General Public License v.2, or (at your option) any later version.
# This program is distributed in the hope that it will be useful, but WITHOUT
# ANY WARRANTY expressed or implied, including the implied warranties of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General
# Public License for more details. You should have received a copy of the
# GNU General Public License along with this program; if not, write to the
# Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA
# 02110-1301, USA. Any Red Hat trademarks that are incorporated in the
# source code or documentation are not subject to the GNU General Public
# License and may only be used or replicated with the express permission of
# Red Hat, Inc.
#
# Red Hat Author(s): David Shea <[email protected]>
try:
import polib
except ImportError:
print("You need to install the python-polib package to read translations")
raise
from pocketlint.pangocheck import is_markup, markup_match
import xml.etree.ElementTree as ET
def | (mofile):
mo = polib.mofile(mofile)
for entry in mo.translated_entries():
if is_markup(entry.msgid):
# If this is a plural, check each of the plural translations
if entry.msgid_plural:
xlations = entry.msgstr_plural
else:
xlations = {None: entry.msgstr}
for plural_id, msgstr in xlations.items():
# Check if the markup is valid at all
try:
# pylint: disable=unescaped-markup
ET.fromstring('<markup>%s</markup>' % msgstr)
except ET.ParseError:
if entry.msgid_plural:
raise AssertionError("Invalid markup translation for %d translation of msgid %s" %
(plural_id, entry.msgid))
else:
raise AssertionError("Invalid markup translation for msgid %s" % entry.msgid)
# Check if the markup has the same number and kind of tags
if not markup_match(entry.msgid, msgstr):
if entry.msgid_plural:
raise AssertionError("Markup does not match for %d translation of msgid %s" %
(plural_id, entry.msgid))
else:
raise AssertionError("Markup does not match for msgid %s" % entry.msgid)
| test_markup | identifier_name |
test_markup.py | # Check translations of pango markup
#
# This will look for translatable strings that appear to contain markup and
# check that the markup in the translation matches.
#
# Copyright (C) 2015 Red Hat, Inc.
#
# This copyrighted material is made available to anyone wishing to use,
# modify, copy, or redistribute it subject to the terms and conditions of
# the GNU General Public License v.2, or (at your option) any later version.
# This program is distributed in the hope that it will be useful, but WITHOUT
# ANY WARRANTY expressed or implied, including the implied warranties of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General
# Public License for more details. You should have received a copy of the
# GNU General Public License along with this program; if not, write to the
# Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA
# 02110-1301, USA. Any Red Hat trademarks that are incorporated in the
# source code or documentation are not subject to the GNU General Public
# License and may only be used or replicated with the express permission of
# Red Hat, Inc.
#
# Red Hat Author(s): David Shea <[email protected]>
try:
import polib
except ImportError:
print("You need to install the python-polib package to read translations")
raise
from pocketlint.pangocheck import is_markup, markup_match
import xml.etree.ElementTree as ET
def test_markup(mofile):
mo = polib.mofile(mofile)
for entry in mo.translated_entries():
if is_markup(entry.msgid):
# If this is a plural, check each of the plural translations
if entry.msgid_plural:
xlations = entry.msgstr_plural
else:
xlations = {None: entry.msgstr}
for plural_id, msgstr in xlations.items():
# Check if the markup is valid at all
try:
# pylint: disable=unescaped-markup
ET.fromstring('<markup>%s</markup>' % msgstr)
except ET.ParseError:
if entry.msgid_plural:
raise AssertionError("Invalid markup translation for %d translation of msgid %s" %
(plural_id, entry.msgid))
else:
raise AssertionError("Invalid markup translation for msgid %s" % entry.msgid)
# Check if the markup has the same number and kind of tags
if not markup_match(entry.msgid, msgstr):
if entry.msgid_plural:
raise AssertionError("Markup does not match for %d translation of msgid %s" %
(plural_id, entry.msgid))
else:
| raise AssertionError("Markup does not match for msgid %s" % entry.msgid) | conditional_block |
|
test_markup.py | # Check translations of pango markup
#
# This will look for translatable strings that appear to contain markup and
# check that the markup in the translation matches.
#
# Copyright (C) 2015 Red Hat, Inc.
#
# This copyrighted material is made available to anyone wishing to use,
# modify, copy, or redistribute it subject to the terms and conditions of
# the GNU General Public License v.2, or (at your option) any later version.
# This program is distributed in the hope that it will be useful, but WITHOUT
# ANY WARRANTY expressed or implied, including the implied warranties of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General
# Public License for more details. You should have received a copy of the
# GNU General Public License along with this program; if not, write to the
# Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA
# 02110-1301, USA. Any Red Hat trademarks that are incorporated in the
# source code or documentation are not subject to the GNU General Public
# License and may only be used or replicated with the express permission of
# Red Hat, Inc.
#
# Red Hat Author(s): David Shea <[email protected]>
try:
import polib
except ImportError:
print("You need to install the python-polib package to read translations")
raise
from pocketlint.pangocheck import is_markup, markup_match
import xml.etree.ElementTree as ET
def test_markup(mofile):
| raise AssertionError("Invalid markup translation for msgid %s" % entry.msgid)
# Check if the markup has the same number and kind of tags
if not markup_match(entry.msgid, msgstr):
if entry.msgid_plural:
raise AssertionError("Markup does not match for %d translation of msgid %s" %
(plural_id, entry.msgid))
else:
raise AssertionError("Markup does not match for msgid %s" % entry.msgid)
| mo = polib.mofile(mofile)
for entry in mo.translated_entries():
if is_markup(entry.msgid):
# If this is a plural, check each of the plural translations
if entry.msgid_plural:
xlations = entry.msgstr_plural
else:
xlations = {None: entry.msgstr}
for plural_id, msgstr in xlations.items():
# Check if the markup is valid at all
try:
# pylint: disable=unescaped-markup
ET.fromstring('<markup>%s</markup>' % msgstr)
except ET.ParseError:
if entry.msgid_plural:
raise AssertionError("Invalid markup translation for %d translation of msgid %s" %
(plural_id, entry.msgid))
else: | identifier_body |
index.d.ts | // Type definitions for animejs 2.0
// Project: http://animejs.com
// Definitions by: Andrew Babin <https://github.com/A-Babin>
// Definitions: https://github.com/DefinitelyTyped/DefinitelyTyped
// TypeScript Version: 2.4
type FunctionBasedParamter = (element: HTMLElement, index: number, length: number) => number;
type AnimeCallbackFunction = (anim: anime.AnimeInstance) => void;
// Allowing null is necessary because DOM queries may not return anything.
type AnimeTarget = string | object | HTMLElement | SVGElement | NodeList | null;
declare namespace anime {
type EasingOptions =
| "linear"
| "easeInQuad"
| "easeInCubic"
| "easeInQuart"
| "easeInQuint"
| "easeInSine"
| "easeInExpo"
| "easeInCirc"
| "easeInBack"
| "easeInElastic"
| "easeOutQuad"
| "easeOutCubic"
| "easeOutQuart"
| "easeOutQuint"
| "easeOutSine"
| "easeOutExpo"
| "easeOutCirc"
| "easeOutBack"
| "easeOutElastic"
| "easeInOutQuad"
| "easeInOutCubic"
| "easeInOutQuart"
| "easeInOutQuint"
| "easeInOutSine"
| "easeInOutExpo"
| "easeInOutCirc"
| "easeInOutBack"
| "easeInOutElastic";
type DirectionOptions = "reverse" | "alternate" | "normal";
interface AnimeInstanceParams {
loop?: number | boolean;
autoplay?: boolean;
direction?: DirectionOptions | string;
begin?: AnimeCallbackFunction;
run?: AnimeCallbackFunction;
update?: AnimeCallbackFunction;
complete?: AnimeCallbackFunction;
}
interface AnimeAnimParams {
targets: AnimeTarget | ReadonlyArray<AnimeTarget>;
duration?: number | FunctionBasedParamter;
delay?: number | FunctionBasedParamter;
elasticity?: number | FunctionBasedParamter;
round?: number | boolean | FunctionBasedParamter;
easing?: EasingOptions | string | ReadonlyArray<number>;
begin?: AnimeCallbackFunction;
run?: AnimeCallbackFunction;
update?: AnimeCallbackFunction;
complete?: AnimeCallbackFunction;
[AnyAnimatedProperty: string]: any;
}
interface AnimeParams extends AnimeInstanceParams, AnimeAnimParams {
// Just need this to merge both Params interfaces.
}
interface AnimeInstance {
play(): void;
pause(): void;
restart(): void;
reverse(): void;
seek(time: number): void;
began: boolean;
paused: boolean;
completed: boolean;
finished: Promise<void>;
begin: AnimeCallbackFunction;
run: AnimeCallbackFunction;
update: AnimeCallbackFunction;
complete: AnimeCallbackFunction;
autoplay: boolean;
currentTime: number;
delay: number;
direction: string;
duration: number;
loop: number | boolean;
offset: number;
progress: number;
remaining: number;
reversed: boolean;
animatables: ReadonlyArray<object>;
animations: ReadonlyArray<object>;
}
interface AnimeTimelineAnimParams extends AnimeAnimParams {
offset: number | string | FunctionBasedParamter;
}
interface AnimeTimelineInstance extends AnimeInstance {
add(params: AnimeAnimParams): AnimeTimelineInstance;
}
// Helpers
const speed: number;
const running: AnimeInstance[];
const easings: { [EasingFunction: string]: (t: number) => any };
function remove(targets: AnimeTarget | ReadonlyArray<AnimeTarget>): void;
function getValue(targets: AnimeTarget, prop: string): string | number;
function path(path: string | HTMLElement | SVGElement | null, percent?: number): (prop: string) => {
el: HTMLElement | SVGElement,
property: string,
totalLength: number
};
function setDashoffset(el: HTMLElement | SVGElement | null): number;
function bezier(x1: number, y1: number, x2: number, y2: number): (t: number) => number;
// Timeline | function random(min: number, max: number): number;
}
declare function anime(params: anime.AnimeParams): anime.AnimeInstance;
export = anime;
export as namespace anime; | function timeline(params?: AnimeInstanceParams | ReadonlyArray<AnimeInstance>): AnimeTimelineInstance; | random_line_split |
run_CNN_SAT.py | # Copyright 2015 Tianchuan Du University of Delaware
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# THIS CODE IS PROVIDED *AS IS* BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, EITHER EXPRESS OR IMPLIED, INCLUDING WITHOUT LIMITATION ANY IMPLIED
# WARRANTIES OR CONDITIONS OF TITLE, FITNESS FOR A PARTICULAR PURPOSE,
# MERCHANTABLITY OR NON-INFRINGEMENT.
# See the Apache 2 License for the specific language governing permissions and
# limitations under the License.
import cPickle
import gzip
import numpy
import os
import sys
import theano
from theano.tensor.shared_randomstreams import RandomStreams
import time
from io_func.model_io import _nnet2file, _file2nnet, _cfg2file, log
from learning.sgd import train_sgd_verbose, validate_by_minibatch_verbose
from models.cnn_sat import CNN_SAT
import theano.tensor as T
from utils.network_config import NetworkConfig
from utils.utils import parse_arguments
# Implements the Speaker Adaptive Training of DNNs proposed in the following papers:
# [1] Yajie Miao, Hao Zhang, Florian Metze. "Towards Speaker Adaptive Training of Deep
# Neural Network Acoustic Models". Interspeech 2014.
# [2] Yajie Miao, Lu Jiang, Hao Zhang, Florian Metze. "Improvements to Speaker Adaptive
# Training of Deep Neural Networks". SLT 2014.
if __name__ == '__main__':
# check the arguments
arg_elements = [sys.argv[i] for i in range(1, len(sys.argv))]
arguments = parse_arguments(arg_elements)
required_arguments = ['train_data', 'valid_data', 'si_nnet_spec', 'si_conv_nnet_spec', 'wdir', 'adapt_nnet_spec', 'init_model']
for arg in required_arguments:
if arguments.has_key(arg) == False:
print "Error: the argument %s has to be specified" % (arg); exit(1)
# mandatory arguments
train_data_spec = arguments['train_data']; valid_data_spec = arguments['valid_data']
si_nnet_spec = arguments['si_nnet_spec']
si_conv_nnet_spec = arguments['si_conv_nnet_spec']
adapt_nnet_spec = arguments['adapt_nnet_spec'];
wdir = arguments['wdir']
init_model_file = arguments['init_model']
# parse network configuration from arguments, and initialize data reading
cfg_si = NetworkConfig(); cfg_si.model_type = 'CNN'
cfg_si.parse_config_cnn(arguments, '10:' + si_nnet_spec, si_conv_nnet_spec)
cfg_si.init_data_reading(train_data_spec, valid_data_spec)
# parse the structure of the i-vector network
cfg_adapt = NetworkConfig()
net_split = adapt_nnet_spec.split(':')
adapt_nnet_spec = ''
for n in xrange(len(net_split) - 1):
adapt_nnet_spec += net_split[n] + ':'
cfg_adapt.parse_config_dnn(arguments, adapt_nnet_spec + '0')
numpy_rng = numpy.random.RandomState(89677)
theano_rng = RandomStreams(numpy_rng.randint(2 ** 30))
log('> ... initializing the model')
# setup up the model
dnn = CNN_SAT(numpy_rng=numpy_rng, theano_rng = theano_rng, cfg_si = cfg_si, cfg_adapt = cfg_adapt)
# read the initial DNN (the SI DNN which has been well trained)
# _file2nnet(dnn.cnn_si.layers, filename = init_model_file)
_file2nnet(dnn.cnn_si.layers, filename = 'BKUP/nnet.param.si')
_file2nnet(dnn.dnn_adapt.layers, filename = 'BKUP/nnet.param.adapt')
# get the training and validation functions for adaptation network training
dnn.params = dnn.dnn_adapt.params # only update the parameters of the i-vector nnet
dnn.delta_params = dnn.dnn_adapt.delta_params
log('> ... getting the finetuning functions for iVecNN')
train_fn, valid_fn = dnn.build_finetune_functions(
(cfg_si.train_x, cfg_si.train_y), (cfg_si.valid_x, cfg_si.valid_y),
batch_size = cfg_adapt.batch_size)
log('> ... learning the adaptation network')
cfg = cfg_adapt
while (cfg.lrate.get_rate() != 0):
# one epoch of sgd training
# train_error = train_sgd_verbose(train_fn, cfg_si.train_sets, cfg_si.train_xy,
# cfg.batch_size, cfg.lrate.get_rate(), cfg.momentum)
# log('> epoch %d, training error %f ' % (cfg.lrate.epoch, 100*numpy.mean(train_error)) + '(%)')
# validation
|
# save the model and network configuration
if cfg.param_output_file != '':
_nnet2file(dnn.dnn_adapt.layers, filename = cfg.param_output_file + '.adapt',
input_factor = cfg_adapt.input_dropout_factor, factor = cfg_adapt.dropout_factor)
_nnet2file(dnn.cnn_si.layers, filename = cfg.param_output_file + '.si',
input_factor = cfg_si.input_dropout_factor, factor = cfg_si.dropout_factor)
log('> ... the final PDNN model parameter is ' + cfg.param_output_file + ' (.si, .adapt)')
if cfg.cfg_output_file != '':
_cfg2file(cfg_adapt, filename=cfg.cfg_output_file + '.adapt')
_cfg2file(cfg_si, filename=cfg.cfg_output_file + '.si')
log('> ... the final PDNN model config is ' + cfg.cfg_output_file + ' (.si, .adapt)')
# output the model into Kaldi-compatible format
if cfg.kaldi_output_file != '':
dnn.cnn_si.fc_dnn.write_model_to_kaldi(cfg.kaldi_output_file + '.si')
dnn.dnn_adapt.write_model_to_kaldi(cfg.kaldi_output_file + '.adapt', with_softmax = False)
log('> ... the final Kaldi model is ' + cfg.kaldi_output_file + ' (.si, .adapt)')
| valid_error = validate_by_minibatch_verbose(valid_fn, cfg_si.valid_sets, cfg_si.valid_xy, cfg.batch_size)
log('> epoch %d, lrate %f, validation error %f ' % (cfg.lrate.epoch, cfg.lrate.get_rate(), 100*numpy.mean(valid_error)) + '(%)')
cfg.lrate.get_next_rate(current_error = 100 * numpy.mean(valid_error))
cfg.lrate.rate = 0 | conditional_block |
run_CNN_SAT.py | # Copyright 2015 Tianchuan Du University of Delaware
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# THIS CODE IS PROVIDED *AS IS* BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, EITHER EXPRESS OR IMPLIED, INCLUDING WITHOUT LIMITATION ANY IMPLIED
# WARRANTIES OR CONDITIONS OF TITLE, FITNESS FOR A PARTICULAR PURPOSE,
# MERCHANTABLITY OR NON-INFRINGEMENT.
# See the Apache 2 License for the specific language governing permissions and
# limitations under the License.
import cPickle
import gzip
import numpy
import os
import sys
import theano
from theano.tensor.shared_randomstreams import RandomStreams
import time
from io_func.model_io import _nnet2file, _file2nnet, _cfg2file, log
from learning.sgd import train_sgd_verbose, validate_by_minibatch_verbose
from models.cnn_sat import CNN_SAT
import theano.tensor as T
from utils.network_config import NetworkConfig |
# Implements the Speaker Adaptive Training of DNNs proposed in the following papers:
# [1] Yajie Miao, Hao Zhang, Florian Metze. "Towards Speaker Adaptive Training of Deep
# Neural Network Acoustic Models". Interspeech 2014.
# [2] Yajie Miao, Lu Jiang, Hao Zhang, Florian Metze. "Improvements to Speaker Adaptive
# Training of Deep Neural Networks". SLT 2014.
if __name__ == '__main__':
# check the arguments
arg_elements = [sys.argv[i] for i in range(1, len(sys.argv))]
arguments = parse_arguments(arg_elements)
required_arguments = ['train_data', 'valid_data', 'si_nnet_spec', 'si_conv_nnet_spec', 'wdir', 'adapt_nnet_spec', 'init_model']
for arg in required_arguments:
if arguments.has_key(arg) == False:
print "Error: the argument %s has to be specified" % (arg); exit(1)
# mandatory arguments
train_data_spec = arguments['train_data']; valid_data_spec = arguments['valid_data']
si_nnet_spec = arguments['si_nnet_spec']
si_conv_nnet_spec = arguments['si_conv_nnet_spec']
adapt_nnet_spec = arguments['adapt_nnet_spec'];
wdir = arguments['wdir']
init_model_file = arguments['init_model']
# parse network configuration from arguments, and initialize data reading
cfg_si = NetworkConfig(); cfg_si.model_type = 'CNN'
cfg_si.parse_config_cnn(arguments, '10:' + si_nnet_spec, si_conv_nnet_spec)
cfg_si.init_data_reading(train_data_spec, valid_data_spec)
# parse the structure of the i-vector network
cfg_adapt = NetworkConfig()
net_split = adapt_nnet_spec.split(':')
adapt_nnet_spec = ''
for n in xrange(len(net_split) - 1):
adapt_nnet_spec += net_split[n] + ':'
cfg_adapt.parse_config_dnn(arguments, adapt_nnet_spec + '0')
numpy_rng = numpy.random.RandomState(89677)
theano_rng = RandomStreams(numpy_rng.randint(2 ** 30))
log('> ... initializing the model')
# setup up the model
dnn = CNN_SAT(numpy_rng=numpy_rng, theano_rng = theano_rng, cfg_si = cfg_si, cfg_adapt = cfg_adapt)
# read the initial DNN (the SI DNN which has been well trained)
# _file2nnet(dnn.cnn_si.layers, filename = init_model_file)
_file2nnet(dnn.cnn_si.layers, filename = 'BKUP/nnet.param.si')
_file2nnet(dnn.dnn_adapt.layers, filename = 'BKUP/nnet.param.adapt')
# get the training and validation functions for adaptation network training
dnn.params = dnn.dnn_adapt.params # only update the parameters of the i-vector nnet
dnn.delta_params = dnn.dnn_adapt.delta_params
log('> ... getting the finetuning functions for iVecNN')
train_fn, valid_fn = dnn.build_finetune_functions(
(cfg_si.train_x, cfg_si.train_y), (cfg_si.valid_x, cfg_si.valid_y),
batch_size = cfg_adapt.batch_size)
log('> ... learning the adaptation network')
cfg = cfg_adapt
while (cfg.lrate.get_rate() != 0):
# one epoch of sgd training
# train_error = train_sgd_verbose(train_fn, cfg_si.train_sets, cfg_si.train_xy,
# cfg.batch_size, cfg.lrate.get_rate(), cfg.momentum)
# log('> epoch %d, training error %f ' % (cfg.lrate.epoch, 100*numpy.mean(train_error)) + '(%)')
# validation
valid_error = validate_by_minibatch_verbose(valid_fn, cfg_si.valid_sets, cfg_si.valid_xy, cfg.batch_size)
log('> epoch %d, lrate %f, validation error %f ' % (cfg.lrate.epoch, cfg.lrate.get_rate(), 100*numpy.mean(valid_error)) + '(%)')
cfg.lrate.get_next_rate(current_error = 100 * numpy.mean(valid_error))
cfg.lrate.rate = 0
# save the model and network configuration
if cfg.param_output_file != '':
_nnet2file(dnn.dnn_adapt.layers, filename = cfg.param_output_file + '.adapt',
input_factor = cfg_adapt.input_dropout_factor, factor = cfg_adapt.dropout_factor)
_nnet2file(dnn.cnn_si.layers, filename = cfg.param_output_file + '.si',
input_factor = cfg_si.input_dropout_factor, factor = cfg_si.dropout_factor)
log('> ... the final PDNN model parameter is ' + cfg.param_output_file + ' (.si, .adapt)')
if cfg.cfg_output_file != '':
_cfg2file(cfg_adapt, filename=cfg.cfg_output_file + '.adapt')
_cfg2file(cfg_si, filename=cfg.cfg_output_file + '.si')
log('> ... the final PDNN model config is ' + cfg.cfg_output_file + ' (.si, .adapt)')
# output the model into Kaldi-compatible format
if cfg.kaldi_output_file != '':
dnn.cnn_si.fc_dnn.write_model_to_kaldi(cfg.kaldi_output_file + '.si')
dnn.dnn_adapt.write_model_to_kaldi(cfg.kaldi_output_file + '.adapt', with_softmax = False)
log('> ... the final Kaldi model is ' + cfg.kaldi_output_file + ' (.si, .adapt)') | from utils.utils import parse_arguments
| random_line_split |
colorbeams.py | """
Color beams pattern
"""
from .pattern import Pattern
import colorsys
import time
class ColorBeams(Pattern):
@staticmethod
def getHue(hue):
hsv = colorsys.hsv_to_rgb(hue, 1, 1)
return int(hsv[0] * 255), int(hsv[1] * 255), int(hsv[2] * 255)
@staticmethod
def highlight(strip, i, hue = 0.5):
i = i % len(strip)
# set the color of this pixel
strip[i] = ColorBeams.getHue(hue)
for x in range(15):
index = (i - x) % len(strip)
decay = pow(0.7, x)
# strip[index] = (int(strip[index][0] * decay), int(strip[index][1] * decay), int(strip[index][2] * decay))
strip[index] = (int(strip[i][0] * decay), int(strip[i][1] * decay), int(strip[i][2] * decay)) |
def __init__(self):
pass
@classmethod
def get_id(self):
return 11
@classmethod
def update(self, strip, state):
# use the time to determine the offset
t = ColorBeams.__get_time()
offset = int(((t % state.delay) / state.delay) * len(strip))
for y in range(0, len(strip), 50):
ColorBeams.highlight(strip, offset + y, (5 * y / len(strip)) % 1) |
@staticmethod
def __get_time():
return time.time() * 1000 | random_line_split |
colorbeams.py | """
Color beams pattern
"""
from .pattern import Pattern
import colorsys
import time
class ColorBeams(Pattern):
@staticmethod
def | (hue):
hsv = colorsys.hsv_to_rgb(hue, 1, 1)
return int(hsv[0] * 255), int(hsv[1] * 255), int(hsv[2] * 255)
@staticmethod
def highlight(strip, i, hue = 0.5):
i = i % len(strip)
# set the color of this pixel
strip[i] = ColorBeams.getHue(hue)
for x in range(15):
index = (i - x) % len(strip)
decay = pow(0.7, x)
# strip[index] = (int(strip[index][0] * decay), int(strip[index][1] * decay), int(strip[index][2] * decay))
strip[index] = (int(strip[i][0] * decay), int(strip[i][1] * decay), int(strip[i][2] * decay))
@staticmethod
def __get_time():
return time.time() * 1000
def __init__(self):
pass
@classmethod
def get_id(self):
return 11
@classmethod
def update(self, strip, state):
# use the time to determine the offset
t = ColorBeams.__get_time()
offset = int(((t % state.delay) / state.delay) * len(strip))
for y in range(0, len(strip), 50):
ColorBeams.highlight(strip, offset + y, (5 * y / len(strip)) % 1)
| getHue | identifier_name |
colorbeams.py | """
Color beams pattern
"""
from .pattern import Pattern
import colorsys
import time
class ColorBeams(Pattern):
@staticmethod
def getHue(hue):
hsv = colorsys.hsv_to_rgb(hue, 1, 1)
return int(hsv[0] * 255), int(hsv[1] * 255), int(hsv[2] * 255)
@staticmethod
def highlight(strip, i, hue = 0.5):
i = i % len(strip)
# set the color of this pixel
strip[i] = ColorBeams.getHue(hue)
for x in range(15):
index = (i - x) % len(strip)
decay = pow(0.7, x)
# strip[index] = (int(strip[index][0] * decay), int(strip[index][1] * decay), int(strip[index][2] * decay))
strip[index] = (int(strip[i][0] * decay), int(strip[i][1] * decay), int(strip[i][2] * decay))
@staticmethod
def __get_time():
return time.time() * 1000
def __init__(self):
pass
@classmethod
def get_id(self):
|
@classmethod
def update(self, strip, state):
# use the time to determine the offset
t = ColorBeams.__get_time()
offset = int(((t % state.delay) / state.delay) * len(strip))
for y in range(0, len(strip), 50):
ColorBeams.highlight(strip, offset + y, (5 * y / len(strip)) % 1)
| return 11 | identifier_body |
colorbeams.py | """
Color beams pattern
"""
from .pattern import Pattern
import colorsys
import time
class ColorBeams(Pattern):
@staticmethod
def getHue(hue):
hsv = colorsys.hsv_to_rgb(hue, 1, 1)
return int(hsv[0] * 255), int(hsv[1] * 255), int(hsv[2] * 255)
@staticmethod
def highlight(strip, i, hue = 0.5):
i = i % len(strip)
# set the color of this pixel
strip[i] = ColorBeams.getHue(hue)
for x in range(15):
|
@staticmethod
def __get_time():
return time.time() * 1000
def __init__(self):
pass
@classmethod
def get_id(self):
return 11
@classmethod
def update(self, strip, state):
# use the time to determine the offset
t = ColorBeams.__get_time()
offset = int(((t % state.delay) / state.delay) * len(strip))
for y in range(0, len(strip), 50):
ColorBeams.highlight(strip, offset + y, (5 * y / len(strip)) % 1)
| index = (i - x) % len(strip)
decay = pow(0.7, x)
# strip[index] = (int(strip[index][0] * decay), int(strip[index][1] * decay), int(strip[index][2] * decay))
strip[index] = (int(strip[i][0] * decay), int(strip[i][1] * decay), int(strip[i][2] * decay)) | conditional_block |
facebookConnect.js | /*
* Module : FacebookConnect.js
*
* Setups up the basic code to connect to the facebook JS api.
*
* Requires Config:
* - app.config.facebook.appId
*/
(function(app)
{
var module = app.module("facebookConnect", {
requires : [
"jquery-1.9.1.min"
],
init : function()
{
this.is_loaded = false;
//loading facebook all.js after we've added the fb-root div to avoid fb warning
$('body').prepend('<div id="fb-root"></div>');
app.getScript("//connect.facebook.net/en_US/all.js");
},
scrollTop : function()
{
if( this.is_loaded )
FB.Canvas.scrollTo(0, 0);
}
});
//add listener to window object
window.fbAsyncInit = function ()
{
module.is_loaded = true;
// init the FB JS SDK
FB.init({
appId : app.config.facebook.appId,
status : true, // check the login status upon init?
cookie : true, // set sessions cookies to allow your server to access the session?
xfbml : false // parse XFBML tags on this page?
});
// Grow the canvas to the correct size
FB.Canvas.scrollTo(0, 0);
FB.Canvas.setSize({ height: $('body').height()-100});
setTimeout("FB.Canvas.setAutoGrow()", 500);
//dispact event |
//fix scroll bars
if (self !== top)
{
$("body").css("overflow", "hidden");
}
};
}(app)); | $(document).trigger('facebookConnected'); | random_line_split |
facebookConnect.js | /*
* Module : FacebookConnect.js
*
* Setups up the basic code to connect to the facebook JS api.
*
* Requires Config:
* - app.config.facebook.appId
*/
(function(app)
{
var module = app.module("facebookConnect", {
requires : [
"jquery-1.9.1.min"
],
init : function()
{
this.is_loaded = false;
//loading facebook all.js after we've added the fb-root div to avoid fb warning
$('body').prepend('<div id="fb-root"></div>');
app.getScript("//connect.facebook.net/en_US/all.js");
},
scrollTop : function()
{
if( this.is_loaded )
FB.Canvas.scrollTo(0, 0);
}
});
//add listener to window object
window.fbAsyncInit = function ()
{
module.is_loaded = true;
// init the FB JS SDK
FB.init({
appId : app.config.facebook.appId,
status : true, // check the login status upon init?
cookie : true, // set sessions cookies to allow your server to access the session?
xfbml : false // parse XFBML tags on this page?
});
// Grow the canvas to the correct size
FB.Canvas.scrollTo(0, 0);
FB.Canvas.setSize({ height: $('body').height()-100});
setTimeout("FB.Canvas.setAutoGrow()", 500);
//dispact event
$(document).trigger('facebookConnected');
//fix scroll bars
if (self !== top)
|
};
}(app));
| {
$("body").css("overflow", "hidden");
} | conditional_block |
mod.rs | use lexer::dfa::*;
use lexer::re::Test;
use rust::RustWrite;
use std::io::{self, Write};
#[cfg(test)]
mod test;
/// Generates a fn `__tokenize` based on the given DFA with the following signature:
///
/// ```ignore
/// fn tokenize(text: &str) -> Option<(usize, usize)>
/// ```
///
/// This function returns `None` if there is no matching
/// token. Otherwise, it returns the pair of (NFA index, length) for
/// the next token.
pub fn compile_tokenize_fn<W: Write>(
prefix: &str,
dfa: &DFA,
out: &mut RustWrite<W>)
-> io::Result<()>
{
let mut matcher = Matcher { prefix: prefix, dfa: dfa, out: out };
try!(matcher.tokenize());
Ok(())
}
struct Matcher<'m, W: Write+'m> {
prefix: &'m str,
dfa: &'m DFA,
out: &'m mut RustWrite<W>,
}
impl<'m,W> Matcher<'m,W>
where W: Write
{
fn tokenize(&mut self) -> io::Result<()> {
rust!(self.out, "fn {}tokenize(text: &str) -> Option<(usize, usize)> {{",
self.prefix);
rust!(self.out, "let mut {}chars = text.char_indices();", self.prefix);
rust!(self.out, "let mut {}current_match: Option<(usize, usize)> = None;", self.prefix);
rust!(self.out, "let mut {}current_state: usize = 0;", self.prefix);
rust!(self.out, "loop {{");
rust!(self.out, "match {}current_state {{", self.prefix);
for (index, state) in self.dfa.states.iter().enumerate() {
rust!(self.out, "{} => {{", index);
try!(self.state(state));
rust!(self.out, "}}");
}
rust!(self.out, "_ => {{ panic!(\"invalid state {{}}\", {}current_state); }}",
self.prefix);
rust!(self.out, "}}");
rust!(self.out, "}}");
rust!(self.out, "}}");
Ok(())
}
fn state(&mut self, state: &State) -> io::Result<()> {
// this could be pulled to the top of the loop, but we want to
// encourage LLVM to convert the loop+switch pair into actual
// gotos.
rust!(self.out, "let ({}index, {}ch) = \
match {}chars.next() {{ Some(p) => p, None => return {}current_match }};",
self.prefix, self.prefix, self.prefix, self.prefix);
rust!(self.out, "match {}ch {{", self.prefix);
for &(test, target_state) in &state.test_edges {
match test {
Test::Char(ch) => {
rust!(self.out, "{:?} => {{", ch);
let index = format!("{}index + {}", self.prefix, ch.len_utf8());
try!(self.transition(target_state, &index));
rust!(self.out, "}}");
}
}
}
rust!(self.out, "_ => {{");
let index = format!("{}index + {}ch.len_utf8()", self.prefix, self.prefix);
try!(self.transition(state.other_edge, &index));
rust!(self.out, "}}");
rust!(self.out, "}}");
Ok(())
}
fn | (&mut self,
target_state: DFAStateIndex,
index: &str)
-> io::Result<()> {
match self.dfa.state(target_state).kind {
Kind::Accepts(nfa) => {
rust!(self.out, "{}current_match = Some(({}, {}));",
self.prefix, nfa.index(), index)
}
Kind::Neither => { }
Kind::Reject => {
rust!(self.out, "return {}current_match;", self.prefix);
return Ok(());
}
}
rust!(self.out, "{}current_state = {};", self.prefix, target_state.index());
rust!(self.out, "continue;");
Ok(())
}
}
| transition | identifier_name |
mod.rs | use lexer::dfa::*;
use lexer::re::Test;
use rust::RustWrite;
use std::io::{self, Write};
#[cfg(test)]
mod test;
/// Generates a fn `__tokenize` based on the given DFA with the following signature:
///
/// ```ignore
/// fn tokenize(text: &str) -> Option<(usize, usize)>
/// ```
///
/// This function returns `None` if there is no matching
/// token. Otherwise, it returns the pair of (NFA index, length) for
/// the next token.
pub fn compile_tokenize_fn<W: Write>(
prefix: &str,
dfa: &DFA,
out: &mut RustWrite<W>)
-> io::Result<()>
|
struct Matcher<'m, W: Write+'m> {
prefix: &'m str,
dfa: &'m DFA,
out: &'m mut RustWrite<W>,
}
impl<'m,W> Matcher<'m,W>
where W: Write
{
fn tokenize(&mut self) -> io::Result<()> {
rust!(self.out, "fn {}tokenize(text: &str) -> Option<(usize, usize)> {{",
self.prefix);
rust!(self.out, "let mut {}chars = text.char_indices();", self.prefix);
rust!(self.out, "let mut {}current_match: Option<(usize, usize)> = None;", self.prefix);
rust!(self.out, "let mut {}current_state: usize = 0;", self.prefix);
rust!(self.out, "loop {{");
rust!(self.out, "match {}current_state {{", self.prefix);
for (index, state) in self.dfa.states.iter().enumerate() {
rust!(self.out, "{} => {{", index);
try!(self.state(state));
rust!(self.out, "}}");
}
rust!(self.out, "_ => {{ panic!(\"invalid state {{}}\", {}current_state); }}",
self.prefix);
rust!(self.out, "}}");
rust!(self.out, "}}");
rust!(self.out, "}}");
Ok(())
}
fn state(&mut self, state: &State) -> io::Result<()> {
// this could be pulled to the top of the loop, but we want to
// encourage LLVM to convert the loop+switch pair into actual
// gotos.
rust!(self.out, "let ({}index, {}ch) = \
match {}chars.next() {{ Some(p) => p, None => return {}current_match }};",
self.prefix, self.prefix, self.prefix, self.prefix);
rust!(self.out, "match {}ch {{", self.prefix);
for &(test, target_state) in &state.test_edges {
match test {
Test::Char(ch) => {
rust!(self.out, "{:?} => {{", ch);
let index = format!("{}index + {}", self.prefix, ch.len_utf8());
try!(self.transition(target_state, &index));
rust!(self.out, "}}");
}
}
}
rust!(self.out, "_ => {{");
let index = format!("{}index + {}ch.len_utf8()", self.prefix, self.prefix);
try!(self.transition(state.other_edge, &index));
rust!(self.out, "}}");
rust!(self.out, "}}");
Ok(())
}
fn transition(&mut self,
target_state: DFAStateIndex,
index: &str)
-> io::Result<()> {
match self.dfa.state(target_state).kind {
Kind::Accepts(nfa) => {
rust!(self.out, "{}current_match = Some(({}, {}));",
self.prefix, nfa.index(), index)
}
Kind::Neither => { }
Kind::Reject => {
rust!(self.out, "return {}current_match;", self.prefix);
return Ok(());
}
}
rust!(self.out, "{}current_state = {};", self.prefix, target_state.index());
rust!(self.out, "continue;");
Ok(())
}
}
| {
let mut matcher = Matcher { prefix: prefix, dfa: dfa, out: out };
try!(matcher.tokenize());
Ok(())
} | identifier_body |
mod.rs | use lexer::dfa::*;
use lexer::re::Test;
use rust::RustWrite;
use std::io::{self, Write};
#[cfg(test)]
mod test;
/// Generates a fn `__tokenize` based on the given DFA with the following signature:
///
/// ```ignore
/// fn tokenize(text: &str) -> Option<(usize, usize)>
/// ```
///
/// This function returns `None` if there is no matching
/// token. Otherwise, it returns the pair of (NFA index, length) for
/// the next token.
pub fn compile_tokenize_fn<W: Write>(
prefix: &str,
dfa: &DFA,
out: &mut RustWrite<W>)
-> io::Result<()>
{
let mut matcher = Matcher { prefix: prefix, dfa: dfa, out: out };
try!(matcher.tokenize());
Ok(())
}
struct Matcher<'m, W: Write+'m> {
prefix: &'m str,
dfa: &'m DFA,
out: &'m mut RustWrite<W>,
}
impl<'m,W> Matcher<'m,W>
where W: Write
{
fn tokenize(&mut self) -> io::Result<()> {
rust!(self.out, "fn {}tokenize(text: &str) -> Option<(usize, usize)> {{",
self.prefix);
rust!(self.out, "let mut {}chars = text.char_indices();", self.prefix);
rust!(self.out, "let mut {}current_match: Option<(usize, usize)> = None;", self.prefix);
rust!(self.out, "let mut {}current_state: usize = 0;", self.prefix);
rust!(self.out, "loop {{");
rust!(self.out, "match {}current_state {{", self.prefix);
for (index, state) in self.dfa.states.iter().enumerate() {
rust!(self.out, "{} => {{", index);
try!(self.state(state));
rust!(self.out, "}}");
}
rust!(self.out, "_ => {{ panic!(\"invalid state {{}}\", {}current_state); }}",
self.prefix);
rust!(self.out, "}}");
rust!(self.out, "}}");
rust!(self.out, "}}");
Ok(())
}
fn state(&mut self, state: &State) -> io::Result<()> {
// this could be pulled to the top of the loop, but we want to
// encourage LLVM to convert the loop+switch pair into actual
// gotos.
rust!(self.out, "let ({}index, {}ch) = \
match {}chars.next() {{ Some(p) => p, None => return {}current_match }};",
self.prefix, self.prefix, self.prefix, self.prefix);
rust!(self.out, "match {}ch {{", self.prefix);
for &(test, target_state) in &state.test_edges {
match test {
Test::Char(ch) => {
rust!(self.out, "{:?} => {{", ch);
let index = format!("{}index + {}", self.prefix, ch.len_utf8());
try!(self.transition(target_state, &index));
rust!(self.out, "}}");
}
}
}
rust!(self.out, "_ => {{");
let index = format!("{}index + {}ch.len_utf8()", self.prefix, self.prefix);
try!(self.transition(state.other_edge, &index));
rust!(self.out, "}}");
rust!(self.out, "}}");
Ok(())
}
fn transition(&mut self,
target_state: DFAStateIndex,
index: &str)
-> io::Result<()> {
match self.dfa.state(target_state).kind {
Kind::Accepts(nfa) => {
rust!(self.out, "{}current_match = Some(({}, {}));",
self.prefix, nfa.index(), index)
}
Kind::Neither => |
Kind::Reject => {
rust!(self.out, "return {}current_match;", self.prefix);
return Ok(());
}
}
rust!(self.out, "{}current_state = {};", self.prefix, target_state.index());
rust!(self.out, "continue;");
Ok(())
}
}
| { } | conditional_block |
mod.rs | use lexer::dfa::*;
use lexer::re::Test;
use rust::RustWrite;
use std::io::{self, Write};
#[cfg(test)]
mod test;
/// Generates a fn `__tokenize` based on the given DFA with the following signature:
///
/// ```ignore
/// fn tokenize(text: &str) -> Option<(usize, usize)>
/// ```
///
/// This function returns `None` if there is no matching
/// token. Otherwise, it returns the pair of (NFA index, length) for
/// the next token.
pub fn compile_tokenize_fn<W: Write>(
prefix: &str,
dfa: &DFA,
out: &mut RustWrite<W>)
-> io::Result<()>
{
let mut matcher = Matcher { prefix: prefix, dfa: dfa, out: out };
try!(matcher.tokenize());
Ok(())
}
struct Matcher<'m, W: Write+'m> {
prefix: &'m str,
dfa: &'m DFA,
out: &'m mut RustWrite<W>,
}
impl<'m,W> Matcher<'m,W>
where W: Write
{
fn tokenize(&mut self) -> io::Result<()> {
rust!(self.out, "fn {}tokenize(text: &str) -> Option<(usize, usize)> {{",
self.prefix);
rust!(self.out, "let mut {}chars = text.char_indices();", self.prefix);
rust!(self.out, "let mut {}current_match: Option<(usize, usize)> = None;", self.prefix);
rust!(self.out, "let mut {}current_state: usize = 0;", self.prefix);
rust!(self.out, "loop {{");
rust!(self.out, "match {}current_state {{", self.prefix); | for (index, state) in self.dfa.states.iter().enumerate() {
rust!(self.out, "{} => {{", index);
try!(self.state(state));
rust!(self.out, "}}");
}
rust!(self.out, "_ => {{ panic!(\"invalid state {{}}\", {}current_state); }}",
self.prefix);
rust!(self.out, "}}");
rust!(self.out, "}}");
rust!(self.out, "}}");
Ok(())
}
fn state(&mut self, state: &State) -> io::Result<()> {
// this could be pulled to the top of the loop, but we want to
// encourage LLVM to convert the loop+switch pair into actual
// gotos.
rust!(self.out, "let ({}index, {}ch) = \
match {}chars.next() {{ Some(p) => p, None => return {}current_match }};",
self.prefix, self.prefix, self.prefix, self.prefix);
rust!(self.out, "match {}ch {{", self.prefix);
for &(test, target_state) in &state.test_edges {
match test {
Test::Char(ch) => {
rust!(self.out, "{:?} => {{", ch);
let index = format!("{}index + {}", self.prefix, ch.len_utf8());
try!(self.transition(target_state, &index));
rust!(self.out, "}}");
}
}
}
rust!(self.out, "_ => {{");
let index = format!("{}index + {}ch.len_utf8()", self.prefix, self.prefix);
try!(self.transition(state.other_edge, &index));
rust!(self.out, "}}");
rust!(self.out, "}}");
Ok(())
}
fn transition(&mut self,
target_state: DFAStateIndex,
index: &str)
-> io::Result<()> {
match self.dfa.state(target_state).kind {
Kind::Accepts(nfa) => {
rust!(self.out, "{}current_match = Some(({}, {}));",
self.prefix, nfa.index(), index)
}
Kind::Neither => { }
Kind::Reject => {
rust!(self.out, "return {}current_match;", self.prefix);
return Ok(());
}
}
rust!(self.out, "{}current_state = {};", self.prefix, target_state.index());
rust!(self.out, "continue;");
Ok(())
}
} | random_line_split |
|
test.py | # -*- coding: utf-8 -*-
import base64
import inspect
import json
import logging
import requests
import types
from django.conf import settings
from django.core.management import call_command
from django_nose import FastFixtureTestCase
from functools import wraps
from mock import patch
from tastypie.test import ResourceTestCase, TestApiClient
from rpc_proxy.proxies import get_setting
INITIAL_DATA = ('initial_data',)
TEST_DATA = ('test_data',)
logger = logging.getLogger(__name__)
def mock_request(obj, method, url, **kwargs):
client = TestApiClient()
authentication = 'Basic %s' % base64.b64encode(':'.join([
get_setting('SUPERUSER_USERNAME', None),
get_setting('SUPERUSER_PASSWORD', None),
]))
if method == 'GET':
data = kwargs.get('params', {})
djresponse = client.get(url, data=data, authentication=authentication)
elif method == 'POST':
data = json.loads(kwargs.get('data', '{}'))
djresponse = client.post(url, data=data, authentication=authentication)
elif method == 'PUT':
data = json.loads(kwargs.get('data', '{}'))
djresponse = client.put(url, data=data, authentication=authentication)
elif method == 'PATCH':
data = json.loads(kwargs.get('data', '{}'))
djresponse = client.patch(url, data=data, authentication=authentication)
elif method == 'DELETE':
data = kwargs.get('params', {})
djresponse = client.delete(url, data=data, authentication=authentication)
# convert django.http.HttpResponse to requests.models.Response
response = requests.models.Response()
response.status_code = djresponse.status_code
response.headers = {}
try:
response.headers['content-type'] = djresponse['content-type']
response.headers['location'] = djresponse['location']
except:
pass
response.encoding = requests.utils.get_encoding_from_headers(response.headers)
response._content = djresponse.content
return response
def mock_cache_set(key, value, timeout=None):
# do nothing
pass
def mock_api(func, **decorator_kwargs):
@patch('requests.sessions.Session.request', mock_request)
@patch('tastypie.cache.SimpleCache.set', mock_cache_set)
@wraps(func)
def wrapper(*args, **kwargs):
return func(*args, **kwargs)
return wrapper
class TestCase(FastFixtureTestCase):
"""
Don't be smart in test cases!
"""
fixtures = INITIAL_DATA
def __new__(cls, name):
testcase = super(TestCase, cls).__new__(cls)
if get_setting('API_URL', None):
try:
func_type = types.UnboundMethodType
except:
func_type = types.FunctionType
for name, func in inspect.getmembers(testcase):
if isinstance(func, func_type) and name.startswith('test_'):
setattr(testcase, name, mock_api(func))
return testcase
def setUp(self):
call_command('loaddata', *TEST_DATA)
super(TestCase, self).setUp()
class Proxy(TestCase):
| """
Don't be smart in test cases!
CAVEAT: Proxy classes have to be imported within each test method
to mock the requests
"""
pass | identifier_body |
|
test.py | # -*- coding: utf-8 -*-
import base64
import inspect
import json
import logging
import requests
import types
from django.conf import settings
from django.core.management import call_command
from django_nose import FastFixtureTestCase
from functools import wraps
from mock import patch
from tastypie.test import ResourceTestCase, TestApiClient
from rpc_proxy.proxies import get_setting
INITIAL_DATA = ('initial_data',)
TEST_DATA = ('test_data',)
logger = logging.getLogger(__name__)
def mock_request(obj, method, url, **kwargs):
client = TestApiClient()
authentication = 'Basic %s' % base64.b64encode(':'.join([
get_setting('SUPERUSER_USERNAME', None),
get_setting('SUPERUSER_PASSWORD', None),
]))
if method == 'GET':
data = kwargs.get('params', {})
djresponse = client.get(url, data=data, authentication=authentication)
elif method == 'POST':
data = json.loads(kwargs.get('data', '{}'))
djresponse = client.post(url, data=data, authentication=authentication)
elif method == 'PUT':
data = json.loads(kwargs.get('data', '{}'))
djresponse = client.put(url, data=data, authentication=authentication)
elif method == 'PATCH':
data = json.loads(kwargs.get('data', '{}'))
djresponse = client.patch(url, data=data, authentication=authentication)
elif method == 'DELETE':
data = kwargs.get('params', {})
djresponse = client.delete(url, data=data, authentication=authentication)
# convert django.http.HttpResponse to requests.models.Response
response = requests.models.Response()
response.status_code = djresponse.status_code
response.headers = {}
try:
response.headers['content-type'] = djresponse['content-type']
response.headers['location'] = djresponse['location']
except:
pass
response.encoding = requests.utils.get_encoding_from_headers(response.headers)
response._content = djresponse.content
return response
def mock_cache_set(key, value, timeout=None):
# do nothing
pass
def mock_api(func, **decorator_kwargs):
@patch('requests.sessions.Session.request', mock_request)
@patch('tastypie.cache.SimpleCache.set', mock_cache_set)
@wraps(func)
def wrapper(*args, **kwargs):
return func(*args, **kwargs)
return wrapper
class TestCase(FastFixtureTestCase):
"""
Don't be smart in test cases!
"""
fixtures = INITIAL_DATA
def __new__(cls, name):
testcase = super(TestCase, cls).__new__(cls)
if get_setting('API_URL', None):
try:
func_type = types.UnboundMethodType
except:
func_type = types.FunctionType
for name, func in inspect.getmembers(testcase):
if isinstance(func, func_type) and name.startswith('test_'):
setattr(testcase, name, mock_api(func))
return testcase
def | (self):
call_command('loaddata', *TEST_DATA)
super(TestCase, self).setUp()
class Proxy(TestCase):
"""
Don't be smart in test cases!
CAVEAT: Proxy classes have to be imported within each test method
to mock the requests
"""
pass
| setUp | identifier_name |
test.py | # -*- coding: utf-8 -*-
import base64
import inspect
import json
import logging
import requests
import types
from django.conf import settings
from django.core.management import call_command
from django_nose import FastFixtureTestCase
from functools import wraps
from mock import patch
from tastypie.test import ResourceTestCase, TestApiClient
from rpc_proxy.proxies import get_setting
INITIAL_DATA = ('initial_data',)
TEST_DATA = ('test_data',)
logger = logging.getLogger(__name__)
def mock_request(obj, method, url, **kwargs):
client = TestApiClient()
authentication = 'Basic %s' % base64.b64encode(':'.join([
get_setting('SUPERUSER_USERNAME', None),
get_setting('SUPERUSER_PASSWORD', None),
]))
if method == 'GET':
data = kwargs.get('params', {})
djresponse = client.get(url, data=data, authentication=authentication)
elif method == 'POST':
|
elif method == 'PUT':
data = json.loads(kwargs.get('data', '{}'))
djresponse = client.put(url, data=data, authentication=authentication)
elif method == 'PATCH':
data = json.loads(kwargs.get('data', '{}'))
djresponse = client.patch(url, data=data, authentication=authentication)
elif method == 'DELETE':
data = kwargs.get('params', {})
djresponse = client.delete(url, data=data, authentication=authentication)
# convert django.http.HttpResponse to requests.models.Response
response = requests.models.Response()
response.status_code = djresponse.status_code
response.headers = {}
try:
response.headers['content-type'] = djresponse['content-type']
response.headers['location'] = djresponse['location']
except:
pass
response.encoding = requests.utils.get_encoding_from_headers(response.headers)
response._content = djresponse.content
return response
def mock_cache_set(key, value, timeout=None):
# do nothing
pass
def mock_api(func, **decorator_kwargs):
@patch('requests.sessions.Session.request', mock_request)
@patch('tastypie.cache.SimpleCache.set', mock_cache_set)
@wraps(func)
def wrapper(*args, **kwargs):
return func(*args, **kwargs)
return wrapper
class TestCase(FastFixtureTestCase):
"""
Don't be smart in test cases!
"""
fixtures = INITIAL_DATA
def __new__(cls, name):
testcase = super(TestCase, cls).__new__(cls)
if get_setting('API_URL', None):
try:
func_type = types.UnboundMethodType
except:
func_type = types.FunctionType
for name, func in inspect.getmembers(testcase):
if isinstance(func, func_type) and name.startswith('test_'):
setattr(testcase, name, mock_api(func))
return testcase
def setUp(self):
call_command('loaddata', *TEST_DATA)
super(TestCase, self).setUp()
class Proxy(TestCase):
"""
Don't be smart in test cases!
CAVEAT: Proxy classes have to be imported within each test method
to mock the requests
"""
pass
| data = json.loads(kwargs.get('data', '{}'))
djresponse = client.post(url, data=data, authentication=authentication) | conditional_block |
test.py | # -*- coding: utf-8 -*-
import base64
import inspect
import json
import logging
import requests
import types
from django.conf import settings
from django.core.management import call_command
from django_nose import FastFixtureTestCase
from functools import wraps
from mock import patch
from tastypie.test import ResourceTestCase, TestApiClient
from rpc_proxy.proxies import get_setting
INITIAL_DATA = ('initial_data',)
TEST_DATA = ('test_data',)
logger = logging.getLogger(__name__)
def mock_request(obj, method, url, **kwargs):
client = TestApiClient()
authentication = 'Basic %s' % base64.b64encode(':'.join([
get_setting('SUPERUSER_USERNAME', None),
get_setting('SUPERUSER_PASSWORD', None),
]))
if method == 'GET':
data = kwargs.get('params', {})
djresponse = client.get(url, data=data, authentication=authentication)
elif method == 'POST':
data = json.loads(kwargs.get('data', '{}'))
djresponse = client.post(url, data=data, authentication=authentication)
elif method == 'PUT':
data = json.loads(kwargs.get('data', '{}'))
djresponse = client.put(url, data=data, authentication=authentication)
elif method == 'PATCH':
data = json.loads(kwargs.get('data', '{}'))
djresponse = client.patch(url, data=data, authentication=authentication)
elif method == 'DELETE':
data = kwargs.get('params', {})
djresponse = client.delete(url, data=data, authentication=authentication)
# convert django.http.HttpResponse to requests.models.Response
response = requests.models.Response()
response.status_code = djresponse.status_code
response.headers = {}
try:
response.headers['content-type'] = djresponse['content-type']
response.headers['location'] = djresponse['location'] | return response
def mock_cache_set(key, value, timeout=None):
# do nothing
pass
def mock_api(func, **decorator_kwargs):
@patch('requests.sessions.Session.request', mock_request)
@patch('tastypie.cache.SimpleCache.set', mock_cache_set)
@wraps(func)
def wrapper(*args, **kwargs):
return func(*args, **kwargs)
return wrapper
class TestCase(FastFixtureTestCase):
"""
Don't be smart in test cases!
"""
fixtures = INITIAL_DATA
def __new__(cls, name):
testcase = super(TestCase, cls).__new__(cls)
if get_setting('API_URL', None):
try:
func_type = types.UnboundMethodType
except:
func_type = types.FunctionType
for name, func in inspect.getmembers(testcase):
if isinstance(func, func_type) and name.startswith('test_'):
setattr(testcase, name, mock_api(func))
return testcase
def setUp(self):
call_command('loaddata', *TEST_DATA)
super(TestCase, self).setUp()
class Proxy(TestCase):
"""
Don't be smart in test cases!
CAVEAT: Proxy classes have to be imported within each test method
to mock the requests
"""
pass | except:
pass
response.encoding = requests.utils.get_encoding_from_headers(response.headers)
response._content = djresponse.content
| random_line_split |
immunicity.py | import re
import fnmatch
import urllib2
from kodipopcorntime import plugin
from kodipopcorntime.caching import shelf
PAC_URL = "http://clientconfig.immunicity.org/pacs/all.pac"
CACHE = 1 * 3600 # 1 hour caching
_config = {}
def config():
global _config
if not _config:
with shelf("kodipopcorntime.immunicity.pac_config", ttl=CACHE) as pac_config:
plugin.log.info("Fetching Immunicity PAC file")
pac_data = urllib2.urlopen(PAC_URL).read()
pac_config["server"] = re.search(r"var proxyserver = '(.*)'", pac_data).group(1)
pac_config["domains"] = map(lambda x: x.replace(r"\Z(?ms)", ""), map(fnmatch.translate, re.findall(r"\"(.*?)\",", pac_data)))
_config = pac_config
return _config
def get_proxy_for(url):
| conf = config()
for domain in conf["domains"]:
if re.search(domain, url):
return conf["server"] | identifier_body |
|
immunicity.py | import re
import fnmatch
import urllib2
from kodipopcorntime import plugin
from kodipopcorntime.caching import shelf
PAC_URL = "http://clientconfig.immunicity.org/pacs/all.pac"
CACHE = 1 * 3600 # 1 hour caching
_config = {}
def config():
global _config
if not _config:
|
return _config
def get_proxy_for(url):
conf = config()
for domain in conf["domains"]:
if re.search(domain, url):
return conf["server"]
| with shelf("kodipopcorntime.immunicity.pac_config", ttl=CACHE) as pac_config:
plugin.log.info("Fetching Immunicity PAC file")
pac_data = urllib2.urlopen(PAC_URL).read()
pac_config["server"] = re.search(r"var proxyserver = '(.*)'", pac_data).group(1)
pac_config["domains"] = map(lambda x: x.replace(r"\Z(?ms)", ""), map(fnmatch.translate, re.findall(r"\"(.*?)\",", pac_data)))
_config = pac_config | conditional_block |
immunicity.py | import re
import fnmatch
import urllib2
from kodipopcorntime import plugin
from kodipopcorntime.caching import shelf
PAC_URL = "http://clientconfig.immunicity.org/pacs/all.pac"
CACHE = 1 * 3600 # 1 hour caching
_config = {}
def config():
global _config
if not _config:
with shelf("kodipopcorntime.immunicity.pac_config", ttl=CACHE) as pac_config:
plugin.log.info("Fetching Immunicity PAC file")
pac_data = urllib2.urlopen(PAC_URL).read()
pac_config["server"] = re.search(r"var proxyserver = '(.*)'", pac_data).group(1)
pac_config["domains"] = map(lambda x: x.replace(r"\Z(?ms)", ""), map(fnmatch.translate, re.findall(r"\"(.*?)\",", pac_data)))
_config = pac_config
return _config
def | (url):
conf = config()
for domain in conf["domains"]:
if re.search(domain, url):
return conf["server"]
| get_proxy_for | identifier_name |
immunicity.py | import re
import fnmatch
import urllib2
from kodipopcorntime import plugin
from kodipopcorntime.caching import shelf
PAC_URL = "http://clientconfig.immunicity.org/pacs/all.pac"
CACHE = 1 * 3600 # 1 hour caching
_config = {}
def config():
global _config
if not _config:
with shelf("kodipopcorntime.immunicity.pac_config", ttl=CACHE) as pac_config:
plugin.log.info("Fetching Immunicity PAC file")
pac_data = urllib2.urlopen(PAC_URL).read()
pac_config["server"] = re.search(r"var proxyserver = '(.*)'", pac_data).group(1)
pac_config["domains"] = map(lambda x: x.replace(r"\Z(?ms)", ""), map(fnmatch.translate, re.findall(r"\"(.*?)\",", pac_data)))
_config = pac_config
return _config
def get_proxy_for(url): | for domain in conf["domains"]:
if re.search(domain, url):
return conf["server"] | conf = config() | random_line_split |
recipe-440498.py | # Relative-refs.pyw
"""A short python script for repathing xrefs in Autocad."""
import win32com.client,os, os.path, tkFileDialog
from Tkinter import *
from tkMessageBox import askokcancel
from time import sleep
# Get a COM object for Autocad
acad = win32com.client.Dispatch("AutoCAD.Application")
def repath(filename):
print 'Repathing %s...' %filename
doc = acad.Documents.Open(filename)
blocks = doc.Database.Blocks # Internally xrefs are just blocks!
xrefs = [item for item in blocks if item.IsXRef]
if xrefs:
for xref in xrefs:
old_path = xref.Path
new_path = os.path.join('..\\x-ref\\',os.path.basename(old_path))
xref.Path = new_path
print 'Old path name was %s, new path name is %s.\n' %(old_path, new_path)
try:
doc.Close(True) # Close and save
except: # Something when wrong,
doc.Close(False) # close then report it
raise
class Logger:
"""A filelike object that prints its input on the screen."""
def __init__(self, logfile=None):
"""Takes one argument, a file like object for logging."""
print 'Starting logger...'
if not logfile:
self.logfile = open('relative-refs.log','w')
else:
self.logfile = logfile
sys.stderr = self # Super cheap logging facility...
sys.stdout = self # Just redirect output to a file.
print 'Logger running...'
def write(self, line):
sys.__stdout__.write(line)
self.logfile.write(line)
def close(self):
"""The close method restores stdout and stderr to normal."""
self.logfile.close()
sys.stderr = sys.__stderr__
sys.stdout = sys.__stdout__
class Tktextfile:
"""A file like interface to the Tk text widget."""
def __init__(self, root):
"""Create a scrollable text widget to be written to."""
self.root = root
self.text = Text(root,width=40,height=20)
self.text.pack(side=LEFT, expand=True, fill=BOTH)
scrollbar = Scrollbar(root)
scrollbar.pack(side=RIGHT,fill=Y)
self.text.configure(yscrollcommand=scrollbar.set)
scrollbar.config(command=self.text.yview)
self.text.focus()
def write(self, line):
"""Write method for file like widget."""
self.text.insert(INSERT, line)
self.text.see(END)
def close(self):
"""Fake close method."""
pass
if __name__ == '__main__':
if acad.Visible:
acad.Visible = False
root = Tk()
text = Tktextfile(root)
logger = Logger(text)
dir = tkFileDialog.askdirectory()
answer = askokcancel('RePath','Re path all dwg files in ' + dir + '?')
if answer:
for dirpath, subdirs, files in os.walk(dir):
for name in files:
ext = name.split('.')[-1] or ''
# We want dwg files which are not in the x-ref directory
if ext.lower() == 'dwg' and 'x-ref' not in dirpath.lower():
|
root.update()
acad.Visible = True
| drawing = os.path.join(dirpath, name)
try:
repath(drawing)
except:
print 'Unable to repath drawing %s!' %drawing | conditional_block |
recipe-440498.py | # Relative-refs.pyw
"""A short python script for repathing xrefs in Autocad."""
import win32com.client,os, os.path, tkFileDialog
from Tkinter import *
from tkMessageBox import askokcancel
from time import sleep
# Get a COM object for Autocad
acad = win32com.client.Dispatch("AutoCAD.Application")
def repath(filename):
print 'Repathing %s...' %filename
doc = acad.Documents.Open(filename)
blocks = doc.Database.Blocks # Internally xrefs are just blocks!
xrefs = [item for item in blocks if item.IsXRef]
if xrefs:
for xref in xrefs:
old_path = xref.Path
new_path = os.path.join('..\\x-ref\\',os.path.basename(old_path))
xref.Path = new_path
print 'Old path name was %s, new path name is %s.\n' %(old_path, new_path)
try:
doc.Close(True) # Close and save
except: # Something when wrong,
doc.Close(False) # close then report it
raise
class Logger:
"""A filelike object that prints its input on the screen."""
def __init__(self, logfile=None):
"""Takes one argument, a file like object for logging."""
print 'Starting logger...'
if not logfile:
self.logfile = open('relative-refs.log','w')
else:
self.logfile = logfile
sys.stderr = self # Super cheap logging facility...
sys.stdout = self # Just redirect output to a file.
print 'Logger running...'
def write(self, line):
sys.__stdout__.write(line)
self.logfile.write(line)
def close(self):
"""The close method restores stdout and stderr to normal."""
self.logfile.close()
sys.stderr = sys.__stderr__
sys.stdout = sys.__stdout__
class Tktextfile:
"""A file like interface to the Tk text widget."""
def __init__(self, root):
"""Create a scrollable text widget to be written to."""
self.root = root
self.text = Text(root,width=40,height=20)
self.text.pack(side=LEFT, expand=True, fill=BOTH)
scrollbar = Scrollbar(root)
scrollbar.pack(side=RIGHT,fill=Y)
self.text.configure(yscrollcommand=scrollbar.set)
scrollbar.config(command=self.text.yview)
self.text.focus()
def write(self, line):
|
def close(self):
"""Fake close method."""
pass
if __name__ == '__main__':
if acad.Visible:
acad.Visible = False
root = Tk()
text = Tktextfile(root)
logger = Logger(text)
dir = tkFileDialog.askdirectory()
answer = askokcancel('RePath','Re path all dwg files in ' + dir + '?')
if answer:
for dirpath, subdirs, files in os.walk(dir):
for name in files:
ext = name.split('.')[-1] or ''
# We want dwg files which are not in the x-ref directory
if ext.lower() == 'dwg' and 'x-ref' not in dirpath.lower():
drawing = os.path.join(dirpath, name)
try:
repath(drawing)
except:
print 'Unable to repath drawing %s!' %drawing
root.update()
acad.Visible = True
| """Write method for file like widget."""
self.text.insert(INSERT, line)
self.text.see(END) | identifier_body |
recipe-440498.py | # Relative-refs.pyw
"""A short python script for repathing xrefs in Autocad."""
import win32com.client,os, os.path, tkFileDialog
from Tkinter import *
from tkMessageBox import askokcancel
from time import sleep
# Get a COM object for Autocad
acad = win32com.client.Dispatch("AutoCAD.Application")
def repath(filename):
print 'Repathing %s...' %filename
doc = acad.Documents.Open(filename)
blocks = doc.Database.Blocks # Internally xrefs are just blocks!
xrefs = [item for item in blocks if item.IsXRef]
if xrefs:
for xref in xrefs:
old_path = xref.Path
new_path = os.path.join('..\\x-ref\\',os.path.basename(old_path))
xref.Path = new_path
print 'Old path name was %s, new path name is %s.\n' %(old_path, new_path)
try:
doc.Close(True) # Close and save
except: # Something when wrong,
doc.Close(False) # close then report it
raise
class | :
"""A filelike object that prints its input on the screen."""
def __init__(self, logfile=None):
"""Takes one argument, a file like object for logging."""
print 'Starting logger...'
if not logfile:
self.logfile = open('relative-refs.log','w')
else:
self.logfile = logfile
sys.stderr = self # Super cheap logging facility...
sys.stdout = self # Just redirect output to a file.
print 'Logger running...'
def write(self, line):
sys.__stdout__.write(line)
self.logfile.write(line)
def close(self):
"""The close method restores stdout and stderr to normal."""
self.logfile.close()
sys.stderr = sys.__stderr__
sys.stdout = sys.__stdout__
class Tktextfile:
"""A file like interface to the Tk text widget."""
def __init__(self, root):
"""Create a scrollable text widget to be written to."""
self.root = root
self.text = Text(root,width=40,height=20)
self.text.pack(side=LEFT, expand=True, fill=BOTH)
scrollbar = Scrollbar(root)
scrollbar.pack(side=RIGHT,fill=Y)
self.text.configure(yscrollcommand=scrollbar.set)
scrollbar.config(command=self.text.yview)
self.text.focus()
def write(self, line):
"""Write method for file like widget."""
self.text.insert(INSERT, line)
self.text.see(END)
def close(self):
"""Fake close method."""
pass
if __name__ == '__main__':
if acad.Visible:
acad.Visible = False
root = Tk()
text = Tktextfile(root)
logger = Logger(text)
dir = tkFileDialog.askdirectory()
answer = askokcancel('RePath','Re path all dwg files in ' + dir + '?')
if answer:
for dirpath, subdirs, files in os.walk(dir):
for name in files:
ext = name.split('.')[-1] or ''
# We want dwg files which are not in the x-ref directory
if ext.lower() == 'dwg' and 'x-ref' not in dirpath.lower():
drawing = os.path.join(dirpath, name)
try:
repath(drawing)
except:
print 'Unable to repath drawing %s!' %drawing
root.update()
acad.Visible = True
| Logger | identifier_name |
recipe-440498.py | # Relative-refs.pyw
"""A short python script for repathing xrefs in Autocad."""
import win32com.client,os, os.path, tkFileDialog
from Tkinter import *
from tkMessageBox import askokcancel
from time import sleep
# Get a COM object for Autocad
acad = win32com.client.Dispatch("AutoCAD.Application")
def repath(filename):
print 'Repathing %s...' %filename
doc = acad.Documents.Open(filename)
blocks = doc.Database.Blocks # Internally xrefs are just blocks!
xrefs = [item for item in blocks if item.IsXRef]
if xrefs:
for xref in xrefs:
old_path = xref.Path
new_path = os.path.join('..\\x-ref\\',os.path.basename(old_path))
xref.Path = new_path
print 'Old path name was %s, new path name is %s.\n' %(old_path, new_path)
try:
doc.Close(True) # Close and save
except: # Something when wrong,
doc.Close(False) # close then report it
raise
class Logger:
"""A filelike object that prints its input on the screen."""
def __init__(self, logfile=None):
"""Takes one argument, a file like object for logging."""
print 'Starting logger...'
if not logfile:
self.logfile = open('relative-refs.log','w')
else:
self.logfile = logfile
sys.stderr = self # Super cheap logging facility...
sys.stdout = self # Just redirect output to a file.
print 'Logger running...'
def write(self, line):
sys.__stdout__.write(line)
self.logfile.write(line)
def close(self):
"""The close method restores stdout and stderr to normal."""
self.logfile.close()
sys.stderr = sys.__stderr__
sys.stdout = sys.__stdout__
class Tktextfile:
"""A file like interface to the Tk text widget."""
def __init__(self, root):
"""Create a scrollable text widget to be written to."""
self.root = root
self.text = Text(root,width=40,height=20)
self.text.pack(side=LEFT, expand=True, fill=BOTH)
scrollbar = Scrollbar(root)
scrollbar.pack(side=RIGHT,fill=Y)
self.text.configure(yscrollcommand=scrollbar.set)
scrollbar.config(command=self.text.yview)
self.text.focus()
def write(self, line):
"""Write method for file like widget."""
self.text.insert(INSERT, line)
self.text.see(END)
def close(self):
"""Fake close method."""
pass
if __name__ == '__main__':
if acad.Visible:
acad.Visible = False
root = Tk()
text = Tktextfile(root)
logger = Logger(text)
dir = tkFileDialog.askdirectory()
answer = askokcancel('RePath','Re path all dwg files in ' + dir + '?')
if answer:
for dirpath, subdirs, files in os.walk(dir):
for name in files:
ext = name.split('.')[-1] or ''
# We want dwg files which are not in the x-ref directory
if ext.lower() == 'dwg' and 'x-ref' not in dirpath.lower(): | drawing = os.path.join(dirpath, name)
try:
repath(drawing)
except:
print 'Unable to repath drawing %s!' %drawing
root.update()
acad.Visible = True | random_line_split |
|
services.py | import logging
import requests
import xml.etree.ElementTree as ET
from django.utils.translation import gettext_lazy as _
from churchill.apps.currencies.models import CurrencyValue, Currency, CurrencyValueType
logger = logging.getLogger()
def get_default_currency_id() -> int:
currency, _ = Currency.objects.get_or_create(
name="United States Dollar", iso3="USD"
)
return currency.id
def get_currency_options() -> dict:
return {c.iso3: c.name for c in Currency.objects.all()}
def create_currency_pair(currency, node):
CurrencyValue.objects.create(
currency=currency,
type=CurrencyValueType.BUY,
value=float(node.find("purchase").text),
)
CurrencyValue.objects.create(
currency=currency,
type=CurrencyValueType.SELL,
value=float(node.find("sale").text),
)
def update_currencies():
byn_currency = Currency.objects.first(iso3="BYN")
eur_currency = Currency.objects.first(iso3="EUR")
rub_currency = Currency.objects.first(iso3="RUB")
if not any((byn_currency, eur_currency, rub_currency)):
logger.info(_("No currencies are setup"))
response = requests.get("https://www.mtbank.by/currxml.php?ver=2")
if response.status_code == 200:
for child in ET.fromstring(response.content): | code_to = node.find("codeTo").text
if (
byn_currency
and (code == "USD" and code_to == "BYN")
or (code == "BYN" and code_to == "USD")
):
create_currency_pair(byn_currency, node)
if (
eur_currency
and (code == "USD" and code_to == "EUR")
or (code == "EUR" and code_to == "USD")
):
create_currency_pair(eur_currency, node)
if (
rub_currency
and (code == "USD" and code_to == "RUB")
or (code == "RUB" and code_to == "USD")
):
create_currency_pair(rub_currency, node) | if child.attrib.get("id") == "168,768,968,868":
for node in child.findall("currency"):
code = node.find("code").text | random_line_split |
services.py | import logging
import requests
import xml.etree.ElementTree as ET
from django.utils.translation import gettext_lazy as _
from churchill.apps.currencies.models import CurrencyValue, Currency, CurrencyValueType
logger = logging.getLogger()
def | () -> int:
currency, _ = Currency.objects.get_or_create(
name="United States Dollar", iso3="USD"
)
return currency.id
def get_currency_options() -> dict:
return {c.iso3: c.name for c in Currency.objects.all()}
def create_currency_pair(currency, node):
CurrencyValue.objects.create(
currency=currency,
type=CurrencyValueType.BUY,
value=float(node.find("purchase").text),
)
CurrencyValue.objects.create(
currency=currency,
type=CurrencyValueType.SELL,
value=float(node.find("sale").text),
)
def update_currencies():
byn_currency = Currency.objects.first(iso3="BYN")
eur_currency = Currency.objects.first(iso3="EUR")
rub_currency = Currency.objects.first(iso3="RUB")
if not any((byn_currency, eur_currency, rub_currency)):
logger.info(_("No currencies are setup"))
response = requests.get("https://www.mtbank.by/currxml.php?ver=2")
if response.status_code == 200:
for child in ET.fromstring(response.content):
if child.attrib.get("id") == "168,768,968,868":
for node in child.findall("currency"):
code = node.find("code").text
code_to = node.find("codeTo").text
if (
byn_currency
and (code == "USD" and code_to == "BYN")
or (code == "BYN" and code_to == "USD")
):
create_currency_pair(byn_currency, node)
if (
eur_currency
and (code == "USD" and code_to == "EUR")
or (code == "EUR" and code_to == "USD")
):
create_currency_pair(eur_currency, node)
if (
rub_currency
and (code == "USD" and code_to == "RUB")
or (code == "RUB" and code_to == "USD")
):
create_currency_pair(rub_currency, node)
| get_default_currency_id | identifier_name |
services.py | import logging
import requests
import xml.etree.ElementTree as ET
from django.utils.translation import gettext_lazy as _
from churchill.apps.currencies.models import CurrencyValue, Currency, CurrencyValueType
logger = logging.getLogger()
def get_default_currency_id() -> int:
currency, _ = Currency.objects.get_or_create(
name="United States Dollar", iso3="USD"
)
return currency.id
def get_currency_options() -> dict:
return {c.iso3: c.name for c in Currency.objects.all()}
def create_currency_pair(currency, node):
CurrencyValue.objects.create(
currency=currency,
type=CurrencyValueType.BUY,
value=float(node.find("purchase").text),
)
CurrencyValue.objects.create(
currency=currency,
type=CurrencyValueType.SELL,
value=float(node.find("sale").text),
)
def update_currencies():
|
if (
eur_currency
and (code == "USD" and code_to == "EUR")
or (code == "EUR" and code_to == "USD")
):
create_currency_pair(eur_currency, node)
if (
rub_currency
and (code == "USD" and code_to == "RUB")
or (code == "RUB" and code_to == "USD")
):
create_currency_pair(rub_currency, node)
| byn_currency = Currency.objects.first(iso3="BYN")
eur_currency = Currency.objects.first(iso3="EUR")
rub_currency = Currency.objects.first(iso3="RUB")
if not any((byn_currency, eur_currency, rub_currency)):
logger.info(_("No currencies are setup"))
response = requests.get("https://www.mtbank.by/currxml.php?ver=2")
if response.status_code == 200:
for child in ET.fromstring(response.content):
if child.attrib.get("id") == "168,768,968,868":
for node in child.findall("currency"):
code = node.find("code").text
code_to = node.find("codeTo").text
if (
byn_currency
and (code == "USD" and code_to == "BYN")
or (code == "BYN" and code_to == "USD")
):
create_currency_pair(byn_currency, node) | identifier_body |
services.py | import logging
import requests
import xml.etree.ElementTree as ET
from django.utils.translation import gettext_lazy as _
from churchill.apps.currencies.models import CurrencyValue, Currency, CurrencyValueType
logger = logging.getLogger()
def get_default_currency_id() -> int:
currency, _ = Currency.objects.get_or_create(
name="United States Dollar", iso3="USD"
)
return currency.id
def get_currency_options() -> dict:
return {c.iso3: c.name for c in Currency.objects.all()}
def create_currency_pair(currency, node):
CurrencyValue.objects.create(
currency=currency,
type=CurrencyValueType.BUY,
value=float(node.find("purchase").text),
)
CurrencyValue.objects.create(
currency=currency,
type=CurrencyValueType.SELL,
value=float(node.find("sale").text),
)
def update_currencies():
byn_currency = Currency.objects.first(iso3="BYN")
eur_currency = Currency.objects.first(iso3="EUR")
rub_currency = Currency.objects.first(iso3="RUB")
if not any((byn_currency, eur_currency, rub_currency)):
logger.info(_("No currencies are setup"))
response = requests.get("https://www.mtbank.by/currxml.php?ver=2")
if response.status_code == 200:
for child in ET.fromstring(response.content):
| rub_currency
and (code == "USD" and code_to == "RUB")
or (code == "RUB" and code_to == "USD")
):
create_currency_pair(rub_currency, node)
| if child.attrib.get("id") == "168,768,968,868":
for node in child.findall("currency"):
code = node.find("code").text
code_to = node.find("codeTo").text
if (
byn_currency
and (code == "USD" and code_to == "BYN")
or (code == "BYN" and code_to == "USD")
):
create_currency_pair(byn_currency, node)
if (
eur_currency
and (code == "USD" and code_to == "EUR")
or (code == "EUR" and code_to == "USD")
):
create_currency_pair(eur_currency, node)
if ( | conditional_block |
fix_csxml_character_encoding.py | import sys
import logging
codec_options = ['utf-8', 'latin_1']
logger = logging.getLogger(__name__)
def try_decode(byte_string, codec):
try:
s = byte_string.decode(codec)
return s
except:
return None
def shortest_string(strings):
best_string = None
best_length = None
for s in strings:
if best_string is None or len(s) < best_length:
best_string = s
best_length = len(s)
return best_string
def fix_character_encoding(input_file, output_file):
with open(input_file, 'rb') as f_in:
with open(output_file, 'wb') as f_out:
for line in f_in:
# Try to decode with both latin_1 and utf-8
decoded = [try_decode(line, c) for c in codec_options]
decoded = [d for d in decoded if d is not None]
if len(decoded) == 0:
# Hopefully at least one codec worked
|
else:
# If more than one, choose the codec that gives the best
# length
chosen_string = shortest_string(decoded)
# Write result as ascii, with non-ascii characters escaped
f_out.write(chosen_string.encode('utf-8'))
if __name__ == '__main__':
args = sys.argv[1:]
if len(args) != 2:
logger.error('Expected two arguments: the input file'
' and the output file')
sys.exit(1)
input_file = args[0]
output_file = args[1]
fix_character_encoding(input_file, output_file)
| logger.info('Could not decode: %s' % line)
sys.exit(1) | conditional_block |
fix_csxml_character_encoding.py | import sys
import logging
codec_options = ['utf-8', 'latin_1']
logger = logging.getLogger(__name__)
def try_decode(byte_string, codec):
try:
s = byte_string.decode(codec)
return s
except:
return None
def shortest_string(strings):
|
def fix_character_encoding(input_file, output_file):
with open(input_file, 'rb') as f_in:
with open(output_file, 'wb') as f_out:
for line in f_in:
# Try to decode with both latin_1 and utf-8
decoded = [try_decode(line, c) for c in codec_options]
decoded = [d for d in decoded if d is not None]
if len(decoded) == 0:
# Hopefully at least one codec worked
logger.info('Could not decode: %s' % line)
sys.exit(1)
else:
# If more than one, choose the codec that gives the best
# length
chosen_string = shortest_string(decoded)
# Write result as ascii, with non-ascii characters escaped
f_out.write(chosen_string.encode('utf-8'))
if __name__ == '__main__':
args = sys.argv[1:]
if len(args) != 2:
logger.error('Expected two arguments: the input file'
' and the output file')
sys.exit(1)
input_file = args[0]
output_file = args[1]
fix_character_encoding(input_file, output_file)
| best_string = None
best_length = None
for s in strings:
if best_string is None or len(s) < best_length:
best_string = s
best_length = len(s)
return best_string | identifier_body |
fix_csxml_character_encoding.py | import sys
import logging
codec_options = ['utf-8', 'latin_1']
logger = logging.getLogger(__name__)
def try_decode(byte_string, codec):
try:
s = byte_string.decode(codec)
return s
except:
return None
def | (strings):
best_string = None
best_length = None
for s in strings:
if best_string is None or len(s) < best_length:
best_string = s
best_length = len(s)
return best_string
def fix_character_encoding(input_file, output_file):
with open(input_file, 'rb') as f_in:
with open(output_file, 'wb') as f_out:
for line in f_in:
# Try to decode with both latin_1 and utf-8
decoded = [try_decode(line, c) for c in codec_options]
decoded = [d for d in decoded if d is not None]
if len(decoded) == 0:
# Hopefully at least one codec worked
logger.info('Could not decode: %s' % line)
sys.exit(1)
else:
# If more than one, choose the codec that gives the best
# length
chosen_string = shortest_string(decoded)
# Write result as ascii, with non-ascii characters escaped
f_out.write(chosen_string.encode('utf-8'))
if __name__ == '__main__':
args = sys.argv[1:]
if len(args) != 2:
logger.error('Expected two arguments: the input file'
' and the output file')
sys.exit(1)
input_file = args[0]
output_file = args[1]
fix_character_encoding(input_file, output_file)
| shortest_string | identifier_name |
fix_csxml_character_encoding.py | import sys
import logging
codec_options = ['utf-8', 'latin_1']
logger = logging.getLogger(__name__)
def try_decode(byte_string, codec):
try:
s = byte_string.decode(codec)
return s
except:
return None
def shortest_string(strings):
best_string = None
best_length = None
for s in strings:
if best_string is None or len(s) < best_length:
best_string = s
best_length = len(s)
return best_string
def fix_character_encoding(input_file, output_file):
with open(input_file, 'rb') as f_in:
with open(output_file, 'wb') as f_out:
for line in f_in:
# Try to decode with both latin_1 and utf-8
decoded = [try_decode(line, c) for c in codec_options]
decoded = [d for d in decoded if d is not None]
if len(decoded) == 0:
# Hopefully at least one codec worked
logger.info('Could not decode: %s' % line)
sys.exit(1)
else:
# If more than one, choose the codec that gives the best
# length
chosen_string = shortest_string(decoded)
| args = sys.argv[1:]
if len(args) != 2:
logger.error('Expected two arguments: the input file'
' and the output file')
sys.exit(1)
input_file = args[0]
output_file = args[1]
fix_character_encoding(input_file, output_file) | # Write result as ascii, with non-ascii characters escaped
f_out.write(chosen_string.encode('utf-8'))
if __name__ == '__main__': | random_line_split |
index.tsx | /**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
* | * under the License.
*/
import { DatePicker as AntdDatePicker } from 'antd';
import { styled } from '@superset-ui/core';
const AntdRangePicker = AntdDatePicker.RangePicker;
export const RangePicker = styled(AntdRangePicker)`
border-radius: ${({ theme }) => theme.gridUnit}px;
`;
export const DatePicker = AntdDatePicker; | * Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations | random_line_split |
version.py | #!/usr/bin/python3
"""Script to determine the Pywikibot version (tag, revision and date).
.. versionchanged:: 7.0
version script was moved to the framework scripts folder
"""
#
# (C) Pywikibot team, 2007-2021
#
# Distributed under the terms of the MIT license.
#
import codecs
import os
import sys
import pywikibot
from pywikibot.version import get_toolforge_hostname, getversion
class DummyModule:
"""Fake module instance."""
__version__ = 'n/a'
try:
import setuptools
except ImportError:
setuptools = DummyModule()
try:
import mwparserfromhell
except ImportError:
mwparserfromhell = DummyModule()
try:
import wikitextparser
except ImportError:
wikitextparser = DummyModule()
try:
import requests
except ImportError:
requests = DummyModule()
WMF_CACERT = 'MIIDxTCCAq2gAwIBAgIQAqxcJmoLQJuPC3nyrkYldzANBgkqhkiG9w0BAQUFADBs'
def main(*args: str) -> None:
"""Print pywikibot version and important settings."""
pywikibot.output('Pywikibot: ' + getversion())
pywikibot.output('Release version: ' + pywikibot.__version__)
pywikibot.output('setuptools version: ' + setuptools.__version__)
pywikibot.output('mwparserfromhell version: '
+ mwparserfromhell.__version__)
pywikibot.output('wikitextparser version: ' + wikitextparser.__version__)
pywikibot.output('requests version: ' + requests.__version__)
has_wikimedia_cert = False
if (not hasattr(requests, 'certs')
or not hasattr(requests.certs, 'where')
or not callable(requests.certs.where)):
pywikibot.output(' cacerts: not defined')
elif not os.path.isfile(requests.certs.where()):
pywikibot.output(' cacerts: {} (missing)'.format(
requests.certs.where()))
else:
pywikibot.output(' cacerts: ' + requests.certs.where())
with codecs.open(requests.certs.where(), 'r', 'utf-8') as cert_file:
text = cert_file.read()
if WMF_CACERT in text:
has_wikimedia_cert = True
pywikibot.output(' certificate test: {}'
.format('ok' if has_wikimedia_cert else 'not ok'))
if not has_wikimedia_cert:
pywikibot.output(' Please reinstall requests!')
pywikibot.output('Python: ' + sys.version)
toolforge_env_hostname = get_toolforge_hostname()
if toolforge_env_hostname:
pywikibot.output('Toolforge hostname: ' + toolforge_env_hostname)
# check environment settings
settings = {key for key in os.environ if key.startswith('PYWIKIBOT')}
settings.update(['PYWIKIBOT_DIR', 'PYWIKIBOT_DIR_PWB',
'PYWIKIBOT_NO_USER_CONFIG'])
for environ_name in sorted(settings):
pywikibot.output(
'{}: {}'.format(environ_name,
os.environ.get(environ_name, 'Not set') or "''"))
pywikibot.output('Config base dir: ' + pywikibot.config.base_dir)
for family, usernames in pywikibot.config.usernames.items():
if not usernames:
|
pywikibot.output('Usernames for family {!r}:'.format(family))
for lang, username in usernames.items():
pywikibot.output('\t{}: {}'.format(lang, username))
if __name__ == '__main__':
main()
| continue | conditional_block |
version.py | #!/usr/bin/python3
"""Script to determine the Pywikibot version (tag, revision and date).
.. versionchanged:: 7.0
version script was moved to the framework scripts folder
"""
#
# (C) Pywikibot team, 2007-2021
#
# Distributed under the terms of the MIT license.
#
import codecs
import os
import sys
import pywikibot
from pywikibot.version import get_toolforge_hostname, getversion
class DummyModule:
"""Fake module instance."""
__version__ = 'n/a'
try:
import setuptools
except ImportError:
setuptools = DummyModule()
try:
import mwparserfromhell
except ImportError:
mwparserfromhell = DummyModule()
try:
import wikitextparser
except ImportError:
wikitextparser = DummyModule()
try:
import requests
except ImportError:
requests = DummyModule()
WMF_CACERT = 'MIIDxTCCAq2gAwIBAgIQAqxcJmoLQJuPC3nyrkYldzANBgkqhkiG9w0BAQUFADBs'
def main(*args: str) -> None:
"""Print pywikibot version and important settings."""
pywikibot.output('Pywikibot: ' + getversion())
pywikibot.output('Release version: ' + pywikibot.__version__)
pywikibot.output('setuptools version: ' + setuptools.__version__)
pywikibot.output('mwparserfromhell version: '
+ mwparserfromhell.__version__)
pywikibot.output('wikitextparser version: ' + wikitextparser.__version__)
pywikibot.output('requests version: ' + requests.__version__)
has_wikimedia_cert = False
if (not hasattr(requests, 'certs')
or not hasattr(requests.certs, 'where')
or not callable(requests.certs.where)):
pywikibot.output(' cacerts: not defined')
elif not os.path.isfile(requests.certs.where()):
pywikibot.output(' cacerts: {} (missing)'.format(
requests.certs.where()))
else:
pywikibot.output(' cacerts: ' + requests.certs.where())
with codecs.open(requests.certs.where(), 'r', 'utf-8') as cert_file:
text = cert_file.read()
if WMF_CACERT in text:
has_wikimedia_cert = True
pywikibot.output(' certificate test: {}'
.format('ok' if has_wikimedia_cert else 'not ok'))
if not has_wikimedia_cert:
pywikibot.output(' Please reinstall requests!')
pywikibot.output('Python: ' + sys.version)
toolforge_env_hostname = get_toolforge_hostname()
if toolforge_env_hostname:
pywikibot.output('Toolforge hostname: ' + toolforge_env_hostname)
# check environment settings
settings = {key for key in os.environ if key.startswith('PYWIKIBOT')}
settings.update(['PYWIKIBOT_DIR', 'PYWIKIBOT_DIR_PWB',
'PYWIKIBOT_NO_USER_CONFIG'])
for environ_name in sorted(settings):
pywikibot.output( | if not usernames:
continue
pywikibot.output('Usernames for family {!r}:'.format(family))
for lang, username in usernames.items():
pywikibot.output('\t{}: {}'.format(lang, username))
if __name__ == '__main__':
main() | '{}: {}'.format(environ_name,
os.environ.get(environ_name, 'Not set') or "''"))
pywikibot.output('Config base dir: ' + pywikibot.config.base_dir)
for family, usernames in pywikibot.config.usernames.items(): | random_line_split |
version.py | #!/usr/bin/python3
"""Script to determine the Pywikibot version (tag, revision and date).
.. versionchanged:: 7.0
version script was moved to the framework scripts folder
"""
#
# (C) Pywikibot team, 2007-2021
#
# Distributed under the terms of the MIT license.
#
import codecs
import os
import sys
import pywikibot
from pywikibot.version import get_toolforge_hostname, getversion
class DummyModule:
"""Fake module instance."""
__version__ = 'n/a'
try:
import setuptools
except ImportError:
setuptools = DummyModule()
try:
import mwparserfromhell
except ImportError:
mwparserfromhell = DummyModule()
try:
import wikitextparser
except ImportError:
wikitextparser = DummyModule()
try:
import requests
except ImportError:
requests = DummyModule()
WMF_CACERT = 'MIIDxTCCAq2gAwIBAgIQAqxcJmoLQJuPC3nyrkYldzANBgkqhkiG9w0BAQUFADBs'
def | (*args: str) -> None:
"""Print pywikibot version and important settings."""
pywikibot.output('Pywikibot: ' + getversion())
pywikibot.output('Release version: ' + pywikibot.__version__)
pywikibot.output('setuptools version: ' + setuptools.__version__)
pywikibot.output('mwparserfromhell version: '
+ mwparserfromhell.__version__)
pywikibot.output('wikitextparser version: ' + wikitextparser.__version__)
pywikibot.output('requests version: ' + requests.__version__)
has_wikimedia_cert = False
if (not hasattr(requests, 'certs')
or not hasattr(requests.certs, 'where')
or not callable(requests.certs.where)):
pywikibot.output(' cacerts: not defined')
elif not os.path.isfile(requests.certs.where()):
pywikibot.output(' cacerts: {} (missing)'.format(
requests.certs.where()))
else:
pywikibot.output(' cacerts: ' + requests.certs.where())
with codecs.open(requests.certs.where(), 'r', 'utf-8') as cert_file:
text = cert_file.read()
if WMF_CACERT in text:
has_wikimedia_cert = True
pywikibot.output(' certificate test: {}'
.format('ok' if has_wikimedia_cert else 'not ok'))
if not has_wikimedia_cert:
pywikibot.output(' Please reinstall requests!')
pywikibot.output('Python: ' + sys.version)
toolforge_env_hostname = get_toolforge_hostname()
if toolforge_env_hostname:
pywikibot.output('Toolforge hostname: ' + toolforge_env_hostname)
# check environment settings
settings = {key for key in os.environ if key.startswith('PYWIKIBOT')}
settings.update(['PYWIKIBOT_DIR', 'PYWIKIBOT_DIR_PWB',
'PYWIKIBOT_NO_USER_CONFIG'])
for environ_name in sorted(settings):
pywikibot.output(
'{}: {}'.format(environ_name,
os.environ.get(environ_name, 'Not set') or "''"))
pywikibot.output('Config base dir: ' + pywikibot.config.base_dir)
for family, usernames in pywikibot.config.usernames.items():
if not usernames:
continue
pywikibot.output('Usernames for family {!r}:'.format(family))
for lang, username in usernames.items():
pywikibot.output('\t{}: {}'.format(lang, username))
if __name__ == '__main__':
main()
| main | identifier_name |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.