file_name
large_stringlengths 4
140
| prefix
large_stringlengths 0
12.1k
| suffix
large_stringlengths 0
12k
| middle
large_stringlengths 0
7.51k
| fim_type
large_stringclasses 4
values |
---|---|---|---|---|
demo.py | (modelID):
print "Loading parameters for model", modelID
selection=raw_input('Would you like to download NWB data for model? [Y/N] ')
if selection[0] == 'y' or selection[0] == 'Y':
currModel = Model(modelID, cache_stim = True)
if selection[0] == 'n' or selection[0] == 'N':
currModel = Model(modelID, cache_stim = False)
currModel.init_model()
while(True):
print "Initialized biophysical model", modelID
print '''
Please select from the following options:
1 - Run test pulse on model
2 - Fit model parameter to data
3 - Display static neuron model
4 - Visualize model dynamics
5 - Quit
'''
try:
selection=int(raw_input('Please choose an option above: '))
except ValueError:
print "Invalid selection."
continue
# test pulse example
if selection == 1:
# Run the model with a test pulse of the 'long square' type
print "Running model with a long square current injection pulse of 210pA"
output = currModel.long_square(0.21)
currModel.plot_output()
# fit parameter example
elif selection == 2:
if not currModel.bp.cache_stimulus:
print "Current model was not instantiated with NWB data cached. Please reload the current model and cache experimental stimulus data."
continue
print "Fitting somatic sodium conductance for model", modelID, "to experimental data in sweep 41."
print "Please be patient, this may take some time."
# Define which section and which parameter to fit.
# Here we'll fit the somatic sodium conductance.
currModel.set_fit_section('soma', 0)
currModel.set_parameter_to_fit('gbar_NaV')
# Running the model with an NWB pulse as stimulus takes a
# very long time because of the high sampling rate.
# As a computationally-cheaper approximation for stimuli of
# type Long Square pulse, we can rebuild the stimulus with the
# default (lower) sampling rate in h.IClamp
# currModel.run_nwb_pulse(41) # too slow
output = currModel.long_square(0.21)
# Set the experimental reference sweep and set up the variables for the objective function
currModel.set_reference_sweep(ref_index=41)
currModel.set_up_objective(measure='spike frequency')
# Use SciPy's minimize functions to fit the specified parameter
#results = minimize(currModel.objective_function, currModel.theta, method='Nelder-Mead', tol=1e-3)
#results = minimize(currModel.objective_function, currModel.theta, method='Powell', tol=1e-3)
#results = minimize(currModel.objective_function, currModel.theta, method='COBYLA', tol=1e-5)
currModel.gradient_descent(alpha=0.00005, epsilon=0.001, threshold=0.01, max_cycles=1000)
currModel.plot_fit()
output = currModel.long_square(0.21)
currModel.plot_output()
times = np.array(output['t'])/1000
spikes = detect_putative_spikes(np.array(output['v']), times, 0.1, 1.1)
avg_rate = currModel.average_rate_from_delays(times, spikes, 0.1, 1.1)
print "spike rate for theta of", currModel.theta, ":", avg_rate
# static visualization example
elif selection == 3:
run_visualization(currModel)
elif selection == 4:
run_visualization(currModel, show_simulation_dynamics = True)
elif selection == 5:
quit()
else:
print "Invalid selection."
continue
def run_visualization(currModel, show_simulation_dynamics = False):
print "Setting up visualization..."
morphology = currModel.get_reconstruction()
# Prepare model coordinates for uploading to OpenGL.
tempIndices = []
tempVertices = []
n_index = 0
tempX = []
tempY = []
tempZ = []
tempCol = []
if not show_simulation_dynamics:
print '''
Soma - Red
Axon - Green
Dendrites - Blue
Apical Dendrites - Purple'''
# array of colors to denote individual compartment types
compartmentColors=[[0.0,0.0,0.0,0.0], # padding for index convenience
[1.0, 0.0, 0.0, 1.0], #1: soma - red
[0.0, 1.0, 0.0, 1.0], #2: axon - green
[0.0, 0.0, 1.0, 1.0], #3: dendrites - blue
[1.0, 0.0, 1.0, 1.0]] #4: apical dendrites - purple
color_dim = 4
# used to set up section monitoring for visualization of dynamics
compartmentNames=['none', # padding for index convenience
'soma', #1: soma
'axon', #2: axon
'dend', #3: dendrites - blue
'dend'] #4: apical dendrites - purple
sectionIndices=[0,0,0,0,0]
segmentsPerSection = {}
sec_name = ''
# initialize storage arrays for each vertex.
index = 0
n_compartments = len(morphology.compartment_list)
tempX = [0] * n_compartments
tempY = [0] * n_compartments
tempZ = [0] * n_compartments
tempCol = [0] * n_compartments * color_dim
for n in morphology.compartment_list:
# add parent coords
tempX[n['id']] = n['x']
tempY[n['id']] = -n['y']
tempZ[n['id']] = n['z']
# add color data for parent
col_i = 0
offset = n['id']*color_dim
for cval in compartmentColors[n['type']]:
tempCol[offset+col_i] = cval
col_i += 1
# if at a branch point or an end of a section, set up a vector to monitor that segment's voltage
type = compartmentNames[n['type']]
sec_index = sectionIndices[n['type']]
if not (len(morphology.children_of(n)) == 1): #either branch pt or end
sec_name = type + '[' + str(sec_index) + ']'
sectionIndices[n['type']] += 1
currModel.monitor_section_voltage(type, sec_index)
segmentsPerSection[sec_name] = 1
else:
segmentsPerSection[sec_name] += 1
index += 1
for c in morphology.children_of(n):
# add child coods
tempX[c['id']] = c['x']
tempY[c['id']] = -c['y']
tempZ[c['id']] = c['z']
# add index data:
# draw from parent to child, for each child
tempIndices.append(n['id'])
tempIndices.append(c['id'])
index += 1
# add color data for child
col_i = 0
offset = c['id']*color_dim
for cval in compartmentColors[c['type']]:
tempCol[offset+col_i] = cval
col_i += 1
segmentsPerSection[sec_name] += 1
# get ranges for scaling
maxX = max(tempX)
maxY = max(tempY)
maxZ = max(tempZ)
minX = min(tempX)
minY = min(tempY)
minZ = min(tempZ)
xHalfRange = (maxX - minX)/2.0
yHalfRange = (maxY - minY)/2.0
zHalfRange = (maxZ - minZ)/2.0
longestDimLen = max(xHalfRange, yHalfRange, zHalfRange)
# center coords about 0,0,0, with range -1 to 1
tempX = [((((x-minX)*(2*xHalfRange))/(2*xHalfRange)) - xHalfRange)/longestDimLen for x in tempX]
tempY = [((((y-minY)*(2*yHalfRange))/(2*yHalfRange)) - yHalfRange)/longestDimLen for y in tempY]
tempZ = [((((z-minZ)*(2*zHalfRange))/(2*zHalfRange)) - zHalfRange)/longestDimLen for z in tempZ]
# convert everything to a numpy array so OpenGL can use it
indexData = np.array(tempIndices, dtype='uint16')
vertexData = np.array([tempX,tempY,tempZ | main | identifier_name |
|
demo.py | "Initialized biophysical model", modelID
print '''
Please select from the following options:
1 - Run test pulse on model
2 - Fit model parameter to data
3 - Display static neuron model
4 - Visualize model dynamics
5 - Quit
'''
try:
selection=int(raw_input('Please choose an option above: '))
except ValueError:
print "Invalid selection."
continue
# test pulse example
if selection == 1:
# Run the model with a test pulse of the 'long square' type
print "Running model with a long square current injection pulse of 210pA"
output = currModel.long_square(0.21)
currModel.plot_output()
# fit parameter example
elif selection == 2:
if not currModel.bp.cache_stimulus:
print "Current model was not instantiated with NWB data cached. Please reload the current model and cache experimental stimulus data."
continue
print "Fitting somatic sodium conductance for model", modelID, "to experimental data in sweep 41."
print "Please be patient, this may take some time."
# Define which section and which parameter to fit.
# Here we'll fit the somatic sodium conductance.
currModel.set_fit_section('soma', 0)
currModel.set_parameter_to_fit('gbar_NaV')
# Running the model with an NWB pulse as stimulus takes a
# very long time because of the high sampling rate.
# As a computationally-cheaper approximation for stimuli of
# type Long Square pulse, we can rebuild the stimulus with the
# default (lower) sampling rate in h.IClamp
# currModel.run_nwb_pulse(41) # too slow
output = currModel.long_square(0.21)
# Set the experimental reference sweep and set up the variables for the objective function
currModel.set_reference_sweep(ref_index=41)
currModel.set_up_objective(measure='spike frequency')
# Use SciPy's minimize functions to fit the specified parameter
#results = minimize(currModel.objective_function, currModel.theta, method='Nelder-Mead', tol=1e-3)
#results = minimize(currModel.objective_function, currModel.theta, method='Powell', tol=1e-3)
#results = minimize(currModel.objective_function, currModel.theta, method='COBYLA', tol=1e-5)
currModel.gradient_descent(alpha=0.00005, epsilon=0.001, threshold=0.01, max_cycles=1000)
currModel.plot_fit()
output = currModel.long_square(0.21)
currModel.plot_output()
times = np.array(output['t'])/1000
spikes = detect_putative_spikes(np.array(output['v']), times, 0.1, 1.1)
avg_rate = currModel.average_rate_from_delays(times, spikes, 0.1, 1.1)
print "spike rate for theta of", currModel.theta, ":", avg_rate
# static visualization example
elif selection == 3:
run_visualization(currModel)
elif selection == 4:
run_visualization(currModel, show_simulation_dynamics = True)
elif selection == 5:
quit()
else:
print "Invalid selection."
continue
def run_visualization(currModel, show_simulation_dynamics = False):
print "Setting up visualization..."
morphology = currModel.get_reconstruction()
# Prepare model coordinates for uploading to OpenGL.
tempIndices = []
tempVertices = []
n_index = 0
tempX = []
tempY = []
tempZ = []
tempCol = []
if not show_simulation_dynamics:
print '''
Soma - Red
Axon - Green
Dendrites - Blue
Apical Dendrites - Purple'''
# array of colors to denote individual compartment types
compartmentColors=[[0.0,0.0,0.0,0.0], # padding for index convenience
[1.0, 0.0, 0.0, 1.0], #1: soma - red
[0.0, 1.0, 0.0, 1.0], #2: axon - green
[0.0, 0.0, 1.0, 1.0], #3: dendrites - blue
[1.0, 0.0, 1.0, 1.0]] #4: apical dendrites - purple
color_dim = 4
# used to set up section monitoring for visualization of dynamics
compartmentNames=['none', # padding for index convenience
'soma', #1: soma
'axon', #2: axon
'dend', #3: dendrites - blue
'dend'] #4: apical dendrites - purple
sectionIndices=[0,0,0,0,0]
segmentsPerSection = {}
sec_name = ''
# initialize storage arrays for each vertex.
index = 0
n_compartments = len(morphology.compartment_list)
tempX = [0] * n_compartments
tempY = [0] * n_compartments
tempZ = [0] * n_compartments
tempCol = [0] * n_compartments * color_dim
for n in morphology.compartment_list:
# add parent coords
tempX[n['id']] = n['x']
tempY[n['id']] = -n['y']
tempZ[n['id']] = n['z']
# add color data for parent
col_i = 0
offset = n['id']*color_dim
for cval in compartmentColors[n['type']]:
|
# if at a branch point or an end of a section, set up a vector to monitor that segment's voltage
type = compartmentNames[n['type']]
sec_index = sectionIndices[n['type']]
if not (len(morphology.children_of(n)) == 1): #either branch pt or end
sec_name = type + '[' + str(sec_index) + ']'
sectionIndices[n['type']] += 1
currModel.monitor_section_voltage(type, sec_index)
segmentsPerSection[sec_name] = 1
else:
segmentsPerSection[sec_name] += 1
index += 1
for c in morphology.children_of(n):
# add child coods
tempX[c['id']] = c['x']
tempY[c['id']] = -c['y']
tempZ[c['id']] = c['z']
# add index data:
# draw from parent to child, for each child
tempIndices.append(n['id'])
tempIndices.append(c['id'])
index += 1
# add color data for child
col_i = 0
offset = c['id']*color_dim
for cval in compartmentColors[c['type']]:
tempCol[offset+col_i] = cval
col_i += 1
segmentsPerSection[sec_name] += 1
# get ranges for scaling
maxX = max(tempX)
maxY = max(tempY)
maxZ = max(tempZ)
minX = min(tempX)
minY = min(tempY)
minZ = min(tempZ)
xHalfRange = (maxX - minX)/2.0
yHalfRange = (maxY - minY)/2.0
zHalfRange = (maxZ - minZ)/2.0
longestDimLen = max(xHalfRange, yHalfRange, zHalfRange)
# center coords about 0,0,0, with range -1 to 1
tempX = [((((x-minX)*(2*xHalfRange))/(2*xHalfRange)) - xHalfRange)/longestDimLen for x in tempX]
tempY = [((((y-minY)*(2*yHalfRange))/(2*yHalfRange)) - yHalfRange)/longestDimLen for y in tempY]
tempZ = [((((z-minZ)*(2*zHalfRange))/(2*zHalfRange)) - zHalfRange)/longestDimLen for z in tempZ]
# convert everything to a numpy array so OpenGL can use it
indexData = np.array(tempIndices, dtype='uint16')
vertexData = np.array([tempX,tempY,tempZ], dtype='float32')
tempCol = np.array(tempCol, dtype='float32')
vertexData = np.append(vertexData.transpose().flatten(), tempCol)
#################### /Preparing Model Coords
# Set up the Visualization instance
n_vertices = len(tempX)
currVis = Visualization(data=vertexData, indices=indexData, nVert=n_vertices, colorDim=color_dim)
if show_simulation_dynamics:
currModel.run_test_pulse(amp=0.25, delay=20.0 | tempCol[offset+col_i] = cval
col_i += 1 | conditional_block |
fs.rs | x| Error::from_sys(x))
}
}
impl Drop for File {
fn drop(&mut self) {
let _ = close(self.fd);
}
}
#[derive(Copy, Clone, Eq, PartialEq)]
pub struct FileType {
dir: bool,
file: bool,
}
impl FileType {
pub fn is_dir(&self) -> bool {
self.dir
}
pub fn is_file(&self) -> bool {
self.file
}
pub fn is_symlink(&self) -> bool {
false
}
}
impl ::os::unix::fs::FileTypeExt for FileType {
fn is_block_device(&self) -> bool { false }
fn is_char_device(&self) -> bool { false }
fn is_fifo(&self) -> bool { false }
fn is_socket(&self) -> bool { false }
}
pub struct OpenOptions {
read: bool,
write: bool,
append: bool,
create: bool,
truncate: bool,
mode: u16,
}
impl OpenOptions {
pub fn new() -> OpenOptions {
OpenOptions {
read: false,
write: false,
append: false,
create: false,
truncate: false,
mode: 0,
}
}
pub fn read(&mut self, read: bool) -> &mut OpenOptions {
self.read = read;
self
}
pub fn write(&mut self, write: bool) -> &mut OpenOptions {
self.write = write;
self
}
pub fn append(&mut self, append: bool) -> &mut OpenOptions {
self.append = append;
self
}
pub fn create(&mut self, create: bool) -> &mut OpenOptions {
self.create = create;
self
}
pub fn truncate(&mut self, truncate: bool) -> &mut OpenOptions {
self.truncate = truncate;
self
}
pub fn open<P: AsRef<Path>>(&self, path: P) -> Result<File> {
let mut flags = 0;
if self.read && self.write {
flags |= O_RDWR;
} else if self.read {
flags |= O_RDONLY;
} else if self.write {
flags |= O_WRONLY;
}
if self.append {
flags |= O_APPEND;
}
if self.create {
flags |= O_CREAT;
}
if self.truncate {
flags |= O_TRUNC;
}
flags |= (self.mode & MODE_PERM) as usize;
let path_str = path.as_ref().as_os_str().as_inner();
open(path_str, flags).map(|fd| unsafe { File::from_raw_fd(fd) }).map_err(|x| Error::from_sys(x))
}
}
impl ::os::unix::fs::OpenOptionsExt for OpenOptions {
fn mode(&mut self, mode: u32) -> &mut Self {
self.mode = mode as u16;
self
}
}
pub struct Metadata {
stat: Stat
}
impl Metadata {
pub fn file_type(&self) -> FileType {
FileType {
dir: self.stat.st_mode & MODE_DIR == MODE_DIR,
file: self.stat.st_mode & MODE_FILE == MODE_FILE
}
}
pub fn is_dir(&self) -> bool {
self.stat.st_mode & MODE_DIR == MODE_DIR
}
pub fn is_file(&self) -> bool {
self.stat.st_mode & MODE_FILE == MODE_FILE
}
pub fn len(&self) -> u64 {
self.stat.st_size
}
pub fn permissions(&self) -> Permissions {
Permissions {
mode: self.stat.st_mode & MODE_PERM
}
}
}
impl ::os::unix::fs::MetadataExt for Metadata {
fn mode(&self) -> u32 {
self.stat.st_mode as u32
}
fn uid(&self) -> u32 {
self.stat.st_uid
}
fn gid(&self) -> u32 {
self.stat.st_gid
}
fn size(&self) -> u64 {
self.stat.st_size
}
}
pub struct Permissions {
mode: u16
}
impl Permissions {
pub fn readonly(&self) -> bool {
self.mode & 0o222 == 0
}
pub fn set_readonly(&mut self, readonly: bool) {
if readonly {
self.mode &= !0o222;
} else {
self.mode |= 0o222;
}
}
}
impl ::os::unix::fs::PermissionsExt for Permissions {
fn mode(&self) -> u32 {
self.mode as u32
}
fn set_mode(&mut self, mode: u32) {
self.mode = mode as u16;
}
fn from_mode(mode: u32) -> Self {
Permissions {
mode: mode as u16
}
}
}
pub struct DirEntry {
path: PathBuf,
}
impl DirEntry {
pub fn file_name(&self) -> &Path {
unsafe { mem::transmute(self.path.file_name().unwrap().to_str().unwrap()) }
}
pub fn file_type(&self) -> Result<FileType> {
self.metadata().map(|metadata| metadata.file_type())
}
pub fn metadata(&self) -> Result<Metadata> {
metadata(&self.path)
}
pub fn path(&self) -> PathBuf {
self.path.clone()
}
}
pub struct ReadDir {
path: PathBuf,
file: BufReader<File>,
}
impl Iterator for ReadDir {
type Item = Result<DirEntry>;
fn next(&mut self) -> Option<Result<DirEntry>> {
let mut name = String::new();
match self.file.read_line(&mut name) {
Ok(0) => None,
Ok(_) => {
if name.ends_with('\n') {
name.pop();
}
let mut path = self.path.clone();
path.push(name);
Some(Ok(DirEntry {
path: path
}))
},
Err(err) => Some(Err(err))
}
}
}
/// Find the canonical path of a file
pub fn canonicalize<P: AsRef<Path>>(path: P) -> Result<PathBuf> {
match File::open(path) {
Ok(file) => {
match file.path() {
Ok(realpath) => Ok(realpath),
Err(err) => Err(err)
}
},
Err(err) => Err(err)
}
}
/// Get information about a file
pub fn metadata<P: AsRef<Path>>(path: P) -> Result<Metadata> {
try!(File::open(path)).metadata()
}
/// Get information about a file without following symlinks
/// Warning: Redox does not currently support symlinks
pub fn symlink_metadata<P: AsRef<Path>>(path: P) -> Result<Metadata> {
metadata(path)
}
/// Create a new directory, using a path
/// The default mode of the directory is 775
pub fn create_dir<P: AsRef<Path>>(path: P) -> Result<()> {
let path_str = path.as_ref().as_os_str().as_inner();
mkdir(path_str, 0o775).and(Ok(())).map_err(|x| Error::from_sys(x))
}
/// Recursively create a directory and all of its parent components if they are missing.
pub fn create_dir_all<P: AsRef<Path>>(path: P) -> Result<()> {
if let Some(parent) = path.as_ref().parent() {
try!(create_dir_all(&parent));
}
if let Err(_err) = metadata(&path) {
try!(create_dir(&path));
}
Ok(())
}
/// Copy the contents of one file to another
pub fn copy<P: AsRef<Path>, Q: AsRef<Path>>(from: P, to: Q) -> Result<u64> {
let mut infile = try!(File::open(from));
let mut outfile = try!(File::create(to));
io::copy(&mut infile, &mut outfile)
}
/// Rename a file or directory to a new name
pub fn rename<P: AsRef<Path>, Q: AsRef<Path>>(from: P, to: Q) -> Result<()> {
try!(copy(Path::new(from.as_ref()), to));
remove_file(from)
}
/// Return an iterator over the entries within a directory
pub fn read_dir<P: AsRef<Path>>(path: P) -> Result<ReadDir> {
let path_buf = path.as_ref().to_owned();
File::open(&path_buf).map(|file| ReadDir { path: path_buf, file: BufReader::new(file) })
}
/// Removes an existing, empty directory
pub fn remove_dir<P: AsRef<Path>>(path: P) -> Result<()> {
let path_str = path.as_ref().as_os_str().as_inner();
rmdir(path_str).and(Ok(())).map_err(|x| Error::from_sys(x))
}
/// Removes a directory at this path, after removing all its contents. Use carefully!
pub fn remove_dir_all<P: AsRef<Path>>(path: P) -> Result<()> {
for child in try!(read_dir(&path)) {
let child = try!(child);
if try!(child.file_type()).is_dir() {
try!(remove_dir_all(&child.path()));
} else | {
try!(remove_file(&child.path()));
} | conditional_block |
|
fs.rs | from_raw_fd(fd) }).map_err(|x| Error::from_sys(x))
}
/// Duplicate the file
pub fn dup(&self, buf: &[u8]) -> Result<File> {
dup(self.fd, buf).map(|fd| unsafe { File::from_raw_fd(fd) }).map_err(|x| Error::from_sys(x))
}
/// Get information about a file
pub fn metadata(&self) -> Result<Metadata> {
let mut stat = Stat::default();
try!(fstat(self.fd, &mut stat).map_err(|x| Error::from_sys(x)));
Ok(Metadata {
stat: stat
})
}
/// Get the canonical path of the file
pub fn path(&self) -> Result<PathBuf> {
let mut buf: [u8; 4096] = [0; 4096];
match fpath(self.fd, &mut buf) {
Ok(count) => Ok(PathBuf::from(unsafe { String::from_utf8_unchecked(Vec::from(&buf[0..count])) })),
Err(err) => Err(Error::from_sys(err)),
}
}
/// Flush the file data and metadata
pub fn sync_all(&mut self) -> Result<()> {
fsync(self.fd).and(Ok(())).map_err(|x| Error::from_sys(x))
}
/// Flush the file data
pub fn sync_data(&mut self) -> Result<()> {
fsync(self.fd).and(Ok(())).map_err(|x| Error::from_sys(x))
}
/// Truncates the file
pub fn set_len(&self, size: u64) -> Result<()> {
ftruncate(self.fd, size as usize).and(Ok(())).map_err(|x| Error::from_sys(x))
}
}
impl AsRawFd for File {
fn as_raw_fd(&self) -> RawFd {
self.fd
}
}
impl FromRawFd for File {
unsafe fn from_raw_fd(fd: RawFd) -> Self {
File {
fd: fd
}
}
}
impl IntoRawFd for File {
fn into_raw_fd(self) -> RawFd {
let fd = self.fd;
mem::forget(self);
fd
}
}
impl Read for File {
fn read(&mut self, buf: &mut [u8]) -> Result<usize> {
read(self.fd, buf).map_err(|x| Error::from_sys(x))
}
}
impl Write for File {
fn write(&mut self, buf: &[u8]) -> Result<usize> {
write(self.fd, buf).map_err(|x| Error::from_sys(x))
}
fn flush(&mut self) -> Result<()> {
fsync(self.fd).and(Ok(())).map_err(|x| Error::from_sys(x))
}
}
impl Seek for File {
/// Seek a given position
fn seek(&mut self, pos: SeekFrom) -> Result<u64> {
let (whence, offset) = match pos {
SeekFrom::Start(offset) => (SEEK_SET, offset as isize),
SeekFrom::Current(offset) => (SEEK_CUR, offset as isize),
SeekFrom::End(offset) => (SEEK_END, offset as isize),
};
lseek(self.fd, offset, whence).map(|position| position as u64).map_err(|x| Error::from_sys(x))
}
}
impl Drop for File {
fn drop(&mut self) {
let _ = close(self.fd);
}
}
#[derive(Copy, Clone, Eq, PartialEq)]
pub struct FileType {
dir: bool,
file: bool,
}
impl FileType {
pub fn is_dir(&self) -> bool {
self.dir
}
pub fn is_file(&self) -> bool {
self.file
}
pub fn is_symlink(&self) -> bool {
false
}
}
impl ::os::unix::fs::FileTypeExt for FileType {
fn is_block_device(&self) -> bool { false }
fn is_char_device(&self) -> bool { false }
fn is_fifo(&self) -> bool { false }
fn is_socket(&self) -> bool { false }
}
pub struct OpenOptions {
read: bool,
write: bool,
append: bool,
create: bool,
truncate: bool,
mode: u16,
}
impl OpenOptions {
pub fn new() -> OpenOptions {
OpenOptions {
read: false,
write: false,
append: false,
create: false,
truncate: false,
mode: 0,
}
}
pub fn read(&mut self, read: bool) -> &mut OpenOptions {
self.read = read;
self
}
pub fn write(&mut self, write: bool) -> &mut OpenOptions {
self.write = write;
self
}
pub fn append(&mut self, append: bool) -> &mut OpenOptions {
self.append = append;
self
}
pub fn create(&mut self, create: bool) -> &mut OpenOptions {
self.create = create;
self
}
pub fn truncate(&mut self, truncate: bool) -> &mut OpenOptions {
self.truncate = truncate;
self
}
pub fn open<P: AsRef<Path>>(&self, path: P) -> Result<File> {
let mut flags = 0;
if self.read && self.write {
flags |= O_RDWR;
} else if self.read {
flags |= O_RDONLY;
} else if self.write {
flags |= O_WRONLY;
}
if self.append {
flags |= O_APPEND;
}
if self.create {
flags |= O_CREAT;
}
if self.truncate {
flags |= O_TRUNC;
}
flags |= (self.mode & MODE_PERM) as usize;
let path_str = path.as_ref().as_os_str().as_inner();
open(path_str, flags).map(|fd| unsafe { File::from_raw_fd(fd) }).map_err(|x| Error::from_sys(x))
}
}
impl ::os::unix::fs::OpenOptionsExt for OpenOptions {
fn mode(&mut self, mode: u32) -> &mut Self {
self.mode = mode as u16;
self
}
}
pub struct Metadata {
stat: Stat
}
impl Metadata {
pub fn file_type(&self) -> FileType {
FileType {
dir: self.stat.st_mode & MODE_DIR == MODE_DIR,
file: self.stat.st_mode & MODE_FILE == MODE_FILE
}
}
pub fn is_dir(&self) -> bool {
self.stat.st_mode & MODE_DIR == MODE_DIR
}
pub fn is_file(&self) -> bool {
self.stat.st_mode & MODE_FILE == MODE_FILE
}
pub fn len(&self) -> u64 {
self.stat.st_size
}
pub fn permissions(&self) -> Permissions {
Permissions {
mode: self.stat.st_mode & MODE_PERM
}
}
}
impl ::os::unix::fs::MetadataExt for Metadata {
fn mode(&self) -> u32 {
self.stat.st_mode as u32
}
fn uid(&self) -> u32 {
self.stat.st_uid
}
fn gid(&self) -> u32 {
self.stat.st_gid
}
fn size(&self) -> u64 {
self.stat.st_size
}
}
pub struct Permissions {
mode: u16
}
impl Permissions {
pub fn readonly(&self) -> bool {
self.mode & 0o222 == 0
}
pub fn set_readonly(&mut self, readonly: bool) {
if readonly {
self.mode &= !0o222;
} else {
self.mode |= 0o222;
}
}
}
impl ::os::unix::fs::PermissionsExt for Permissions {
fn mode(&self) -> u32 {
self.mode as u32
}
fn set_mode(&mut self, mode: u32) { | self.mode = mode as u16;
}
fn from_mode(mode: u32) -> Self {
Permissions {
mode: mode as u16
}
}
}
pub struct DirEntry {
path: PathBuf,
}
impl DirEntry {
pub fn file_name(&self) -> &Path {
unsafe { mem::transmute(self.path.file_name().unwrap().to_str().unwrap()) }
}
pub fn file_type(&self) -> Result<FileType> {
self.metadata().map(|metadata| metadata.file_type())
}
pub fn metadata(&self) -> Result<Metadata> {
metadata(&self.path)
}
pub fn path(&self) -> PathBuf {
self.path.clone()
}
}
pub struct ReadDir {
path: PathBuf,
file: BufReader<File>,
}
impl Iterator for ReadDir {
type Item = Result<DirEntry>;
fn next(&mut self) -> Option<Result<DirEntry>> {
let mut name = String::new();
match self.file.read_line(&mut name) {
Ok(0) => None,
Ok(_) => {
if name.ends_with('\n') {
name.pop();
}
let mut path = self.path.clone();
path.push(name);
Some(Ok(D | random_line_split |
|
fs.rs | from_raw_fd(fd) }).map_err(|x| Error::from_sys(x))
}
/// Duplicate the file
pub fn dup(&self, buf: &[u8]) -> Result<File> {
dup(self.fd, buf).map(|fd| unsafe { File::from_raw_fd(fd) }).map_err(|x| Error::from_sys(x))
}
/// Get information about a file
pub fn metadata(&self) -> Result<Metadata> {
let mut stat = Stat::default();
try!(fstat(self.fd, &mut stat).map_err(|x| Error::from_sys(x)));
Ok(Metadata {
stat: stat
})
}
/// Get the canonical path of the file
pub fn path(&self) -> Result<PathBuf> {
let mut buf: [u8; 4096] = [0; 4096];
match fpath(self.fd, &mut buf) {
Ok(count) => Ok(PathBuf::from(unsafe { String::from_utf8_unchecked(Vec::from(&buf[0..count])) })),
Err(err) => Err(Error::from_sys(err)),
}
}
/// Flush the file data and metadata
pub fn sync_all(&mut self) -> Result<()> {
fsync(self.fd).and(Ok(())).map_err(|x| Error::from_sys(x))
}
/// Flush the file data
pub fn sync_data(&mut self) -> Result<()> {
fsync(self.fd).and(Ok(())).map_err(|x| Error::from_sys(x))
}
/// Truncates the file
pub fn set_len(&self, size: u64) -> Result<()> {
ftruncate(self.fd, size as usize).and(Ok(())).map_err(|x| Error::from_sys(x))
}
}
impl AsRawFd for File {
fn as_raw_fd(&self) -> RawFd |
}
impl FromRawFd for File {
unsafe fn from_raw_fd(fd: RawFd) -> Self {
File {
fd: fd
}
}
}
impl IntoRawFd for File {
fn into_raw_fd(self) -> RawFd {
let fd = self.fd;
mem::forget(self);
fd
}
}
impl Read for File {
fn read(&mut self, buf: &mut [u8]) -> Result<usize> {
read(self.fd, buf).map_err(|x| Error::from_sys(x))
}
}
impl Write for File {
fn write(&mut self, buf: &[u8]) -> Result<usize> {
write(self.fd, buf).map_err(|x| Error::from_sys(x))
}
fn flush(&mut self) -> Result<()> {
fsync(self.fd).and(Ok(())).map_err(|x| Error::from_sys(x))
}
}
impl Seek for File {
/// Seek a given position
fn seek(&mut self, pos: SeekFrom) -> Result<u64> {
let (whence, offset) = match pos {
SeekFrom::Start(offset) => (SEEK_SET, offset as isize),
SeekFrom::Current(offset) => (SEEK_CUR, offset as isize),
SeekFrom::End(offset) => (SEEK_END, offset as isize),
};
lseek(self.fd, offset, whence).map(|position| position as u64).map_err(|x| Error::from_sys(x))
}
}
impl Drop for File {
fn drop(&mut self) {
let _ = close(self.fd);
}
}
#[derive(Copy, Clone, Eq, PartialEq)]
pub struct FileType {
dir: bool,
file: bool,
}
impl FileType {
pub fn is_dir(&self) -> bool {
self.dir
}
pub fn is_file(&self) -> bool {
self.file
}
pub fn is_symlink(&self) -> bool {
false
}
}
impl ::os::unix::fs::FileTypeExt for FileType {
fn is_block_device(&self) -> bool { false }
fn is_char_device(&self) -> bool { false }
fn is_fifo(&self) -> bool { false }
fn is_socket(&self) -> bool { false }
}
pub struct OpenOptions {
read: bool,
write: bool,
append: bool,
create: bool,
truncate: bool,
mode: u16,
}
impl OpenOptions {
pub fn new() -> OpenOptions {
OpenOptions {
read: false,
write: false,
append: false,
create: false,
truncate: false,
mode: 0,
}
}
pub fn read(&mut self, read: bool) -> &mut OpenOptions {
self.read = read;
self
}
pub fn write(&mut self, write: bool) -> &mut OpenOptions {
self.write = write;
self
}
pub fn append(&mut self, append: bool) -> &mut OpenOptions {
self.append = append;
self
}
pub fn create(&mut self, create: bool) -> &mut OpenOptions {
self.create = create;
self
}
pub fn truncate(&mut self, truncate: bool) -> &mut OpenOptions {
self.truncate = truncate;
self
}
pub fn open<P: AsRef<Path>>(&self, path: P) -> Result<File> {
let mut flags = 0;
if self.read && self.write {
flags |= O_RDWR;
} else if self.read {
flags |= O_RDONLY;
} else if self.write {
flags |= O_WRONLY;
}
if self.append {
flags |= O_APPEND;
}
if self.create {
flags |= O_CREAT;
}
if self.truncate {
flags |= O_TRUNC;
}
flags |= (self.mode & MODE_PERM) as usize;
let path_str = path.as_ref().as_os_str().as_inner();
open(path_str, flags).map(|fd| unsafe { File::from_raw_fd(fd) }).map_err(|x| Error::from_sys(x))
}
}
impl ::os::unix::fs::OpenOptionsExt for OpenOptions {
fn mode(&mut self, mode: u32) -> &mut Self {
self.mode = mode as u16;
self
}
}
pub struct Metadata {
stat: Stat
}
impl Metadata {
pub fn file_type(&self) -> FileType {
FileType {
dir: self.stat.st_mode & MODE_DIR == MODE_DIR,
file: self.stat.st_mode & MODE_FILE == MODE_FILE
}
}
pub fn is_dir(&self) -> bool {
self.stat.st_mode & MODE_DIR == MODE_DIR
}
pub fn is_file(&self) -> bool {
self.stat.st_mode & MODE_FILE == MODE_FILE
}
pub fn len(&self) -> u64 {
self.stat.st_size
}
pub fn permissions(&self) -> Permissions {
Permissions {
mode: self.stat.st_mode & MODE_PERM
}
}
}
impl ::os::unix::fs::MetadataExt for Metadata {
fn mode(&self) -> u32 {
self.stat.st_mode as u32
}
fn uid(&self) -> u32 {
self.stat.st_uid
}
fn gid(&self) -> u32 {
self.stat.st_gid
}
fn size(&self) -> u64 {
self.stat.st_size
}
}
pub struct Permissions {
mode: u16
}
impl Permissions {
pub fn readonly(&self) -> bool {
self.mode & 0o222 == 0
}
pub fn set_readonly(&mut self, readonly: bool) {
if readonly {
self.mode &= !0o222;
} else {
self.mode |= 0o222;
}
}
}
impl ::os::unix::fs::PermissionsExt for Permissions {
fn mode(&self) -> u32 {
self.mode as u32
}
fn set_mode(&mut self, mode: u32) {
self.mode = mode as u16;
}
fn from_mode(mode: u32) -> Self {
Permissions {
mode: mode as u16
}
}
}
pub struct DirEntry {
path: PathBuf,
}
impl DirEntry {
pub fn file_name(&self) -> &Path {
unsafe { mem::transmute(self.path.file_name().unwrap().to_str().unwrap()) }
}
pub fn file_type(&self) -> Result<FileType> {
self.metadata().map(|metadata| metadata.file_type())
}
pub fn metadata(&self) -> Result<Metadata> {
metadata(&self.path)
}
pub fn path(&self) -> PathBuf {
self.path.clone()
}
}
pub struct ReadDir {
path: PathBuf,
file: BufReader<File>,
}
impl Iterator for ReadDir {
type Item = Result<DirEntry>;
fn next(&mut self) -> Option<Result<DirEntry>> {
let mut name = String::new();
match self.file.read_line(&mut name) {
Ok(0) => None,
Ok(_) => {
if name.ends_with('\n') {
name.pop();
}
let mut path = self.path.clone();
path.push(name);
Some( | {
self.fd
} | identifier_body |
fs.rs | pub fn sync_data(&mut self) -> Result<()> {
fsync(self.fd).and(Ok(())).map_err(|x| Error::from_sys(x))
}
/// Truncates the file
pub fn set_len(&self, size: u64) -> Result<()> {
ftruncate(self.fd, size as usize).and(Ok(())).map_err(|x| Error::from_sys(x))
}
}
impl AsRawFd for File {
fn as_raw_fd(&self) -> RawFd {
self.fd
}
}
impl FromRawFd for File {
unsafe fn from_raw_fd(fd: RawFd) -> Self {
File {
fd: fd
}
}
}
impl IntoRawFd for File {
fn into_raw_fd(self) -> RawFd {
let fd = self.fd;
mem::forget(self);
fd
}
}
impl Read for File {
fn read(&mut self, buf: &mut [u8]) -> Result<usize> {
read(self.fd, buf).map_err(|x| Error::from_sys(x))
}
}
impl Write for File {
fn write(&mut self, buf: &[u8]) -> Result<usize> {
write(self.fd, buf).map_err(|x| Error::from_sys(x))
}
fn flush(&mut self) -> Result<()> {
fsync(self.fd).and(Ok(())).map_err(|x| Error::from_sys(x))
}
}
impl Seek for File {
/// Seek a given position
fn seek(&mut self, pos: SeekFrom) -> Result<u64> {
let (whence, offset) = match pos {
SeekFrom::Start(offset) => (SEEK_SET, offset as isize),
SeekFrom::Current(offset) => (SEEK_CUR, offset as isize),
SeekFrom::End(offset) => (SEEK_END, offset as isize),
};
lseek(self.fd, offset, whence).map(|position| position as u64).map_err(|x| Error::from_sys(x))
}
}
impl Drop for File {
fn drop(&mut self) {
let _ = close(self.fd);
}
}
#[derive(Copy, Clone, Eq, PartialEq)]
pub struct FileType {
dir: bool,
file: bool,
}
impl FileType {
pub fn is_dir(&self) -> bool {
self.dir
}
pub fn is_file(&self) -> bool {
self.file
}
pub fn is_symlink(&self) -> bool {
false
}
}
impl ::os::unix::fs::FileTypeExt for FileType {
fn is_block_device(&self) -> bool { false }
fn is_char_device(&self) -> bool { false }
fn is_fifo(&self) -> bool { false }
fn is_socket(&self) -> bool { false }
}
pub struct OpenOptions {
read: bool,
write: bool,
append: bool,
create: bool,
truncate: bool,
mode: u16,
}
impl OpenOptions {
pub fn new() -> OpenOptions {
OpenOptions {
read: false,
write: false,
append: false,
create: false,
truncate: false,
mode: 0,
}
}
pub fn read(&mut self, read: bool) -> &mut OpenOptions {
self.read = read;
self
}
pub fn write(&mut self, write: bool) -> &mut OpenOptions {
self.write = write;
self
}
pub fn append(&mut self, append: bool) -> &mut OpenOptions {
self.append = append;
self
}
pub fn create(&mut self, create: bool) -> &mut OpenOptions {
self.create = create;
self
}
pub fn truncate(&mut self, truncate: bool) -> &mut OpenOptions {
self.truncate = truncate;
self
}
pub fn open<P: AsRef<Path>>(&self, path: P) -> Result<File> {
let mut flags = 0;
if self.read && self.write {
flags |= O_RDWR;
} else if self.read {
flags |= O_RDONLY;
} else if self.write {
flags |= O_WRONLY;
}
if self.append {
flags |= O_APPEND;
}
if self.create {
flags |= O_CREAT;
}
if self.truncate {
flags |= O_TRUNC;
}
flags |= (self.mode & MODE_PERM) as usize;
let path_str = path.as_ref().as_os_str().as_inner();
open(path_str, flags).map(|fd| unsafe { File::from_raw_fd(fd) }).map_err(|x| Error::from_sys(x))
}
}
impl ::os::unix::fs::OpenOptionsExt for OpenOptions {
fn mode(&mut self, mode: u32) -> &mut Self {
self.mode = mode as u16;
self
}
}
pub struct Metadata {
stat: Stat
}
impl Metadata {
pub fn file_type(&self) -> FileType {
FileType {
dir: self.stat.st_mode & MODE_DIR == MODE_DIR,
file: self.stat.st_mode & MODE_FILE == MODE_FILE
}
}
pub fn is_dir(&self) -> bool {
self.stat.st_mode & MODE_DIR == MODE_DIR
}
pub fn is_file(&self) -> bool {
self.stat.st_mode & MODE_FILE == MODE_FILE
}
pub fn len(&self) -> u64 {
self.stat.st_size
}
pub fn permissions(&self) -> Permissions {
Permissions {
mode: self.stat.st_mode & MODE_PERM
}
}
}
impl ::os::unix::fs::MetadataExt for Metadata {
fn mode(&self) -> u32 {
self.stat.st_mode as u32
}
fn uid(&self) -> u32 {
self.stat.st_uid
}
fn gid(&self) -> u32 {
self.stat.st_gid
}
fn size(&self) -> u64 {
self.stat.st_size
}
}
pub struct Permissions {
mode: u16
}
impl Permissions {
pub fn readonly(&self) -> bool {
self.mode & 0o222 == 0
}
pub fn set_readonly(&mut self, readonly: bool) {
if readonly {
self.mode &= !0o222;
} else {
self.mode |= 0o222;
}
}
}
impl ::os::unix::fs::PermissionsExt for Permissions {
fn mode(&self) -> u32 {
self.mode as u32
}
fn set_mode(&mut self, mode: u32) {
self.mode = mode as u16;
}
fn from_mode(mode: u32) -> Self {
Permissions {
mode: mode as u16
}
}
}
pub struct DirEntry {
path: PathBuf,
}
impl DirEntry {
pub fn file_name(&self) -> &Path {
unsafe { mem::transmute(self.path.file_name().unwrap().to_str().unwrap()) }
}
pub fn file_type(&self) -> Result<FileType> {
self.metadata().map(|metadata| metadata.file_type())
}
pub fn metadata(&self) -> Result<Metadata> {
metadata(&self.path)
}
pub fn path(&self) -> PathBuf {
self.path.clone()
}
}
pub struct ReadDir {
path: PathBuf,
file: BufReader<File>,
}
impl Iterator for ReadDir {
type Item = Result<DirEntry>;
fn next(&mut self) -> Option<Result<DirEntry>> {
let mut name = String::new();
match self.file.read_line(&mut name) {
Ok(0) => None,
Ok(_) => {
if name.ends_with('\n') {
name.pop();
}
let mut path = self.path.clone();
path.push(name);
Some(Ok(DirEntry {
path: path
}))
},
Err(err) => Some(Err(err))
}
}
}
/// Find the canonical path of a file
pub fn canonicalize<P: AsRef<Path>>(path: P) -> Result<PathBuf> {
match File::open(path) {
Ok(file) => {
match file.path() {
Ok(realpath) => Ok(realpath),
Err(err) => Err(err)
}
},
Err(err) => Err(err)
}
}
/// Get information about a file
pub fn metadata<P: AsRef<Path>>(path: P) -> Result<Metadata> {
try!(File::open(path)).metadata()
}
/// Get information about a file without following symlinks
/// Warning: Redox does not currently support symlinks
pub fn symlink_metadata<P: AsRef<Path>>(path: P) -> Result<Metadata> {
metadata(path)
}
/// Create a new directory, using a path
/// The default mode of the directory is 775
pub fn create_dir<P: AsRef<Path>>(path: P) -> Result<()> {
let path_str = path.as_ref().as_os_str().as_inner();
mkdir(path_str, 0o775).and(Ok(())).map_err(|x| Error::from_sys(x))
}
/// Recursively create a directory and all of its parent components if they are missing.
pub fn | create_dir_all | identifier_name |
|
dnp3.go | File
case 27: // Delete File
case 28: // Get File Info
case 29: // Authenticate File
case 30: // Abort File
case 31: // Activate Config
case 32: // Authentication Request
case 33: // Authentication Error
case 129: // Response
case 130: // Unsolicited Response
case 131: // Authentication Response
}
objTypeField := binary.BigEndian.Uint16([]byte{data[objectStart], data[objectStart+1]})
objectGroup := objTypeField & 0xFF00
objectVariation := objTypeField & 0x00FF
object := d.hexConvert([]byte{data[objectStart], data[objectStart+1]})
objectPrefixCode := data[objectStart+2] & 0x70 // OPC
objectRangeSpecifierCode := data[objectStart+2] & 0x0F // RSC
fmt.Println(object)
fmt.Println(objectGroup)
fmt.Println(objectVariation)
fmt.Printf("Prefix Code %d\n", objectPrefixCode)
fmt.Println(ObjPrefixCodes[objectPrefixCode])
fmt.Printf("Range Specifier Code %d\n", objectRangeSpecifierCode) // 6 means no range field
fmt.Println(ObjRangeSpecifierCodes[objectRangeSpecifierCode])
fmt.Println(d.hexConvert([]byte{data[objectStart+3]}))
offset := objectStart + 3
rangebytes := 0
fmt.Println(offset)
switch objectRangeSpecifierCode {
case 0:
// start := offset
numItems := int(data[offset+1]) - int(data[offset]) + 1
rangebytes = 2
fmt.Println(numItems)
pointAddress := int(data[offset])
fmt.Println(pointAddress)
// num_items = ( tvb_get_guint8(tvb, offset+1) - tvb_get_guint8(tvb, offset) + 1);
// proto_item_set_generated(range_item);
// al_ptaddr = tvb_get_guint8(tvb, offset);
// proto_tree_add_item(range_tree, hf_dnp3_al_range_start8, tvb, offset, 1, ENC_LITTLE_ENDIAN);
// proto_tree_add_item(range_tree, hf_dnp3_al_range_stop8, tvb, offset + 1, 1, ENC_LITTLE_ENDIAN);
// rangebytes = 2;
case 1:
case 2:
case 3:
case 4:
case 5:
case 6:
case 7:
case 8:
case 9:
case 10:
case 11:
case 12:
case 13:
case 14:
case 15:
}
/* Move offset past any range field */
offset += rangebytes
fmt.Println(offset)
// RSCArrayFirst := []byte{0, 1, 2, 3, 4, 5}
// if d.contains(RSCArrayFirst, objectRangeSpecifierCode) {
// }
/* Special handling for Octet string objects as the variation is the length of the string */
// temp = objTypeField & 0xFF00
// if (temp == AL_OBJ_OCT) || (temp == AL_OBJ_OCT_EVT) {
// al_oct_len = al_obj & 0xFF
// al_obj = temp
// }
// objectGroup := data[objectStart] & 0x0f
// objectGroup := fmt.Sprintf("0x%x%x", data[objectStart], data[objectStart+1])
// fmt.Println(objectGroup)
// objectGroup, _ := strconv.Atoi(fmt.Sprintf("%d", data[objectStart]))
// objectVariation, _ := strconv.Atoi(fmt.Sprintf("%d", data[objectStart+1]))
// fmt.Println(objectGroup)
// fmt.Println(objectVariation)
/* Index Size (3-bits x111xxxx) */
// /* When Qualifier Code != 11 */
// #define AL_OBJQL_PREFIX_NI 0x00 /* Objects are Packed with no index */
// #define AL_OBJQL_PREFIX_1O 0x01 /* Objects are prefixed w/ 1-octet index */
// #define AL_OBJQL_PREFIX_2O 0x02 /* Objects are prefixed w/ 2-octet index */
// #define AL_OBJQL_PREFIX_4O 0x03 /* Objects are prefixed w/ 4-octet index */
// #define AL_OBJQL_PREFIX_1OS 0x04 /* Objects are prefixed w/ 1-octet object size */
// #define AL_OBJQL_PREFIX_2OS 0x05 /* Objects are prefixed w/ 2-octet object size */
// #define AL_OBJQL_PREFIX_4OS 0x06 /* Objects are prefixed w/ 4-octet object size */
// /* When Qualifier Code == 11 */
// #define AL_OBJQL_IDX11_1OIS 0x01 /* 1 octet identifier size */
// #define AL_OBJQL_IDX11_2OIS 0x02 /* 2 octet identifier size */
// #define AL_OBJQL_IDX11_4OIS 0x03 /* 4 octet identifier size */
// /* Qualifier Code (4-bits) */
// /* 4-bits ( xxxx1111 ) */
// #define AL_OBJQL_RANGE_SSI8 0x00 /* 00 8-bit Start and Stop Indices in Range Field */
// #define AL_OBJQL_RANGE_SSI16 0x01 /* 01 16-bit Start and Stop Indices in Range Field */
// #define AL_OBJQL_RANGE_SSI32 0x02 /* 02 32-bit Start and Stop Indices in Range Field */
// #define AL_OBJQL_RANGE_AA8 0x03 /* 03 8-bit Absolute Address in Range Field */
// #define AL_OBJQL_RANGE_AA16 0x04 /* 04 16-bit Absolute Address in Range Field */
// #define AL_OBJQL_RANGE_AA32 0x05 /* 05 32-bit Absolute Address in Range Field */
// #define AL_OBJQL_RANGE_R0 0x06 /* 06 Length of Range field is 0 (no range field) */
// #define AL_OBJQL_RANGE_SF8 0x07 /* 07 8-bit Single Field Quantity */
// #define AL_OBJQL_RANGE_SF16 0x08 /* 08 16-bit Single Field Quantity */
// #define AL_OBJQL_RANGE_SF32 0x09 /* 09 32-bit Single Field Quantity */
// /* 0x0A 10 Reserved */
// #define AL_OBJQL_RANGE_FF 0x0B /* 11 Free-format Qualifier, range field has 1 octet count of objects */
// /* 0x0C 12 Reserved */
// /* 0x0D 13 Reserved */
// /* 0x0E 14 Reserved */
// /* 0x0F 15 Reserved */
/***************************************************************************/
/* Application Layer Data Object Qualifier */
/***************************************************************************/
// /* Bit-Masks */
// #define AL_OBJQ_PREFIX 0x70 /* x111xxxx Masks Prefix from Qualifier */
// #define AL_OBJQ_RANGE 0x0F /* xxxx1111 Masks Range from Qualifier */
// objectQualifier := fmt.Sprintf("0x%d", data[objectStart+2])
// fmt.Println(objectQualifier)
// src = []byte{data[objectStart], data[objectStart+1]}
// dst = make([]byte, hex.EncodedLen(len(src)))
// hex.Encode(dst, src)
// prefixCode := fmt.Sprintf("0x%s", dst)
// fmt.Println(prefixCode)
}
func (d *DNP3) IsDNP3(bytesRead []byte) bool {
if len(bytesRead) >= MIN_HEADER_LENGTH && binary.BigEndian.Uint16(bytesRead[0:2]) == START_FIELD {
return true
}
return false
}
func (d *DNP3) isMaster(bytesRead []byte) bool {
intValue := int((bytesRead[3] & 0x80) >> 7)
var boolValue bool = intValue != 0
return boolValue
}
func (d *DNP3) hexConvert(byteArray []byte) string {
return "0x" + hex.EncodeToString(byteArray)
}
func (d *DNP3) isMultiPart(bytesRead []byte) bool {
var FirstOfMulti01 byte = 0x40
var NotFirstNotLast00 byte = 0x00 | var FinalFrame10 byte = 0x80
var OneFrame11 byte = 0xC0
TpFinFir := bytesRead[10] & 0xC0
switch TpFinFir { | random_line_split |
|
dnp3.go | 3: "Objects prefixed with 4-octet index",
4: "Objects prefixed with 1-octet object size",
5: "Objects prefixed with 2-octet object size",
6: "Objects prefixed with 4-octet object size",
7: "Reserved",
}
/***************************************************************************/
/* Application Layer Object Prefix codes bits */
/***************************************************************************/
var ObjRangeSpecifierCodes = map[byte]string{
0: "8-bit Start and Stop Indices in Range Field",
1: "16-bit Start and Stop Indices in Range Field",
2: "32-bit Start and Stop Indices in Range Field",
3: "8-bit Absolute Address in Range Field",
4: "16-bit Absolute Address in Range Field",
5: "32-bit Absolute Address in Range Field",
6: "Length of Range field is 0 (no range field)",
7: "8-bit Single Field Quantity",
8: "16-bit Single Field Quantity",
9: "32-bit Single Field Quantity",
10: "Reserved",
11: "Free-format Qualifier, range field has 1 octet count of objects",
12: "Reserved",
13: "Reserved",
14: "Reserved",
15: "Reserved",
}
var (
errDNP3PacketTooShort = errors.New("DNS packet too short")
)
type DNP3 struct {
BaseLayer // Stores the packet bytes and payload bytes.
DNP3DataLinkLayer DNP3DataLinkLayer
DNP3TransportLayer DNP3TransportLayer
DNP3ApplicationLayer DNP3ApplicationLayer
SomeByte byte
AnotherByte byte
restOfData []byte
}
type DNP3DataLinkLayer struct {
Start string
Length int
Control struct {
ControlByte string
IsMaster int `json:"Is Master"`
PRM int `json:"Primary"`
FCB int `json:"Frame Count Bit"`
FCV int `json:"Frame Count Valid"`
FUNC string `json:"Function Code"`
}
Destination int
Source int
CRC string
}
type DNP3TransportLayer struct {
TransportByte string
Final int
First int
Sequence int
}
type DNP3ApplicationLayer struct {
Control struct {
ControlByte string
First int
Final int
Confirm int
Unsolicited int
Sequence int
}
Function string `json:"Function Code"`
IINCode string `json:"Internal Indication (IIN)"`
}
type DNP3AppObject struct {
Group int
Variation int
Qualifier int
RangeStart int
RangeStop int
DataType int
Length int
}
func (d *DNP3) LayerType() gopacket.LayerType { return LayerTypeDNP3 }
func (d *DNP3) LayerContents() []byte {
return []byte{d.SomeByte, d.AnotherByte}
}
func (d *DNP3) LayerPayload() []byte {
return d.restOfData
}
func (d *DNP3) DecodeFromBytes(data []byte, df gopacket.DecodeFeedback) error {
// If the data block is too short to be a DNP3 layer, then return an error.
if len(data) < 10 {
df.SetTruncated()
return errDNP3PacketTooShort
}
d.linkLayer(data)
d.transportLayer(data)
d.applicationLayer(data)
return nil
}
func decodeDNP3(data []byte, p gopacket.PacketBuilder) error {
// Attempt to decode the byte slice.
d := &DNP3{}
err := d.DecodeFromBytes(data, p)
if err != nil {
return err
}
// If the decoding worked, add the layer to the packet and set it
// as the application layer too, if there isn't already one.
p.AddLayer(d)
p.SetApplicationLayer(d)
d.BaseLayer = BaseLayer{Contents: data[:len(data)]}
d.BaseLayer.Payload = nil
return p.NextDecoder(gopacket.LayerTypePayload)
}
// CanDecode implements gopacket.DecodingLayer.
func (d *DNP3) CanDecode() gopacket.LayerClass {
return LayerTypeDNP3
}
// NextLayerType returns the layer type contained by this DecodingLayer.
func (d *DNP3) | () gopacket.LayerType {
return gopacket.LayerTypePayload
}
// Payload returns nil, since TLS encrypted payload is inside TLSAppDataRecord
func (d *DNP3) Payload() []byte {
return nil
}
func appObject(bytesRead []byte) {
object := bytesRead[22:]
// indexSize := uint(object[2] & 0x70 >> 4)
// QualifierCode := uint(object[2] & 0x0F)
// fmt.Println(indexSize)
// fmt.Println(QualifierCode)
group := int(object[0])
variation := int(object[1])
qualifier := int(object[2])
rangeStart := int(object[3])
rangeStop := int(object[4])
dataType := int(object[5])
length := int(object[6])
appObject := DNP3AppObject{
Group: group,
Variation: variation,
Qualifier: qualifier,
RangeStart: rangeStart,
RangeStop: rangeStop,
DataType: dataType,
Length: length,
}
out, err := json.Marshal(appObject)
if err != nil {
panic(err)
}
fmt.Println(string(out))
}
func (d *DNP3) linkLayer(data []byte) {
start := d.hexConvert(data[0:2])
d.DNP3DataLinkLayer.Start = start
length := int(data[2])
d.DNP3DataLinkLayer.Length = length
ctlControl := d.hexConvert([]byte{data[3]})
d.DNP3DataLinkLayer.Control.ControlByte = ctlControl
IsMaster := int((data[3] & 0x80) >> 7)
d.DNP3DataLinkLayer.Control.IsMaster = IsMaster
PRM := int((data[3] & 0x40) >> 6)
d.DNP3DataLinkLayer.Control.PRM = PRM
FCB := int((data[3] & 0x20) >> 5)
d.DNP3DataLinkLayer.Control.FCB = FCB
FCV := int((data[3] & 0x10) >> 4)
d.DNP3DataLinkLayer.Control.FCV = FCV
FUNCCODE := data[3] & 0x0F
ctlFUNCCODE := fmt.Sprintf("%d", FUNCCODE)
var ctlFUNC string
if PRM == 0x00 {
ctlFUNC = SfCodes[FUNCCODE]
}
if PRM == 0x01 {
ctlFUNC = PfCodes[FUNCCODE]
}
ctlFUNC = ctlFUNC + " (" + ctlFUNCCODE + ")"
d.DNP3DataLinkLayer.Control.FUNC = ctlFUNC
// TODO: make sure 0 to 65535
destination := fmt.Sprintf("%x%x", data[5], data[4])
destinationInt, _ := strconv.Atoi(destination)
d.DNP3DataLinkLayer.Destination = destinationInt
// TODO: make sure 0 to 65535
source := fmt.Sprintf("%x%x", data[7], data[6])
sourceInt, _ := strconv.Atoi(source)
d.DNP3DataLinkLayer.Source = sourceInt
// TODO: Is correct? Hesapla
crc := fmt.Sprintf("0x%x%x", data[9], data[8])
d.DNP3DataLinkLayer.CRC = crc
}
func (d *DNP3) transportLayer(data []byte) {
transport := fmt.Sprintf("0x%x", data[10])
d.DNP3TransportLayer.TransportByte = transport
final := data[10] & 0x80 >> 7
d.DNP3TransportLayer.Final = int(final)
first := data[10] & 0x40 >> 6
d.DNP3TransportLayer.First = int(first)
sequence := data[10] & 0x3f // 6bit
d.DNP3TransportLayer.Sequence = int(sequence)
}
func (d *DNP3) applicationLayer(data []byte) {
// /***************************************************************************/
// /* Application Layer Bit-Masks */
// /***************************************************************************/
// #define DNP3_AL_UNS 0x10
// #define DNP3_AL_CON 0x20
// #define DNP3_AL_FIN 0x40
// #define DNP3_AL_FIR 0x80
// #define DNP3_AL_SEQ 0x0f
// #define DNP3_AL_FUNC 0xff
controlByte := fmt.Sprintf("0x%x", data[11])
d.DNP3ApplicationLayer.Control.ControlByte = controlByte
| NextLayerType | identifier_name |
dnp3.go | 3: "Objects prefixed with 4-octet index",
4: "Objects prefixed with 1-octet object size",
5: "Objects prefixed with 2-octet object size",
6: "Objects prefixed with 4-octet object size",
7: "Reserved",
}
/***************************************************************************/
/* Application Layer Object Prefix codes bits */
/***************************************************************************/
var ObjRangeSpecifierCodes = map[byte]string{
0: "8-bit Start and Stop Indices in Range Field",
1: "16-bit Start and Stop Indices in Range Field",
2: "32-bit Start and Stop Indices in Range Field",
3: "8-bit Absolute Address in Range Field",
4: "16-bit Absolute Address in Range Field",
5: "32-bit Absolute Address in Range Field",
6: "Length of Range field is 0 (no range field)",
7: "8-bit Single Field Quantity",
8: "16-bit Single Field Quantity",
9: "32-bit Single Field Quantity",
10: "Reserved",
11: "Free-format Qualifier, range field has 1 octet count of objects",
12: "Reserved",
13: "Reserved",
14: "Reserved",
15: "Reserved",
}
var (
errDNP3PacketTooShort = errors.New("DNS packet too short")
)
type DNP3 struct {
BaseLayer // Stores the packet bytes and payload bytes.
DNP3DataLinkLayer DNP3DataLinkLayer
DNP3TransportLayer DNP3TransportLayer
DNP3ApplicationLayer DNP3ApplicationLayer
SomeByte byte
AnotherByte byte
restOfData []byte
}
type DNP3DataLinkLayer struct {
Start string
Length int
Control struct {
ControlByte string
IsMaster int `json:"Is Master"`
PRM int `json:"Primary"`
FCB int `json:"Frame Count Bit"`
FCV int `json:"Frame Count Valid"`
FUNC string `json:"Function Code"`
}
Destination int
Source int
CRC string
}
type DNP3TransportLayer struct {
TransportByte string
Final int
First int
Sequence int
}
type DNP3ApplicationLayer struct {
Control struct {
ControlByte string
First int
Final int
Confirm int
Unsolicited int
Sequence int
}
Function string `json:"Function Code"`
IINCode string `json:"Internal Indication (IIN)"`
}
type DNP3AppObject struct {
Group int
Variation int
Qualifier int
RangeStart int
RangeStop int
DataType int
Length int
}
func (d *DNP3) LayerType() gopacket.LayerType { return LayerTypeDNP3 }
func (d *DNP3) LayerContents() []byte {
return []byte{d.SomeByte, d.AnotherByte}
}
func (d *DNP3) LayerPayload() []byte |
func (d *DNP3) DecodeFromBytes(data []byte, df gopacket.DecodeFeedback) error {
// If the data block is too short to be a DNP3 layer, then return an error.
if len(data) < 10 {
df.SetTruncated()
return errDNP3PacketTooShort
}
d.linkLayer(data)
d.transportLayer(data)
d.applicationLayer(data)
return nil
}
func decodeDNP3(data []byte, p gopacket.PacketBuilder) error {
// Attempt to decode the byte slice.
d := &DNP3{}
err := d.DecodeFromBytes(data, p)
if err != nil {
return err
}
// If the decoding worked, add the layer to the packet and set it
// as the application layer too, if there isn't already one.
p.AddLayer(d)
p.SetApplicationLayer(d)
d.BaseLayer = BaseLayer{Contents: data[:len(data)]}
d.BaseLayer.Payload = nil
return p.NextDecoder(gopacket.LayerTypePayload)
}
// CanDecode implements gopacket.DecodingLayer.
func (d *DNP3) CanDecode() gopacket.LayerClass {
return LayerTypeDNP3
}
// NextLayerType returns the layer type contained by this DecodingLayer.
func (d *DNP3) NextLayerType() gopacket.LayerType {
return gopacket.LayerTypePayload
}
// Payload returns nil, since TLS encrypted payload is inside TLSAppDataRecord
func (d *DNP3) Payload() []byte {
return nil
}
func appObject(bytesRead []byte) {
object := bytesRead[22:]
// indexSize := uint(object[2] & 0x70 >> 4)
// QualifierCode := uint(object[2] & 0x0F)
// fmt.Println(indexSize)
// fmt.Println(QualifierCode)
group := int(object[0])
variation := int(object[1])
qualifier := int(object[2])
rangeStart := int(object[3])
rangeStop := int(object[4])
dataType := int(object[5])
length := int(object[6])
appObject := DNP3AppObject{
Group: group,
Variation: variation,
Qualifier: qualifier,
RangeStart: rangeStart,
RangeStop: rangeStop,
DataType: dataType,
Length: length,
}
out, err := json.Marshal(appObject)
if err != nil {
panic(err)
}
fmt.Println(string(out))
}
func (d *DNP3) linkLayer(data []byte) {
start := d.hexConvert(data[0:2])
d.DNP3DataLinkLayer.Start = start
length := int(data[2])
d.DNP3DataLinkLayer.Length = length
ctlControl := d.hexConvert([]byte{data[3]})
d.DNP3DataLinkLayer.Control.ControlByte = ctlControl
IsMaster := int((data[3] & 0x80) >> 7)
d.DNP3DataLinkLayer.Control.IsMaster = IsMaster
PRM := int((data[3] & 0x40) >> 6)
d.DNP3DataLinkLayer.Control.PRM = PRM
FCB := int((data[3] & 0x20) >> 5)
d.DNP3DataLinkLayer.Control.FCB = FCB
FCV := int((data[3] & 0x10) >> 4)
d.DNP3DataLinkLayer.Control.FCV = FCV
FUNCCODE := data[3] & 0x0F
ctlFUNCCODE := fmt.Sprintf("%d", FUNCCODE)
var ctlFUNC string
if PRM == 0x00 {
ctlFUNC = SfCodes[FUNCCODE]
}
if PRM == 0x01 {
ctlFUNC = PfCodes[FUNCCODE]
}
ctlFUNC = ctlFUNC + " (" + ctlFUNCCODE + ")"
d.DNP3DataLinkLayer.Control.FUNC = ctlFUNC
// TODO: make sure 0 to 65535
destination := fmt.Sprintf("%x%x", data[5], data[4])
destinationInt, _ := strconv.Atoi(destination)
d.DNP3DataLinkLayer.Destination = destinationInt
// TODO: make sure 0 to 65535
source := fmt.Sprintf("%x%x", data[7], data[6])
sourceInt, _ := strconv.Atoi(source)
d.DNP3DataLinkLayer.Source = sourceInt
// TODO: Is correct? Hesapla
crc := fmt.Sprintf("0x%x%x", data[9], data[8])
d.DNP3DataLinkLayer.CRC = crc
}
func (d *DNP3) transportLayer(data []byte) {
transport := fmt.Sprintf("0x%x", data[10])
d.DNP3TransportLayer.TransportByte = transport
final := data[10] & 0x80 >> 7
d.DNP3TransportLayer.Final = int(final)
first := data[10] & 0x40 >> 6
d.DNP3TransportLayer.First = int(first)
sequence := data[10] & 0x3f // 6bit
d.DNP3TransportLayer.Sequence = int(sequence)
}
func (d *DNP3) applicationLayer(data []byte) {
// /***************************************************************************/
// /* Application Layer Bit-Masks */
// /***************************************************************************/
// #define DNP3_AL_UNS 0x10
// #define DNP3_AL_CON 0x20
// #define DNP3_AL_FIN 0x40
// #define DNP3_AL_FIR 0x80
// #define DNP3_AL_SEQ 0x0f
// #define DNP3_AL_FUNC 0xff
controlByte := fmt.Sprintf("0x%x", data[11])
d.DNP3ApplicationLayer.Control.ControlByte = controlByte | {
return d.restOfData
} | identifier_body |
dnp3.go | 3: "Objects prefixed with 4-octet index",
4: "Objects prefixed with 1-octet object size",
5: "Objects prefixed with 2-octet object size",
6: "Objects prefixed with 4-octet object size",
7: "Reserved",
}
/***************************************************************************/
/* Application Layer Object Prefix codes bits */
/***************************************************************************/
var ObjRangeSpecifierCodes = map[byte]string{
0: "8-bit Start and Stop Indices in Range Field",
1: "16-bit Start and Stop Indices in Range Field",
2: "32-bit Start and Stop Indices in Range Field",
3: "8-bit Absolute Address in Range Field",
4: "16-bit Absolute Address in Range Field",
5: "32-bit Absolute Address in Range Field",
6: "Length of Range field is 0 (no range field)",
7: "8-bit Single Field Quantity",
8: "16-bit Single Field Quantity",
9: "32-bit Single Field Quantity",
10: "Reserved",
11: "Free-format Qualifier, range field has 1 octet count of objects",
12: "Reserved",
13: "Reserved",
14: "Reserved",
15: "Reserved",
}
var (
errDNP3PacketTooShort = errors.New("DNS packet too short")
)
type DNP3 struct {
BaseLayer // Stores the packet bytes and payload bytes.
DNP3DataLinkLayer DNP3DataLinkLayer
DNP3TransportLayer DNP3TransportLayer
DNP3ApplicationLayer DNP3ApplicationLayer
SomeByte byte
AnotherByte byte
restOfData []byte
}
type DNP3DataLinkLayer struct {
Start string
Length int
Control struct {
ControlByte string
IsMaster int `json:"Is Master"`
PRM int `json:"Primary"`
FCB int `json:"Frame Count Bit"`
FCV int `json:"Frame Count Valid"`
FUNC string `json:"Function Code"`
}
Destination int
Source int
CRC string
}
type DNP3TransportLayer struct {
TransportByte string
Final int
First int
Sequence int
}
type DNP3ApplicationLayer struct {
Control struct {
ControlByte string
First int
Final int
Confirm int
Unsolicited int
Sequence int
}
Function string `json:"Function Code"`
IINCode string `json:"Internal Indication (IIN)"`
}
type DNP3AppObject struct {
Group int
Variation int
Qualifier int
RangeStart int
RangeStop int
DataType int
Length int
}
func (d *DNP3) LayerType() gopacket.LayerType { return LayerTypeDNP3 }
func (d *DNP3) LayerContents() []byte {
return []byte{d.SomeByte, d.AnotherByte}
}
func (d *DNP3) LayerPayload() []byte {
return d.restOfData
}
func (d *DNP3) DecodeFromBytes(data []byte, df gopacket.DecodeFeedback) error {
// If the data block is too short to be a DNP3 layer, then return an error.
if len(data) < 10 {
df.SetTruncated()
return errDNP3PacketTooShort
}
d.linkLayer(data)
d.transportLayer(data)
d.applicationLayer(data)
return nil
}
func decodeDNP3(data []byte, p gopacket.PacketBuilder) error {
// Attempt to decode the byte slice.
d := &DNP3{}
err := d.DecodeFromBytes(data, p)
if err != nil {
return err
}
// If the decoding worked, add the layer to the packet and set it
// as the application layer too, if there isn't already one.
p.AddLayer(d)
p.SetApplicationLayer(d)
d.BaseLayer = BaseLayer{Contents: data[:len(data)]}
d.BaseLayer.Payload = nil
return p.NextDecoder(gopacket.LayerTypePayload)
}
// CanDecode implements gopacket.DecodingLayer.
func (d *DNP3) CanDecode() gopacket.LayerClass {
return LayerTypeDNP3
}
// NextLayerType returns the layer type contained by this DecodingLayer.
func (d *DNP3) NextLayerType() gopacket.LayerType {
return gopacket.LayerTypePayload
}
// Payload returns nil, since TLS encrypted payload is inside TLSAppDataRecord
func (d *DNP3) Payload() []byte {
return nil
}
func appObject(bytesRead []byte) {
object := bytesRead[22:]
// indexSize := uint(object[2] & 0x70 >> 4)
// QualifierCode := uint(object[2] & 0x0F)
// fmt.Println(indexSize)
// fmt.Println(QualifierCode)
group := int(object[0])
variation := int(object[1])
qualifier := int(object[2])
rangeStart := int(object[3])
rangeStop := int(object[4])
dataType := int(object[5])
length := int(object[6])
appObject := DNP3AppObject{
Group: group,
Variation: variation,
Qualifier: qualifier,
RangeStart: rangeStart,
RangeStop: rangeStop,
DataType: dataType,
Length: length,
}
out, err := json.Marshal(appObject)
if err != nil {
panic(err)
}
fmt.Println(string(out))
}
func (d *DNP3) linkLayer(data []byte) {
start := d.hexConvert(data[0:2])
d.DNP3DataLinkLayer.Start = start
length := int(data[2])
d.DNP3DataLinkLayer.Length = length
ctlControl := d.hexConvert([]byte{data[3]})
d.DNP3DataLinkLayer.Control.ControlByte = ctlControl
IsMaster := int((data[3] & 0x80) >> 7)
d.DNP3DataLinkLayer.Control.IsMaster = IsMaster
PRM := int((data[3] & 0x40) >> 6)
d.DNP3DataLinkLayer.Control.PRM = PRM
FCB := int((data[3] & 0x20) >> 5)
d.DNP3DataLinkLayer.Control.FCB = FCB
FCV := int((data[3] & 0x10) >> 4)
d.DNP3DataLinkLayer.Control.FCV = FCV
FUNCCODE := data[3] & 0x0F
ctlFUNCCODE := fmt.Sprintf("%d", FUNCCODE)
var ctlFUNC string
if PRM == 0x00 {
ctlFUNC = SfCodes[FUNCCODE]
}
if PRM == 0x01 |
ctlFUNC = ctlFUNC + " (" + ctlFUNCCODE + ")"
d.DNP3DataLinkLayer.Control.FUNC = ctlFUNC
// TODO: make sure 0 to 65535
destination := fmt.Sprintf("%x%x", data[5], data[4])
destinationInt, _ := strconv.Atoi(destination)
d.DNP3DataLinkLayer.Destination = destinationInt
// TODO: make sure 0 to 65535
source := fmt.Sprintf("%x%x", data[7], data[6])
sourceInt, _ := strconv.Atoi(source)
d.DNP3DataLinkLayer.Source = sourceInt
// TODO: Is correct? Hesapla
crc := fmt.Sprintf("0x%x%x", data[9], data[8])
d.DNP3DataLinkLayer.CRC = crc
}
func (d *DNP3) transportLayer(data []byte) {
transport := fmt.Sprintf("0x%x", data[10])
d.DNP3TransportLayer.TransportByte = transport
final := data[10] & 0x80 >> 7
d.DNP3TransportLayer.Final = int(final)
first := data[10] & 0x40 >> 6
d.DNP3TransportLayer.First = int(first)
sequence := data[10] & 0x3f // 6bit
d.DNP3TransportLayer.Sequence = int(sequence)
}
func (d *DNP3) applicationLayer(data []byte) {
// /***************************************************************************/
// /* Application Layer Bit-Masks */
// /***************************************************************************/
// #define DNP3_AL_UNS 0x10
// #define DNP3_AL_CON 0x20
// #define DNP3_AL_FIN 0x40
// #define DNP3_AL_FIR 0x80
// #define DNP3_AL_SEQ 0x0f
// #define DNP3_AL_FUNC 0xff
controlByte := fmt.Sprintf("0x%x", data[11])
d.DNP3ApplicationLayer.Control.ControlByte = controlByte | {
ctlFUNC = PfCodes[FUNCCODE]
} | conditional_block |
config.js | UENTD_PORT, 10) || 24224;
internals.Config.nginxHost = process.env.UNMS_NGINX_HOST || '127.0.0.1';
internals.Config.nginxPort = parseInt(process.env.UNMS_NGINX_PORT, 10) || 12345;
internals.Config.defaultInternalHttpPort = 8081;
internals.Config.defaultInternalWsPort = 8082;
internals.Config.defaultHttpsPort = 443;
internals.Config.httpPort = parseInt(process.env.HTTP_PORT, 10) || internals.Config.defaultInternalHttpPort;
internals.Config.wsPort = parseInt(process.env.WS_PORT, 10) || internals.Config.defaultInternalWsPort;
internals.Config.publicHttpsPort = parseInt(process.env.PUBLIC_HTTPS_PORT, 10) || internals.Config.defaultHttpsPort;
internals.Config.publicWsPort = parseInt(process.env.PUBLIC_WS_PORT, 10) || internals.Config.publicHttpsPort;
internals.Config.secureLinkSecret = process.env.SECURE_LINK_SECRET || 'enigma';
internals.Config.leChallengeDir = './challenge';
internals.Config.useCustomSslCert = Boolean(process.env.SSL_CERT);
internals.Config.socketRPCTimeout = 15000; // generic rpc call timeout
internals.Config.socketRPCBackupTimeout = 180000; // backup rpc call timeout
internals.Config.pg = {
host: process.env.UNMS_PG_HOST || '127.0.0.1',
port: parseInt(process.env.UNMS_PG_PORT, 10) || 5432,
database: process.env.UNMS_PG_DATABASE || 'unms',
user: process.env.UNMS_PG_USER || 'postgres',
password: process.env.UNMS_PG_PASSWORD || '',
schema: process.env.UNMS_PG_SCHEMA || 'public',
};
internals.Config.rabbitMQ = {
host: process.env.UNMS_RABBITMQ_HOST || '127.0.0.1',
port: parseInt(process.env.UNMS_RABBITMQ_PORT, 10) || 5672,
exchange: 'unms',
};
internals.Config.publicDir = path.join(__dirname, 'public');
internals.Config.templatePaths = map(viewPath => path.join(__dirname, viewPath), [
'/public',
'/lib/api-docs/templates',
]);
internals.Config.publicPaths = map(viewPath => path.join(__dirname, viewPath), [
'/public',
'/lib/api-docs/public',
]);
// UNMS cloud related settings
internals.Config.cloudSettings = internals.Config.cloud ? {
domain: process.env.CLOUD_DOMAIN,
googleMapsApiKey: process.env.CLOUD_MAPS_API_KEY,
mapsProvider: process.env.CLOUD_MAPS_PROVIDER,
billingApiUrl: process.env.CLOUD_BILLING_API_URL,
ssoApiUrl: process.env.CLOUD_SSO_API_URL,
maintenanceWindowApiUrl: process.env.CLOUD_MAINTENANCE_WINDOW_API_URL,
redisInstanceId: process.env.UNMS_REDISDB_INSTANCE_ID,
smtpHostname: process.env.CLOUD_SMTP_HOSTNAME,
smtpPort: process.env.CLOUD_SMTP_PORT,
smtpUsername: process.env.CLOUD_SMTP_USERNAME,
smtpPassword: process.env.CLOUD_SMTP_PASSWORD,
smtpSender: process.env.CLOUD_SMTP_SENDER_ADDRESS,
smtpTimeout: process.env.CLOUD_SMTP_TIMEOUT,
smtpTlsAllowUnauthorized: process.env.CLOUD_SMTP_TLS_ALLOW_UNAUTHORIZED,
smtpSecurityMode: process.env.CLOUD_SMTP_SECURITY_MODE,
smtpAuthEnabled: process.env.CLOUD_SMTP_AUTH_ENABLED,
storage: {
gcsProjectId: process.env.CLOUD_GCS_PROJECT_ID,
gcsKeyFilename: process.env.CLOUD_GCS_KEY_FILENAME,
},
} : null;
internals.Config.branch = process.env.BRANCH || 'master';
internals.Config.unmsLatestVersionUrl =
`https://api.github.com/repos/Ubiquiti-App/UNMS/contents/latest-version?ref=${internals.Config.branch}`;
// discovery
internals.Config.discoveryScanTimeout = toMs('second', 5);
internals.Config.discoveryIpRangeMaxSize = 65536; // 2^16
// Reporting
internals.Config.sentryDSN =
'https://3a0cdfa562074c6ca018c01b666d37c5:[email protected]/120911';
// production log token in Logentries
internals.Config.logentriesToken = 'b740d3fa-9176-43b8-b259-58a6660590dc';
// affects password hashing time, see https://github.com/kelektiv/node.bcrypt.js#a-note-on-rounds
internals.Config.pwdSaltRounds = 8;
internals.Config.jwtToken = 'x-auth-token';
internals.Config.authStrategy = 'jwt';
internals.Config.defaultEmail = '[email protected]';
internals.Config.defaultUsername = 'ubnt';
internals.Config.defaultPwd = 'ubntubnt';
internals.Config.passwordTokenExpiry = toMs('minute', 30);
internals.Config.sessionTimeout = toMs('minute', 30);
internals.Config.extendedSessionTimeout = toMs('hour', 24);
internals.Config.hashAlgorithm = 'aes-256-cbc';
internals.Config.totpAuthSecretOptions = {
issuer: 'UNMS',
};
internals.Config.httpConnection = {
port: internals.Config.httpPort,
routes: {
cors: {
origin: ['*'],
headers: ['Accept', 'Authorization', 'Content-Type', 'If-None-Match', internals.Config.jwtToken],
exposedHeaders: [internals.Config.jwtToken],
},
files: {
relativeTo: internals.Config.publicDir,
},
security: true,
},
state: {
strictHeader: false,
},
};
internals.Config.deviceConfigBackup = {
dir: './data/config-backups',
minimumFiles: 6,
ttl: toMs('day', 30),
multiBackup: {
dir: 'multi',
ttl: toMs('hour', 1),
},
fileMaxBytes: 10000000,
queue: {
delay: toMs('minute', 2),
concurency: 5,
},
};
internals.Config.deviceTypes = {
all: ['olt', 'erouter'],
};
internals.Config.deviceServices = {
ntpServers: {
ntpServer1: '0.pool.ntp.org',
ntpServer2: '1.pool.ntp.org',
},
};
internals.Config.statisticsIntervals = {
minute: { length: toMs('minute', 2), period: toMs('second', 1) },
hour: { length: toMs('hour', 1), period: toMs('second', 15) },
day: { length: toMs('day', 1), period: toMs('minute', 15) },
month: { length: toMs('month', 1), period: toMs('hour', 8) },
quarter: { length: toMs('month', 3), period: toMs('hour', 24) },
year: { length: toMs('year', 1), period: toMs('hour', 106) },
};
internals.Config.statisticsIntervalMapping = {
[IntervalEnum.Hour]: ['hour'],
[IntervalEnum.Day]: ['day'],
[IntervalEnum.Month]: ['month'],
[IntervalEnum.Quarter]: ['quarter'],
[IntervalEnum.Year]: ['year'],
};
const findMaxStatisticsIntervalPeriod = flow(
map(name => internals.Config.statisticsIntervals[name].period),
max
);
internals.Config.statisticsIntervalPeriodMapping = reduce(
internals.Config.statisticsIntervalMapping,
(acc, values, interval) => (Object.assign(acc, { [interval]: findMaxStatisticsIntervalPeriod(values) })),
{}
);
const findMaxStatisticsIntervalLength = flow(
map(name => internals.Config.statisticsIntervals[name].length),
max
);
internals.Config.statisticsIntervalLengthMapping = reduce(
internals.Config.statisticsIntervalMapping,
(acc, values, interval) => (Object.assign(acc, { [interval]: findMaxStatisticsIntervalLength(values) })),
{}
);
internals.Config.deviceLog = {
[DevicePropertyEnum.Cpu]: {
name: 'CPU',
unit: '%',
limit: 90,
interval: toMs('second', 30),
level: LogLevelEnum.Warning,
logType: LogTypeEnum.DeviceCpuOverLimit,
},
[DevicePropertyEnum.Ram]: {
name: 'RAM',
unit: '%',
limit: 90,
interval: toMs('second', 10),
level: LogLevelEnum.Warning,
logType: LogTypeEnum.DeviceRamOverLimit,
},
};
internals.Config.outages = {
unmsStartGracePeriod: toMs('minute', 2),
lagGracePeriod: toMs('second', 5), // avoid responding to possible lagging device's communication
maxAge: toMs('month', 3),
};
internals.Config.siteImages = { | random_line_split |
||
packet.go | kCidDefaultLength is the length of connection ID we generate.
// TODO: make this configurable.
const kCidDefaultLength = 5
// ConnectionId identifies the connection that a packet belongs to.
type ConnectionId []byte
// String stringifies a connection ID in the natural way.
func (c ConnectionId) String() string {
return hex.EncodeToString(c)
}
// EncodeLength produces the length encoding used in the long packet header.
func (c ConnectionId) EncodeLength() byte {
if len(c) == 0 {
return 0
}
assert(len(c) >= 4 && len(c) <= 18)
return byte(len(c) - 3)
}
// The PDU definition for the header.
// These types are capitalized so that |codec| can use them.
type packetHeader struct {
// Type is the on-the-wire form of the packet type.
// Consult getHeaderType if you want a value that corresponds to the
// definition of packetType.
Type packetType
Version VersionNumber
ConnectionIDLengths byte
DestinationConnectionID ConnectionId
SourceConnectionID ConnectionId
TokenLength uint8
Token []byte
PayloadLength uint64 `tls:"varint"`
// In order to decode a short header, the length of the connection
// ID must be set in |shortCidLength| before decoding.
shortCidLength uintptr
}
func (p packetHeader) String() string {
ht := "SHORT"
if p.Type.isLongHeader() {
ht = "LONG"
}
return fmt.Sprintf("%s PT=%v", ht, p.getHeaderType())
}
func (p *packetHeader) getHeaderType() packetType {
if p.Type.isLongHeader() {
return p.Type & 0x7f
}
return packetTypeProtectedShort
}
type packet struct {
packetHeader
PacketNumber uint64 // Never more than 32 bits on the wire.
payload []byte
}
// This reads from p.ConnectionIDLengths.
func (p packetHeader) ConnectionIDLengths__length() uintptr {
if p.Type.isLongHeader() {
return 1
}
return 0
}
func (p packetHeader) TokenLength__length() uintptr {
if p.getHeaderType() != packetTypeInitial {
assert(len(p.Token) == 0)
return 0
}
return 1
}
func (p packetHeader) Token__length() uintptr {
if p.getHeaderType() != packetTypeInitial {
assert(len(p.Token) == 0)
return 0
}
return uintptr(p.TokenLength)
}
func (p packetHeader) DestinationConnectionID__length() uintptr {
if !p.Type.isLongHeader() {
return p.shortCidLength
}
l := p.ConnectionIDLengths >> 4
if l != 0 {
l += 3
}
return uintptr(l)
}
func (p packetHeader) SourceConnectionID__length() uintptr {
if !p.Type.isLongHeader() {
return 0
}
l := p.ConnectionIDLengths & 0xf
if l != 0 {
l += 3
}
return uintptr(l)
}
func (p packetHeader) PayloadLength__length() uintptr {
if p.Type.isLongHeader() {
return codecDefaultSize
}
return 0
}
func (p packetHeader) Version__length() uintptr {
if p.Type.isLongHeader() {
return 4
}
return 0
}
func newPacket(pt packetType, destCid ConnectionId, srcCid ConnectionId, ver VersionNumber, pn uint64, payload []byte, aeadOverhead int) *packet {
if pt == packetTypeProtectedShort {
// Only support writing the 32-bit packet number.
pt = packetType(0x2 | packetFlagShortHeader)
srcCid = nil
} else {
pt = pt | packetType(packetFlagLongHeader)
}
lengths := (destCid.EncodeLength() << 4) | srcCid.EncodeLength()
return &packet{
packetHeader: packetHeader{
Type: pt,
ConnectionIDLengths: lengths,
DestinationConnectionID: destCid,
SourceConnectionID: srcCid,
Version: ver,
PayloadLength: uint64(len(payload) + 4 + aeadOverhead),
},
PacketNumber: pn,
payload: payload,
}
}
type versionNegotiationPacket struct {
Versions []byte
}
func newVersionNegotiationPacket(versions []VersionNumber) *versionNegotiationPacket {
var buf bytes.Buffer
for _, v := range versions {
buf.Write(encodeArgs(v))
}
return &versionNegotiationPacket{buf.Bytes()}
}
/*
We don't use these.
func encodePacket(c ConnectionState, aead Aead, p *Packet) ([]byte, error) {
hdr, err := encode(&p.packetHeader)
if err != nil {
return nil, err
}
b, err := aead.protect(p.packetHeader.PacketNumber, hdr, p.payload)
if err != nil {
return nil, err
}
return encodeArgs(hdr, b), nil
}
func decodePacket(c ConnectionState, aead Aead, b []byte) (*Packet, error) {
// Parse the header
var hdr packetHeader
br, err := decode(&hdr, b)
if err != nil {
return nil, err
}
hdr.PacketNumber = c.expandPacketNumber(hdr.PacketNumber)
pt, err := aead.unprotect(hdr.PacketNumber, b[0:br], b[br:])
if err != nil {
return nil, err
}
return &Packet{hdr, pt}, nil
}
*/
func dumpPacket(payload []byte) string {
first := true
ret := fmt.Sprintf("%d=[", len(payload))
for len(payload) > 0 {
if !first {
ret += ", "
}
first = false
n, f, err := decodeFrame(payload)
if err != nil {
ret += fmt.Sprintf("Undecoded: [%x]", payload)
break
}
payload = payload[n:]
// TODO([email protected]): Not sure why %v doesn't work
ret += f.String()
}
ret += "]"
return ret
}
type pneCipherFactory interface {
create(sample []byte) cipher.Stream
}
type pneCipherFactoryAES struct {
block cipher.Block
}
func newPneCipherFactoryAES(key []byte) pneCipherFactory {
inner, err := aes.NewCipher(key)
assert(err == nil)
if err != nil {
return nil
}
return &pneCipherFactoryAES{block: inner}
}
func (f *pneCipherFactoryAES) create(sample []byte) cipher.Stream {
if len(sample) != 16 {
return nil
}
return cipher.NewCTR(f.block, sample)
}
func xorPacketNumber(hdr *packetHeader, hdrlen int, pnbuf []byte, p []byte, factory pneCipherFactory) error {
logf(logTypeTrace, "PNE Operation: hdrlen=%v, hdr=%x, payload=%x", hdrlen, p[:hdrlen], p)
// The packet must be at least long enough to contain
// the header, plus a minimum 1-byte PN, plus the sample.
sample_length := 16
if sample_length > len(p)-(hdrlen+1) {
logf(logTypePacket, "Packet too short")
return nil
}
// Now compute the offset
sample_offset := hdrlen + 4
if sample_offset+sample_length > len(p) {
sample_offset = len(p) - sample_length
}
sample := p[sample_offset : sample_offset+sample_length]
logf(logTypeTrace, "PNE sample_offset=%d sample=%x", sample_offset, sample)
stream := factory.create(sample)
stream.XORKeyStream(pnbuf, p[hdrlen:hdrlen+len(pnbuf)])
return nil
}
var pnPatterns = []struct {
prefix byte
mask byte
length int
}{
{
0, 0x80, 1,
},
{
0x80, 0xc0, 2,
},
{
0xc0, 0xc0, 4,
},
}
const ()
func encodePacketNumber(pn uint64, l int) []byte {
var buf bytes.Buffer
i := 0
for i, _ = range pnPatterns {
if pnPatterns[i].length == l {
break
}
}
uintEncodeInt(&buf, pn, uintptr(l))
b := buf.Bytes()
b[0] &= ^pnPatterns[i].mask
b[0] |= pnPatterns[i].prefix
return b
}
func decodePacketNumber(buf []byte) (uint64, int, error) {
if len(buf) < 1 {
return 0, 0, fmt.Errorf("Zero-length packet number")
}
i := 0
for i, _ = range pnPatterns {
if pnPatterns[i].mask&buf[0] == pnPatterns[i].prefix {
break
}
}
pat := &pnPatterns[i]
if len(buf) < pat.length | {
return 0, 0, fmt.Errorf("Buffer too short for packet number (%v < %v)", len(buf), pat.length)
} | conditional_block |
|
packet.go | 9 0 1
+-+-+-+-+-+-+-+-+
|0|K|1|1|0|R R R|
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
| Destination Connection ID (0..144) ...
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
| Packet Number (8/16/32) ...
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
| Protected Payload (*) ...
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
*/
const (
packetFlagLongHeader = byte(0x80)
packetFlagK = byte(0x40)
packetFlagShortHeader = byte(0x30)
)
// This packet type differs considerably from the spec. It includes both
// long and short headers in the same value space. Long headers are from
// 0-0x7f (inclusive); short headers are always represented as 0xff.
type packetType byte
const (
packetTypeInitial = packetType(0x7f)
packetTypeRetry = packetType(0x7e)
packetTypeHandshake = packetType(0x7d)
packetType0RTTProtected = packetType(0x7c)
packetTypeProtectedShort = packetType(0x00) // Not a real type
)
func (pt packetType) isLongHeader() bool {
return pt&packetType(packetFlagLongHeader) != 0
}
func (pt packetType) isProtected() bool {
if !pt.isLongHeader() {
return true
}
switch pt & 0x7f {
case packetTypeInitial, packetTypeHandshake, packetTypeRetry:
return false
}
return true
}
func (pt packetType) String() string {
switch pt {
case packetTypeInitial:
return "Initial"
case packetTypeRetry:
return "Retry"
case packetTypeHandshake:
return "Handshake"
case packetType0RTTProtected:
return "0-RTT"
case packetTypeProtectedShort:
return "1-RTT"
default:
return fmt.Sprintf("%x", uint8(pt))
}
}
// kCidDefaultLength is the length of connection ID we generate.
// TODO: make this configurable.
const kCidDefaultLength = 5
// ConnectionId identifies the connection that a packet belongs to.
type ConnectionId []byte
// String stringifies a connection ID in the natural way.
func (c ConnectionId) String() string {
return hex.EncodeToString(c)
}
// EncodeLength produces the length encoding used in the long packet header.
func (c ConnectionId) EncodeLength() byte {
if len(c) == 0 {
return 0
}
assert(len(c) >= 4 && len(c) <= 18)
return byte(len(c) - 3)
}
// The PDU definition for the header.
// These types are capitalized so that |codec| can use them.
type packetHeader struct {
// Type is the on-the-wire form of the packet type.
// Consult getHeaderType if you want a value that corresponds to the
// definition of packetType.
Type packetType
Version VersionNumber
ConnectionIDLengths byte
DestinationConnectionID ConnectionId
SourceConnectionID ConnectionId
TokenLength uint8
Token []byte
PayloadLength uint64 `tls:"varint"`
// In order to decode a short header, the length of the connection
// ID must be set in |shortCidLength| before decoding.
shortCidLength uintptr
}
func (p packetHeader) String() string {
ht := "SHORT"
if p.Type.isLongHeader() {
ht = "LONG"
}
return fmt.Sprintf("%s PT=%v", ht, p.getHeaderType())
}
func (p *packetHeader) getHeaderType() packetType {
if p.Type.isLongHeader() {
return p.Type & 0x7f
}
return packetTypeProtectedShort
}
type packet struct {
packetHeader
PacketNumber uint64 // Never more than 32 bits on the wire.
payload []byte
}
// This reads from p.ConnectionIDLengths.
func (p packetHeader) ConnectionIDLengths__length() uintptr {
if p.Type.isLongHeader() {
return 1
}
return 0
}
func (p packetHeader) TokenLength__length() uintptr {
if p.getHeaderType() != packetTypeInitial {
assert(len(p.Token) == 0)
return 0
}
return 1
}
func (p packetHeader) Token__length() uintptr {
if p.getHeaderType() != packetTypeInitial {
assert(len(p.Token) == 0)
return 0
}
return uintptr(p.TokenLength)
}
func (p packetHeader) DestinationConnectionID__length() uintptr {
if !p.Type.isLongHeader() {
return p.shortCidLength
}
l := p.ConnectionIDLengths >> 4
if l != 0 {
l += 3
}
return uintptr(l)
}
func (p packetHeader) SourceConnectionID__length() uintptr {
if !p.Type.isLongHeader() {
return 0
}
l := p.ConnectionIDLengths & 0xf
if l != 0 {
l += 3
}
return uintptr(l)
}
func (p packetHeader) PayloadLength__length() uintptr {
if p.Type.isLongHeader() {
return codecDefaultSize
}
return 0
}
func (p packetHeader) Version__length() uintptr {
if p.Type.isLongHeader() {
return 4
}
return 0
}
func newPacket(pt packetType, destCid ConnectionId, srcCid ConnectionId, ver VersionNumber, pn uint64, payload []byte, aeadOverhead int) *packet {
if pt == packetTypeProtectedShort {
// Only support writing the 32-bit packet number.
pt = packetType(0x2 | packetFlagShortHeader)
srcCid = nil
} else {
pt = pt | packetType(packetFlagLongHeader)
}
lengths := (destCid.EncodeLength() << 4) | srcCid.EncodeLength()
return &packet{
packetHeader: packetHeader{
Type: pt,
ConnectionIDLengths: lengths,
DestinationConnectionID: destCid,
SourceConnectionID: srcCid,
Version: ver,
PayloadLength: uint64(len(payload) + 4 + aeadOverhead),
},
PacketNumber: pn,
payload: payload,
}
}
type versionNegotiationPacket struct {
Versions []byte
}
func newVersionNegotiationPacket(versions []VersionNumber) *versionNegotiationPacket {
var buf bytes.Buffer
for _, v := range versions {
buf.Write(encodeArgs(v))
}
return &versionNegotiationPacket{buf.Bytes()}
}
/*
We don't use these.
func encodePacket(c ConnectionState, aead Aead, p *Packet) ([]byte, error) {
hdr, err := encode(&p.packetHeader)
if err != nil {
return nil, err
}
b, err := aead.protect(p.packetHeader.PacketNumber, hdr, p.payload)
if err != nil {
return nil, err
}
return encodeArgs(hdr, b), nil
}
func decodePacket(c ConnectionState, aead Aead, b []byte) (*Packet, error) {
// Parse the header
var hdr packetHeader
br, err := decode(&hdr, b)
if err != nil {
return nil, err
}
hdr.PacketNumber = c.expandPacketNumber(hdr.PacketNumber)
pt, err := aead.unprotect(hdr.PacketNumber, b[0:br], b[br:])
if err != nil {
return nil, err
}
return &Packet{hdr, pt}, nil
}
*/
func dumpPacket(payload []byte) string {
first := true
ret := fmt.Sprintf("%d=[", len(payload))
for len(payload) > 0 {
if !first {
ret += ", "
}
first = false
n, f, err := decodeFrame(payload)
if err != nil {
ret += fmt.Sprintf("Undecoded: [%x]", payload)
break
}
payload = payload[n:]
// TODO([email protected]): Not sure why %v doesn't work
ret += f.String()
}
ret += "]"
return ret
}
type pneCipherFactory interface {
create(sample []byte) cipher.Stream
}
type pneCipherFactoryAES struct {
block cipher.Block
}
func newPneCipherFactoryAES(key []byte) pneCipherFactory {
inner, err := aes.NewCipher(key)
assert(err == nil)
if err != nil {
return nil
}
return &pneCipherFactoryAES{block: inner}
}
func (f *pneCipherFactoryAES) create(sample []byte) cipher.Stream {
if len(sample) != 16 {
return nil
}
return cipher.NewCTR(f.block, sample)
}
func xorPacketNumber(hdr *packetHeader, hdrlen int, pnbuf []byte, p []byte, factory pneCipherFactory) error {
logf(logTypeTrace, "PNE Operation: hdrlen=%v, hdr=%x, payload=%x", hdrlen, p[:hdrlen], p)
// The packet must be at least long enough to contain
// the header, plus a minimum 1-byte PN, plus the sample. | sample_length := 16
if sample_length > len(p)-(hdrlen+1) { | random_line_split |
|
packet.go | 2) |
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
|DCIL(4)|SCIL(4)|
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
| Destination Connection ID (0/32..144) ...
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
| Source Connection ID (0/32..144) ...
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
| Token Length (i) ...
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
| Token (*) ...
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
| Length (i) ...
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
| Packet Number (8/16/32) |
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
| Payload (*) ...
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
0 1 2 3
0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 0 1
+-+-+-+-+-+-+-+-+
|0|K|1|1|0|R R R|
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
| Destination Connection ID (0..144) ...
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
| Packet Number (8/16/32) ...
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
| Protected Payload (*) ...
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
*/
const (
packetFlagLongHeader = byte(0x80)
packetFlagK = byte(0x40)
packetFlagShortHeader = byte(0x30)
)
// This packet type differs considerably from the spec. It includes both
// long and short headers in the same value space. Long headers are from
// 0-0x7f (inclusive); short headers are always represented as 0xff.
type packetType byte
const (
packetTypeInitial = packetType(0x7f)
packetTypeRetry = packetType(0x7e)
packetTypeHandshake = packetType(0x7d)
packetType0RTTProtected = packetType(0x7c)
packetTypeProtectedShort = packetType(0x00) // Not a real type
)
func (pt packetType) isLongHeader() bool {
return pt&packetType(packetFlagLongHeader) != 0
}
func (pt packetType) isProtected() bool {
if !pt.isLongHeader() {
return true
}
switch pt & 0x7f {
case packetTypeInitial, packetTypeHandshake, packetTypeRetry:
return false
}
return true
}
func (pt packetType) String() string {
switch pt {
case packetTypeInitial:
return "Initial"
case packetTypeRetry:
return "Retry"
case packetTypeHandshake:
return "Handshake"
case packetType0RTTProtected:
return "0-RTT"
case packetTypeProtectedShort:
return "1-RTT"
default:
return fmt.Sprintf("%x", uint8(pt))
}
}
// kCidDefaultLength is the length of connection ID we generate.
// TODO: make this configurable.
const kCidDefaultLength = 5
// ConnectionId identifies the connection that a packet belongs to.
type ConnectionId []byte
// String stringifies a connection ID in the natural way.
func (c ConnectionId) String() string {
return hex.EncodeToString(c)
}
// EncodeLength produces the length encoding used in the long packet header.
func (c ConnectionId) EncodeLength() byte {
if len(c) == 0 {
return 0
}
assert(len(c) >= 4 && len(c) <= 18)
return byte(len(c) - 3)
}
// The PDU definition for the header.
// These types are capitalized so that |codec| can use them.
type packetHeader struct {
// Type is the on-the-wire form of the packet type.
// Consult getHeaderType if you want a value that corresponds to the
// definition of packetType.
Type packetType
Version VersionNumber
ConnectionIDLengths byte
DestinationConnectionID ConnectionId
SourceConnectionID ConnectionId
TokenLength uint8
Token []byte
PayloadLength uint64 `tls:"varint"`
// In order to decode a short header, the length of the connection
// ID must be set in |shortCidLength| before decoding.
shortCidLength uintptr
}
func (p packetHeader) String() string {
ht := "SHORT"
if p.Type.isLongHeader() {
ht = "LONG"
}
return fmt.Sprintf("%s PT=%v", ht, p.getHeaderType())
}
func (p *packetHeader) getHeaderType() packetType {
if p.Type.isLongHeader() {
return p.Type & 0x7f
}
return packetTypeProtectedShort
}
type packet struct {
packetHeader
PacketNumber uint64 // Never more than 32 bits on the wire.
payload []byte
}
// This reads from p.ConnectionIDLengths.
func (p packetHeader) ConnectionIDLengths__length() uintptr {
if p.Type.isLongHeader() {
return 1
}
return 0
}
func (p packetHeader) TokenLength__length() uintptr {
if p.getHeaderType() != packetTypeInitial {
assert(len(p.Token) == 0)
return 0
}
return 1
}
func (p packetHeader) Token__length() uintptr {
if p.getHeaderType() != packetTypeInitial {
assert(len(p.Token) == 0)
return 0
}
return uintptr(p.TokenLength)
}
func (p packetHeader) DestinationConnectionID__length() uintptr {
if !p.Type.isLongHeader() {
return p.shortCidLength
}
l := p.ConnectionIDLengths >> 4
if l != 0 {
l += 3
}
return uintptr(l)
}
func (p packetHeader) SourceConnectionID__length() uintptr {
if !p.Type.isLongHeader() {
return 0
}
l := p.ConnectionIDLengths & 0xf
if l != 0 {
l += 3
}
return uintptr(l)
}
func (p packetHeader) PayloadLength__length() uintptr {
if p.Type.isLongHeader() {
return codecDefaultSize
}
return 0
}
func (p packetHeader) Version__length() uintptr {
if p.Type.isLongHeader() {
return 4
}
return 0
}
func newPacket(pt packetType, destCid ConnectionId, srcCid ConnectionId, ver VersionNumber, pn uint64, payload []byte, aeadOverhead int) *packet {
if pt == packetTypeProtectedShort {
// Only support writing the 32-bit packet number.
pt = packetType(0x2 | packetFlagShortHeader)
srcCid = nil
} else {
pt = pt | packetType(packetFlagLongHeader)
}
lengths := (destCid.EncodeLength() << 4) | srcCid.EncodeLength()
return &packet{
packetHeader: packetHeader{
Type: pt,
ConnectionIDLengths: lengths,
DestinationConnectionID: destCid,
SourceConnectionID: srcCid,
Version: ver,
PayloadLength: uint64(len(payload) + 4 + aeadOverhead),
},
PacketNumber: pn,
payload: payload,
}
}
type versionNegotiationPacket struct {
Versions []byte
}
func | (versions []VersionNumber) *versionNegotiationPacket {
var buf bytes.Buffer
for _, v := range versions {
buf.Write(encodeArgs(v))
}
return &versionNegotiationPacket{buf.Bytes()}
}
/*
We don't use these.
func encodePacket(c ConnectionState, aead Aead, p *Packet) ([]byte, error) {
hdr, err := encode(&p.packetHeader)
if err != nil {
return nil, err
}
b, err := aead.protect(p.packetHeader.PacketNumber, hdr, p.payload)
if err != nil {
return nil, err
}
return encodeArgs(hdr, b), nil
}
func decodePacket(c ConnectionState, aead Aead, b []byte) (*Packet, error) {
// Parse the header
var hdr packetHeader
br, err := decode(&hdr, b)
if err != nil {
return nil, err
}
hdr.PacketNumber = c.expandPacketNumber(hdr.PacketNumber)
pt, err := aead.unprotect(hdr.PacketNumber, b[0:br], b[br:])
if err != nil {
return nil, err
}
return &Packet{hdr, pt}, nil
}
*/
func dumpPacket(payload []byte) string {
first := true
ret := fmt.Sprintf("%d=[", len(payload))
for len(payload) > 0 {
if !first {
ret += ", "
}
first = false
n, f, err := decodeFrame(payload)
if err != nil {
ret += fmt.Sprintf("Undecoded: [%x]", payload)
break
}
payload = payload[n:]
// TODO([email protected]): Not sure why %v doesn't work
ret += f.String()
}
ret += "]"
return ret
}
type pneCipherFactory interface {
create(sample []byte) cipher.Stream
}
type pneCipherFactoryAES struct {
block cipher.Block
}
| newVersionNegotiationPacket | identifier_name |
packet.go | 2) |
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
|DCIL(4)|SCIL(4)|
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
| Destination Connection ID (0/32..144) ...
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
| Source Connection ID (0/32..144) ...
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
| Token Length (i) ...
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
| Token (*) ...
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
| Length (i) ...
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
| Packet Number (8/16/32) |
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
| Payload (*) ...
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
0 1 2 3
0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 0 1
+-+-+-+-+-+-+-+-+
|0|K|1|1|0|R R R|
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
| Destination Connection ID (0..144) ...
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
| Packet Number (8/16/32) ...
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
| Protected Payload (*) ...
+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
*/
const (
packetFlagLongHeader = byte(0x80)
packetFlagK = byte(0x40)
packetFlagShortHeader = byte(0x30)
)
// This packet type differs considerably from the spec. It includes both
// long and short headers in the same value space. Long headers are from
// 0-0x7f (inclusive); short headers are always represented as 0xff.
type packetType byte
const (
packetTypeInitial = packetType(0x7f)
packetTypeRetry = packetType(0x7e)
packetTypeHandshake = packetType(0x7d)
packetType0RTTProtected = packetType(0x7c)
packetTypeProtectedShort = packetType(0x00) // Not a real type
)
func (pt packetType) isLongHeader() bool {
return pt&packetType(packetFlagLongHeader) != 0
}
func (pt packetType) isProtected() bool |
func (pt packetType) String() string {
switch pt {
case packetTypeInitial:
return "Initial"
case packetTypeRetry:
return "Retry"
case packetTypeHandshake:
return "Handshake"
case packetType0RTTProtected:
return "0-RTT"
case packetTypeProtectedShort:
return "1-RTT"
default:
return fmt.Sprintf("%x", uint8(pt))
}
}
// kCidDefaultLength is the length of connection ID we generate.
// TODO: make this configurable.
const kCidDefaultLength = 5
// ConnectionId identifies the connection that a packet belongs to.
type ConnectionId []byte
// String stringifies a connection ID in the natural way.
func (c ConnectionId) String() string {
return hex.EncodeToString(c)
}
// EncodeLength produces the length encoding used in the long packet header.
func (c ConnectionId) EncodeLength() byte {
if len(c) == 0 {
return 0
}
assert(len(c) >= 4 && len(c) <= 18)
return byte(len(c) - 3)
}
// The PDU definition for the header.
// These types are capitalized so that |codec| can use them.
type packetHeader struct {
// Type is the on-the-wire form of the packet type.
// Consult getHeaderType if you want a value that corresponds to the
// definition of packetType.
Type packetType
Version VersionNumber
ConnectionIDLengths byte
DestinationConnectionID ConnectionId
SourceConnectionID ConnectionId
TokenLength uint8
Token []byte
PayloadLength uint64 `tls:"varint"`
// In order to decode a short header, the length of the connection
// ID must be set in |shortCidLength| before decoding.
shortCidLength uintptr
}
func (p packetHeader) String() string {
ht := "SHORT"
if p.Type.isLongHeader() {
ht = "LONG"
}
return fmt.Sprintf("%s PT=%v", ht, p.getHeaderType())
}
func (p *packetHeader) getHeaderType() packetType {
if p.Type.isLongHeader() {
return p.Type & 0x7f
}
return packetTypeProtectedShort
}
type packet struct {
packetHeader
PacketNumber uint64 // Never more than 32 bits on the wire.
payload []byte
}
// This reads from p.ConnectionIDLengths.
func (p packetHeader) ConnectionIDLengths__length() uintptr {
if p.Type.isLongHeader() {
return 1
}
return 0
}
func (p packetHeader) TokenLength__length() uintptr {
if p.getHeaderType() != packetTypeInitial {
assert(len(p.Token) == 0)
return 0
}
return 1
}
func (p packetHeader) Token__length() uintptr {
if p.getHeaderType() != packetTypeInitial {
assert(len(p.Token) == 0)
return 0
}
return uintptr(p.TokenLength)
}
func (p packetHeader) DestinationConnectionID__length() uintptr {
if !p.Type.isLongHeader() {
return p.shortCidLength
}
l := p.ConnectionIDLengths >> 4
if l != 0 {
l += 3
}
return uintptr(l)
}
func (p packetHeader) SourceConnectionID__length() uintptr {
if !p.Type.isLongHeader() {
return 0
}
l := p.ConnectionIDLengths & 0xf
if l != 0 {
l += 3
}
return uintptr(l)
}
func (p packetHeader) PayloadLength__length() uintptr {
if p.Type.isLongHeader() {
return codecDefaultSize
}
return 0
}
func (p packetHeader) Version__length() uintptr {
if p.Type.isLongHeader() {
return 4
}
return 0
}
func newPacket(pt packetType, destCid ConnectionId, srcCid ConnectionId, ver VersionNumber, pn uint64, payload []byte, aeadOverhead int) *packet {
if pt == packetTypeProtectedShort {
// Only support writing the 32-bit packet number.
pt = packetType(0x2 | packetFlagShortHeader)
srcCid = nil
} else {
pt = pt | packetType(packetFlagLongHeader)
}
lengths := (destCid.EncodeLength() << 4) | srcCid.EncodeLength()
return &packet{
packetHeader: packetHeader{
Type: pt,
ConnectionIDLengths: lengths,
DestinationConnectionID: destCid,
SourceConnectionID: srcCid,
Version: ver,
PayloadLength: uint64(len(payload) + 4 + aeadOverhead),
},
PacketNumber: pn,
payload: payload,
}
}
type versionNegotiationPacket struct {
Versions []byte
}
func newVersionNegotiationPacket(versions []VersionNumber) *versionNegotiationPacket {
var buf bytes.Buffer
for _, v := range versions {
buf.Write(encodeArgs(v))
}
return &versionNegotiationPacket{buf.Bytes()}
}
/*
We don't use these.
func encodePacket(c ConnectionState, aead Aead, p *Packet) ([]byte, error) {
hdr, err := encode(&p.packetHeader)
if err != nil {
return nil, err
}
b, err := aead.protect(p.packetHeader.PacketNumber, hdr, p.payload)
if err != nil {
return nil, err
}
return encodeArgs(hdr, b), nil
}
func decodePacket(c ConnectionState, aead Aead, b []byte) (*Packet, error) {
// Parse the header
var hdr packetHeader
br, err := decode(&hdr, b)
if err != nil {
return nil, err
}
hdr.PacketNumber = c.expandPacketNumber(hdr.PacketNumber)
pt, err := aead.unprotect(hdr.PacketNumber, b[0:br], b[br:])
if err != nil {
return nil, err
}
return &Packet{hdr, pt}, nil
}
*/
func dumpPacket(payload []byte) string {
first := true
ret := fmt.Sprintf("%d=[", len(payload))
for len(payload) > 0 {
if !first {
ret += ", "
}
first = false
n, f, err := decodeFrame(payload)
if err != nil {
ret += fmt.Sprintf("Undecoded: [%x]", payload)
break
}
payload = payload[n:]
// TODO([email protected]): Not sure why %v doesn't work
ret += f.String()
}
ret += "]"
return ret
}
type pneCipherFactory interface {
create(sample []byte) cipher.Stream
}
type pneCipherFactoryAES struct {
block cipher.Block
| {
if !pt.isLongHeader() {
return true
}
switch pt & 0x7f {
case packetTypeInitial, packetTypeHandshake, packetTypeRetry:
return false
}
return true
} | identifier_body |
TCP_echo_server.py | Error:
search_result = False
except ssl.SSLError:
search_result = False
if search_result:
print('\n用户' + account['username'] + '登陆成功!\n')
writer.write('Login Success'.encode())
await writer.drain()
return True
else:
writer.write('Need Email'.encode())
await writer.drain()
email = await reader.read(1024)
verify_email = re.match(r'^[0-9a-zA-Z_]{0,19}@[0-9a-zA-Z]{1,13}\.[com,cn,net]{1,3}(.cn)?$', email.decode())
if verify_email:
email = verify_email.group()
sql = "insert into user(username,password,email) values ('{}','{}','{}')".format(
str(account['username']), str(account['password']), str(email))
try:
cur.execute(sql)
con.commit()
print('\n用户' + account['username'] + '注册成功!\n')
writer.write('Register Success'.encode())
await writer.drain()
return True
except Exception as e:
writer.write('Register Fail'.encode())
await writer.drain()
return False
else:
writer.write('Register Fail'.encode())
await writer.drain()
return False
def creat_server_ssl():
ssl_ctx = ssl.SSLContext(ssl.PROTOCOL_TLS_SERVER)
ssl_ctx.options |= ssl.OP_NO_TLSv1
ssl_ctx.options |= ssl.OP_NO_TLSv1_1
ssl_ctx.options |= ssl.OP_SINGLE_DH_USE
ssl_ctx.options |= ssl.OP_SINGLE_ECDH_USE
ssl_ctx.load_cert_chain(certfile='./server_ssl/mycertfile.pem', keyfile='./server_ssl/mykeyfile.pem')
ssl_ctx.load_verify_locations(cafile='./server_ssl/mycertfile.pem')
ssl_ctx.check_hostname = False
ssl_ctx.verify_mode = ssl.VerifyMode.CERT_REQUIRED
ssl_ctx.set_ciphers('ECDHE-ECDSA-AES256-GCM-SHA384:ECDHE-RSA-AES256-GCM-SHA384')
return ssl_ctx
def set_rule(rule_json):
global rule
rule = json.loads(rule_json)
rules.append(rule)
rpc_server.server_close()
return rule_json
def RPC_server(rpc_port):
global rpc_server
# port = 0
rpc_server = SimpleXMLRPCServer(('localhost', rpc_port)) # 初始化
rpc_server.register_function(set_rule, "set_rule") # 注册函数
print("等待RPC规则配置......")
try:
rpc_server.handle_request() # 保持等待调用状态
print('配置完成......')
except OSError:
print('配置完成......')
return rule
async def handle_echo(reader, writer):
client_addr = writer.get_extra_info('peername')
connect_result = await server_authenticate(reader, writer, SECRET_KEY) # 用户合法性验证
if not connect_result:
print('客户端:' + str(client_addr) + '连接失败')
writer.close()
return
try:
login_result = await user_login(reader, writer)
if not login_result:
user_list.pop(str(client_addr[0]) + str(client_addr[1]))
print('已断开用户连接:', client_addr)
writer.close()
return
except ConnectionResetError:
user_list.pop(str(client_addr[0]) + str(client_addr[1]))
print('用户已断开连接:', client_addr)
writer.close()
return
except ssl.SSLError as e:
return
rpc_port = random.randint(49995,50000)
# await asyncio.sleep(0.01)
rpc_port = 12039
# print(rpc_port)
writer.write(str(rpc_port).encode())
await writer.drain()
loop = asyncio.get_running_loop()
with concurrent.futures.ProcessPoolExecutor() as pool:
config_rule = await loop.run_in_executor(pool, RPC_server, rpc_port)
# config_rule = RPC_server()
ident, src_ip, src_port, dst_ip, dst_port = \
config_rule['ident'], config_rule['src_ip'], config_rule['src_port'],config_rule['dst_ip'],config_rule['dst_port'],
# 此处设置要访问的默认服务器
dst_ip = Client_Ip[0]
dst_port = Client_Port[0]
l_reader, l_writer = await asyncio.open_connection(dst_ip, dst_port)
server_addr = l_writer.get_extra_info('peername')
server_addr_str = str(server_addr[0]) + str(server_addr[1])
hold_server_info(server_addr_str, server_addr, l_reader, l_writer) # 将处理目标客户端(服务器)的请求信息存起来
# print(server_list)
try:
# dest_ip = await reader.read(100) # 首先需要得到客户端的请求ip
# dest_ip = dest_ip.decode()
dest_ip = server_addr # 此处设置默认连接server_addr
print('正在请求:' + str(client_addr) + '请求目的ip:' + str(dest_ip))
find_dest = await connect_dest_server(dest_ip, reader, writer, l_reader, l_writer)
if find_dest:
config_rule['ident'] = 'hi'
config_rule_json = json.dumps(config_rule)
writer.write(config_rule_json.encode())
await writer.drain()
# print(config_rule)
s_reader = find_dest['Reader']
s_writer = find_dest['Writer']
while True:
data = await reader.read(100)
message = data.decode()
if message == 'Heart beat!':
writer.write(message.encode())
print('心跳响应' + message)
continue
if message == '' or message == 'exit':
message = 'Disconnect Request'
s_writer.write(message.encode())
await s_writer.drain()
re_msg = await s_reader.read(1024)
if re_msg.decode() == 'ByeBye':
writer.write(re_msg)
await writer.drain()
user_list.pop(str(client_addr[0]) + str(client_addr[1]))
print('用户已断开连接:', client_addr)
writer.close()
s_writer.close()
break
else:
pass
s_writer.write(data)
await s_writer.drain()
print(str(client_addr) + '正在给' + str(server_addr) + '发送信息:' + data.decode())
try:
re_msg = await s_reader.read(1024)
print('已收到' + str(server_addr) + '的回复:\n' + re_msg.decode())
try:
writer.write(re_msg)
await writer.drain()
print('成功给' + str(client_addr) + '发送回复:\n' + re_msg.decode())
except Exception as e:
print('5',e)
except Exception as e:
print('6',e)
else:
print("请求失败,连接已断开!")
except ConnectionResetError:
message = 'Force Disconnect'
l_writer.write(message.encode())
await l_writer.drain()
l_writer.close()
# user_list.pop(str(client_addr[0]) + str(client_addr[1]))
print('用户已断开连接:', client_addr)
writer.close()
except ssl.SSLError as e:
pass
async def get_general_control():
reader, writer = await asyncio.open_connection(Operat_server_IP[0], Operat_server_Port[0])
while True:
cmd = await reader.read(100)
cmd = cmd.decode()
if cmd == '1' or cmd == 'find -all':
user_info = []
for user in user_list:
user_info.append(user_list[user]['addr'])
re_cmd = transfer_json(user_info, True)
writer.write(re_cmd.encode())
await writer.drain()
elif cmd == '2' or cmd == 'break -all':
for user in user_list:
try:
user_writer = user_list[user]['Writer']
user_writer.write('exit'.encode())
await writer.drain()
user_writer.close()
print('已断开用户连接:', user_list[user]['addr'])
except Exception as e:
print('7',e)
user_list.clear()
writer.write('-----全部关闭成功-----'.encode())
await writer.drain()
else:
writer.write('-----命令有误!请重试-----'.encode())
await writer.drain()
async def main():
ssl_server = creat_server_ssl()
server = await asyncio.start_server(handle_echo, Server_Ip[0], Server_Port[0], ssl=ssl_server)
# server.socket返回内部的服务器套接字列表副本
addr = server.sockets[0].getsockname()
print('成功开启服务器:', addr)
print('等待客户端连接...\n')
try:
await get_general_control()
except Exception as e:
print('8',e)
async with server:
# 开始接受连接,直到协程被取消。 serve_forever 任务的取消将导致服务器被关闭。
await server.serve_forever()
def open_agent_server():
asyncio.run(main())
con, cur = create_sqlite_db()
if __name__ == '__main__':
open_agent_server()
| identifier_body |
||
TCP_echo_server.py | _connection.decode(), method=False)
if ensure_connection['code'] == 'Accept connection':
try:
connect_success = {'local_addr': local_addr, 'request_addr': dest_addr, 'code': 'Ready'}
connect_success = transfer_json(connect_success, method=True)
local_writer.write(connect_success.encode())
await local_writer.drain()
print('请求成功:' + str(local_addr) + '正在与' + str(dest_addr) + '通讯...\n')
dest = {'addr': dest_addr, 'Reader': dest_reader, 'Writer': dest_writer}
return dest
except ConnectionResetError:
# print('已断开用户连接:', local_addr)
return False
except Exception as e:
print('0',e)
elif ensure_connection['code'] == 'Refuse connection':
try:
connect_fail = {'local_addr': local_addr, 'request_addr': dest_addr, 'code': 'No'}
connect_fail = transfer_json(connect_fail, method=True)
local_writer.write(connect_fail.encode())
await local_writer.drain()
print('请求失败:' + str(dest_addr) + '拒绝与' + str(local_addr) + '通讯...\n')
dest_writer.close()
local_writer.close()
return False
except Exception as e:
print('1',e)
else:
pass
except Exception as e:
print('2',e)
pass
except Exception as e:
print('3',e)
pass
def hold_user_info(ip, addr, reader, writer):
"""存储已连接客户端的相关内容"""
user = {'addr': addr, 'Reader': reader, 'Writer': writer}
user_list[ip] = user
def hold_server_info(ip, addr, reader, writer):
"""存储已连接目标服务器(客户端)的相关内容"""
user = {'addr': addr, 'Reader': reader, 'Writer': writer}
server_list[ip] = user
async def server_authenticate(reader, writer, secret_key):
"""客户端合法认证"""
message = os.urandom(32) # 随机产生 n=32 个字节的字符串
writer.write(message)
await writer.drain()
s = hashlib.sha512()
s.update(message + secret_key.encode('utf-8')) # 加密
digest = s.hexdigest()
response = await reader.read(1024)
if digest == response.decode('utf-8'):
client_addr = writer.get_extra_info('peername')
client_addr_str = str(client_addr[0]) + str(client_addr[1]) # 拼接ip和port
hold_user_info(client_addr_str, client_addr, reader, writer)
print('\n客户端:' + str(client_addr) + '连接成功\n')
return digest
else:
writer.write('connection_error'.encode()) # 若连接失败,发送错误信息
writer.close()
async def user_login(reader, writer):
global search_result, account
try:
search_result = None
account = await reader.read(1024)
account = transfer_json(account.decode(), False)
sql = "select * from user where username = '{}' and password = '{}'".format(account['username'],
account['password'])
cur.execute(sql)
search_result = cur.fetchall()
except sqlite3.OperationalError:
search_result = False
except ssl.SSLError:
search_result = False
if search_result:
print('\n用户' + account['username'] + '登陆成功!\n')
writer.write('Login Success'.encode())
await writer.drain()
return True
else:
writer.write('Need Email'.encode())
await writer.drain()
email = await reader.read(1024)
verify_email = re.match(r'^[0-9a-zA-Z_]{0,19}@[0-9a-zA-Z]{1,13}\.[com,cn,net]{1,3}(.cn)?$', email.decode())
if verify_email:
email = verify_email.group()
sql = "insert into user(username,password,email) values ('{}','{}','{}')".format(
str(account['username']), str(account['password']), str(email))
try:
cur.execute(sql)
con.commit()
print('\n用户' + account['username'] + '注册成功!\n')
writer.write('Register Success'.encode())
await writer.drain()
return True
except Exception as e:
writer.write('Register Fail'.encode())
await writer.drain()
return False
else:
writer.write('Register Fail'.encode())
await writer.drain()
return False
def creat_server_ssl():
ssl_ctx = ssl.SSLContext(ssl.PROTOCOL_TLS_SERVER)
ssl_ctx.options |= ssl.OP_NO_TLSv1
ssl_ctx.options |= ssl.OP_NO_TLSv1_1
ssl_ctx.options |= ssl.OP_SINGLE_DH_USE
ssl_ctx.options |= ssl.OP_SINGLE_ECDH_USE
ssl_ctx.load_cert_chain(certfile='./server_ssl/mycertfile.pem', keyfile='./server_ssl/mykeyfile.pem')
ssl_ctx.load_verify_locations(cafile='./server_ssl/mycertfile.pem')
ssl_ctx.check_hostname = False
ssl_ctx.verify_mode = ssl.VerifyMode.CERT_REQUIRED
ssl_ctx.set_ciphers('ECDHE-ECDSA-AES256-GCM-SHA384:ECDHE-RSA-AES256-GCM-SHA384')
return ssl_ctx
def set_rule(rule_json):
global rule
rule = json.loads(rule_json)
rules.append(rule)
rpc_server.server_close()
return rule_json
def RPC_server(rpc_port):
global rpc_server
# port = 0
rpc_server = SimpleXMLRPCServer(('localhost', rpc_port)) # 初始化
rpc_server.register_function(set_rule, "set_rule") # 注册函数
print("等待RPC规则配置......")
try:
rpc_server.handle_request() # 保持等待调用状态
print('配置完成......')
except OSError:
print('配置完成......')
return rule
async def handle_echo(reader, writer):
client_addr = writer.get_extra_info('peername')
connect_result = await server_authenticate(reader, writer, SECRET_KEY) # 用户合法性验证
if not connect_result:
print('客户端:' + str(client_addr) + '连接失败')
writer.close()
return
try:
login_result = await user_login(reader, writer)
if not login_result:
user_list.pop(str(client_addr[0]) + str(client_addr[1]))
print('已断开用户连接:', client_addr)
writer.close()
return
except ConnectionResetError:
user_list.pop(str(client_addr[0]) + str(client_addr[1]))
print('用户已断开连接:', client_addr)
writer.close()
return
except ssl.SSLError as e:
return
rpc_port = random.randint(49995,50000)
# await asyncio.sleep(0.01)
rpc_port = 12039
# print(rpc_port)
writer.write(str(rpc_port).encode())
await writer.drain()
loop = asyncio.get_running_loop()
with concurrent.futures.ProcessPoolExecutor() as pool:
config_rule = await loop.run_in_executor(pool, RPC_server, rpc_port)
# config_rule = RPC_server()
ident, src_ip, src_port, dst_ip, dst_port = \
config_rule['ident'], config_rule['src_ip'], config_rule['src_port'],config_rule['dst_ip'],config_rule['dst_port'],
# 此处设置要访问的默认服务器
dst_ip = Client_Ip[0]
dst_port = Client_Port[0]
l_reader, l_writer = await asyncio.open_connection(dst_ip, dst_port)
server_addr = l_writer.get_extra_info('peername')
server_addr_str = str(server_addr[0]) + str(server_addr[1])
hold_server_info(server_addr_str, server_addr, l_reader, l_writer) # 将处理目标客户端(服务器)的请求信息存起来
# print(server_list)
try:
# dest_ip = await reader.read(100) # 首先需要得到客户端的请求ip
# dest_ip = dest_ip.decode()
dest_ip = server_addr # 此处设置默认连接server_addr |
print('正在请求:' + str(client_addr) + '请求目的ip:' + str(dest_ip))
find_dest = await connect_dest_server(dest_ip, reader, writer, l_reader, l_writer)
if find_dest:
config_rule['ident'] = 'hi'
config_rule_json = json.dumps(config_rule)
writer.write(config_rule_json.encode())
await writer.drain()
# print(config_rule)
s_reader = find_dest['Reader']
s_writer = find_dest['Writer']
while True:
data = await reader.read(100)
message = data.decode()
if message == 'Heart beat!':
writer.write(message.encode())
print('心跳响应' + message)
continue
if message == '' or message == 'exit':
message = 'Disconnect Request'
s_writer.write(message.encode())
await s_writer.drain()
re_msg = await s_reader.read(1024)
| random_line_split |
|
TCP_echo_server.py | .decode(), method=False)
if ensure_connection['code'] == 'Accept connection':
try:
connect_success = {'local_addr': local_addr, 'request_addr': dest_addr, 'code': 'Ready'}
connect_success = transfer_json(connect_success, method=True)
local_writer.write(connect_success.encode())
await local_writer.drain()
print('请求成功:' + str(local_addr) + '正在与' + str(dest_addr) + '通讯...\n')
dest = {'addr': dest_addr, 'Reader': dest_reader, 'Writer': dest_writer}
return dest
except ConnectionResetError:
# print('已断开用户连接:', local_addr)
return False
except Exception as e:
print('0',e)
elif ensure_connection['code'] == 'Refuse connection':
try:
connect_fail = {'local_addr': local_addr, 'request_addr': dest_addr, 'code': 'No'}
connect_fail = transfer_json(connect_fail, method=True)
local_writer.write(connect_fail.encode())
await local_writer.drain()
print('请求失败:' + str(dest_addr) + '拒绝与' + str(local_addr) + '通讯...\n')
dest_writer.close()
local_writer.close()
return False
except Exception as e:
print('1',e)
else:
pass
except Exception as e:
print('2',e)
pass
except Exception as e:
print('3',e)
pass
def hold_user_info(ip, addr, reader, writer):
"""存储已连接客户端的相关内容"""
user = {'addr': addr, 'Reader': reader, 'Writer': writer}
user_list[ip] = user
def hold_server_info(ip, addr, reader, writer):
"""存储已连接目标服务器(客户端)的相关内容"""
user = {'addr': addr, 'Reader': reader, 'Writer': writer}
server_list[ip] = user
async def server_authenticate(reader, writer, secret_key):
"""客户端合法认证"""
message = os.urandom(32) # 随机产生 n=32 个字节的字符串
writer.write(message)
await writer.drain()
s = hashlib.sha512()
s.update(message + secret_key.encode('utf-8')) # 加密
digest = s.hexdigest()
response = await reader.read(1024)
if digest == response.decode('utf-8'):
client_addr = writer.get_extra_info('peername')
client_addr_str = str(client_addr[0]) + str(client_addr[1]) # 拼接ip和port
hold_user_info(client_addr_str, client_addr, reader, writer)
print('\n客户端:' + str(client_addr) + '连接成功\n')
return digest
else:
writer.write('connection_error'.encode()) # 若连接失败,发送错误信息
writer.close()
async def user_login(reader, writer):
global search_result, account
try:
search_result = None
account = await reader.read(1024)
account = transfer_json(account.decode(), False)
sql = "select * from user where username = '{}' and password = '{}'".format(account['username'],
| account['password'])
cur.execute(sql)
search_result = cur.fetchall()
except sqlite3.OperationalError:
search_result = False
except ssl.SSLError:
search_result = False
if search_result:
print('\n用户' + account['username'] + '登陆成功!\n')
writer.write('Login Success'.encode())
await writer.drain()
return True
else:
writer.write('Need Email'.encode())
await writer.drain()
email = await reader.read(1024)
verify_email = re.match(r'^[0-9a-zA-Z_]{0,19}@[0-9a-zA-Z]{1,13}\.[com,cn,net]{1,3}(.cn)?$', email.decode())
if verify_email:
email = verify_email.group()
sql = "insert into user(username,password,email) values ('{}','{}','{}')".format(
str(account['username']), str(account['password']), str(email))
try:
cur.execute(sql)
con.commit()
print('\n用户' + account['username'] + '注册成功!\n')
writer.write('Register Success'.encode())
await writer.drain()
return True
except Exception as e:
writer.write('Register Fail'.encode())
await writer.drain()
return False
else:
writer.write('Register Fail'.encode())
await writer.drain()
return False
def creat_server_ssl():
ssl_ctx = ssl.SSLContext(ssl.PROTOCOL_TLS_SERVER)
ssl_ctx.options |= ssl.OP_NO_TLSv1
ssl_ctx.options |= ssl.OP_NO_TLSv1_1
ssl_ctx.options |= ssl.OP_SINGLE_DH_USE
ssl_ctx.options |= ssl.OP_SINGLE_ECDH_USE
ssl_ctx.load_cert_chain(certfile='./server_ssl/mycertfile.pem', keyfile='./server_ssl/mykeyfile.pem')
ssl_ctx.load_verify_locations(cafile='./server_ssl/mycertfile.pem')
ssl_ctx.check_hostname = False
ssl_ctx.verify_mode = ssl.VerifyMode.CERT_REQUIRED
ssl_ctx.set_ciphers('ECDHE-ECDSA-AES256-GCM-SHA384:ECDHE-RSA-AES256-GCM-SHA384')
return ssl_ctx
def set_rule(rule_json):
global rule
rule = json.loads(rule_json)
rules.append(rule)
rpc_server.server_close()
return rule_json
def RPC_server(rpc_port):
global rpc_server
# port = 0
rpc_server = SimpleXMLRPCServer(('localhost', rpc_port)) # 初始化
rpc_server.register_function(set_rule, "set_rule") # 注册函数
print("等待RPC规则配置......")
try:
rpc_server.handle_request() # 保持等待调用状态
print('配置完成......')
except OSError:
print('配置完成......')
return rule
async def handle_echo(reader, writer):
client_addr = writer.get_extra_info('peername')
connect_result = await server_authenticate(reader, writer, SECRET_KEY) # 用户合法性验证
if not connect_result:
print('客户端:' + str(client_addr) + '连接失败')
writer.close()
return
try:
login_result = await user_login(reader, writer)
if not login_result:
user_list.pop(str(client_addr[0]) + str(client_addr[1]))
print('已断开用户连接:', client_addr)
writer.close()
return
except ConnectionResetError:
user_list.pop(str(client_addr[0]) + str(client_addr[1]))
print('用户已断开连接:', client_addr)
writer.close()
return
except ssl.SSLError as e:
return
rpc_port = random.randint(49995,50000)
# await asyncio.sleep(0.01)
rpc_port = 12039
# print(rpc_port)
writer.write(str(rpc_port).encode())
await writer.drain()
loop = asyncio.get_running_loop()
with concurrent.futures.ProcessPoolExecutor() as pool:
config_rule = await loop.run_in_executor(pool, RPC_server, rpc_port)
# config_rule = RPC_server()
ident, src_ip, src_port, dst_ip, dst_port = \
config_rule['ident'], config_rule['src_ip'], config_rule['src_port'],config_rule['dst_ip'],config_rule['dst_port'],
# 此处设置要访问的默认服务器
dst_ip = Client_Ip[0]
dst_port = Client_Port[0]
l_reader, l_writer = await asyncio.open_connection(dst_ip, dst_port)
server_addr = l_writer.get_extra_info('peername')
server_addr_str = str(server_addr[0]) + str(server_addr[1])
hold_server_info(server_addr_str, server_addr, l_reader, l_writer) # 将处理目标客户端(服务器)的请求信息存起来
# print(server_list)
try:
# dest_ip = await reader.read(100) # 首先需要得到客户端的请求ip
# dest_ip = dest_ip.decode()
dest_ip = server_addr # 此处设置默认连接server_addr
print('正在请求:' + str(client_addr) + '请求目的ip:' + str(dest_ip))
find_dest = await connect_dest_server(dest_ip, reader, writer, l_reader, l_writer)
if find_dest:
config_rule['ident'] = 'hi'
config_rule_json = json.dumps(config_rule)
writer.write(config_rule_json.encode())
await writer.drain()
# print(config_rule)
s_reader = find_dest['Reader']
s_writer = find_dest['Writer']
while True:
data = await reader.read(100)
message = data.decode()
if message == 'Heart beat!':
writer.write(message.encode())
print('心跳响应' + message)
continue
if message == '' or message == 'exit':
message = 'Disconnect Request'
s_writer.write(message.encode())
await s_writer.drain()
re_msg = await s_reader.read(1024)
| identifier_name |
|
TCP_echo_server.py | (connect_success.encode())
await local_writer.drain()
print('请求成功:' + str(local_addr) + '正在与' + str(dest_addr) + '通讯...\n')
dest = {'addr': dest_addr, 'Reader': dest_reader, 'Writer': dest_writer}
return dest
except ConnectionResetError:
# print('已断开用户连接:', local_addr)
return False
except Exception as e:
print('0',e)
elif ensure_connection['code'] == 'Refuse connection':
try:
connect_fail = {'local_addr': local_addr, 'request_addr': dest_addr, 'code': 'No'}
connect_fail = transfer_json(connect_fail, method=True)
local_writer.write(connect_fail.encode())
await local_writer.drain()
print('请求失败:' + str(dest_addr) + '拒绝与' + str(local_addr) + '通讯...\n')
dest_writer.close()
local_writer.close()
return False
except Exception as e:
print('1',e)
else:
pass
except Exception as e:
print('2',e)
pass
except Exception as e:
print('3',e)
pass
def hold_user_info(ip, addr, reader, writer):
"""存储已连接客户端的相关内容"""
user = {'addr': addr, 'Reader': reader, 'Writer': writer}
user_list[ip] = user
def hold_server_info(ip, addr, reader, writer):
"""存储已连接目标服务器(客户端)的相关内容"""
user = {'addr': addr, 'Reader': reader, 'Writer': writer}
server_list[ip] = user
async def server_authenticate(reader, writer, secret_key):
"""客户端合法认证"""
message = os.urandom(32) # 随机产生 n=32 个字节的字符串
writer.write(message)
await writer.drain()
s = hashlib.sha512()
s.update(message + secret_key.encode('utf-8')) # 加密
digest = s.hexdigest()
response = await reader.read(1024)
if digest == response.decode('utf-8'):
client_addr = writer.get_extra_info('peername')
client_addr_str = str(client_addr[0]) + str(client_addr[1]) # 拼接ip和port
hold_user_info(client_addr_str, client_addr, reader, writer)
print('\n客户端:' + str(client_addr) + '连接成功\n')
return digest
else:
writer.write('connection_error'.encode()) # 若连接失败,发送错误信息
writer.close()
async def user_login(reader, writer):
global search_result, account
try:
search_result = None
account = await reader.read(1024)
account = transfer_json(account.decode(), False)
sql = "select * from user where username = '{}' and password = '{}'".format(account['username'],
account['password'])
cur.execute(sql)
search_result = cur.fetchall()
except sqlite3.OperationalError:
search_result = False
except ssl.SSLError:
search_result = False
if search_result:
print('\n用户' + account['username'] + '登陆成功!\n')
writer.write('Login Success'.encode())
await writer.drain()
return True
else:
writer.write('Need Email'.encode())
await writer.drain()
email = await reader.read(1024)
verify_email = re.match(r'^[0-9a-zA-Z_]{0,19}@[0-9a-zA-Z]{1,13}\.[com,cn,net]{1,3}(.cn)?$', email.decode())
if verify_email:
email = verify_email.group()
sql = "insert into user(username,password,email) values ('{}','{}','{}')".format(
str(account['username']), str(account['password']), str(email))
try:
cur.execute(sql)
con.commit()
print('\n用户' + account['username'] + '注册成功!\n')
writer.write('Register Success'.encode())
await writer.drain()
return True
except Exception as e:
writer.write('Register Fail'.encode())
await writer.drain()
return False
else:
writer.write('Register Fail'.encode())
await writer.drain()
return False
def creat_server_ssl():
ssl_ctx = ssl.SSLContext(ssl.PROTOCOL_TLS_SERVER)
ssl_ctx.options |= ssl.OP_NO_TLSv1
ssl_ctx.options |= ssl.OP_NO_TLSv1_1
ssl_ctx.options |= ssl.OP_SINGLE_DH_USE
ssl_ctx.options |= ssl.OP_SINGLE_ECDH_USE
ssl_ctx.load_cert_chain(certfile='./server_ssl/mycertfile.pem', keyfile='./server_ssl/mykeyfile.pem')
ssl_ctx.load_verify_locations(cafile='./server_ssl/mycertfile.pem')
ssl_ctx.check_hostname = False
ssl_ctx.verify_mode = ssl.VerifyMode.CERT_REQUIRED
ssl_ctx.set_ciphers('ECDHE-ECDSA-AES256-GCM-SHA384:ECDHE-RSA-AES256-GCM-SHA384')
return ssl_ctx
def set_rule(rule_json):
global rule
rule = json.loads(rule_json)
rules.append(rule)
rpc_server.server_close()
return rule_json
def RPC_server(rpc_port):
global rpc_server
# port = 0
rpc_server = SimpleXMLRPCServer(('localhost', rpc_port)) # 初始化
rpc_server.register_function(set_rule, "set_rule") # 注册函数
print("等待RPC规则配置......")
try:
rpc_server.handle_request() # 保持等待调用状态
print('配置完成......')
except OSError:
print('配置完成......')
return rule
async def handle_echo(reader, writer):
client_addr = writer.get_extra_info('peername')
connect_result = await server_authenticate(reader, writer, SECRET_KEY) # 用户合法性验证
if not connect_result:
print('客户端:' + str(client_addr) + '连接失败')
writer.close()
return
try:
login_result = await user_login(reader, writer)
if not login_result:
user_list.pop(str(client_addr[0]) + str(client_addr[1]))
print('已断开用户连接:', client_addr)
writer.close()
return
except ConnectionResetError:
user_list.pop(str(client_addr[0]) + str(client_addr[1]))
print('用户已断开连接:', client_addr)
writer.close()
return
except ssl.SSLError as e:
return
rpc_port = random.randint(49995,50000)
# await asyncio.sleep(0.01)
rpc_port = 12039
# print(rpc_port)
writer.write(str(rpc_port).encode())
await writer.drain()
loop = asyncio.get_running_loop()
with concurrent.futures.ProcessPoolExecutor() as pool:
config_rule = await loop.run_in_executor(pool, RPC_server, rpc_port)
# config_rule = RPC_server()
ident, src_ip, src_port, dst_ip, dst_port = \
config_rule['ident'], config_rule['src_ip'], config_rule['src_port'],config_rule['dst_ip'],config_rule['dst_port'],
# 此处设置要访问的默认服务器
dst_ip = Client_Ip[0]
dst_port = Client_Port[0]
l_reader, l_writer = await asyncio.open_connection(dst_ip, dst_port)
server_addr = l_writer.get_extra_info('peername')
server_addr_str = str(server_addr[0]) + str(server_addr[1])
hold_server_info(server_addr_str, server_addr, l_reader, l_writer) # 将处理目标客户端(服务器)的请求信息存起来
# print(server_list)
try:
# dest_ip = await reader.read(100) # 首先需要得到客户端的请求ip
# dest_ip = dest_ip.decode()
dest_ip = server_addr # 此处设置默认连接server_addr
print('正在请求:' + str(client_addr) + '请求目的ip:' + str(dest_ip))
find_dest = await connect_dest_server(dest_ip, reader, writer, l_reader, l_writer)
if find_dest:
config_rule['ident'] = 'hi'
config_rule_json = json.dumps(config_rule)
writer.write(config_rule_json.encode())
await writer.drain()
# print(config_rule)
s_reader = find_dest['Reader']
s_writer = find_dest['Writer']
while True:
data = await reader.read(100)
message = data.decode()
if message == 'Heart beat!':
writer.write(message.encode())
print('心跳响应' + message)
continue
if message == '' or message == 'exit':
message = 'Disconnect Request'
s_writer.write(message.encode())
await s_writer.drain()
re_msg = await s_reader.read(1024)
if re_msg.decode() == 'ByeBye':
writer.write(re_msg)
await writer.drain()
user_list.pop(str(client_addr[0]) + str(client_addr[1]))
| print('用户已断开连接:', client_addr)
writer.close()
s_w | conditional_block |
|
localization.rs | .clone()
} else {
let string = fs::read_to_string(&path).unwrap_or_else(|_| {
if (res_id, locale) == ("builtin.ftl", "en-US") {
FALLBACK_STRINGS.to_string()
} else {
error!("missing resouce {}/{}", locale, res_id);
String::new()
}
});
let res = match FluentResource::try_new(string) {
Ok(res) => Arc::new(res),
Err((res, _err)) => Arc::new(res),
};
self.resources.insert(path, res.clone());
res
}
}
/// Return the best localization bundle for the provided `LanguageIdentifier`.
fn get_bundle(&mut self, locale: &LanguageIdentifier, resource_ids: &[String]) -> BundleStack {
let resolved_locales = self.resolve_locales(locale.clone());
debug!("resolved: {}", PrintLocales(resolved_locales.as_slice()));
let mut stack = Vec::new();
for locale in &resolved_locales {
let mut bundle = FluentBundle::new(&resolved_locales);
for res_id in resource_ids {
let res = self.get_resource(&res_id, &locale.to_string());
bundle.add_resource(res).unwrap();
}
stack.push(bundle);
}
BundleStack(stack)
}
/// Given a locale, returns the best set of available locales.
pub(crate) fn resolve_locales(&self, locale: LanguageIdentifier) -> Vec<LanguageIdentifier> {
negotiate_languages(
&[locale],
&self.locales,
Some(&self.default_locale),
NegotiationStrategy::Filtering,
)
.into_iter()
.map(|l| l.to_owned())
.collect()
}
}
impl L10nManager {
/// Create a new localization manager.
///
/// `resources` is a list of file names that contain strings. `base_dir`
/// is a path to a directory that includes per-locale subdirectories.
///
/// This directory should be of the structure `base_dir/{locale}/{resource}`,
/// where '{locale}' is a valid BCP47 language tag, and {resource} is a `.ftl`
/// included in `resources`.
pub fn new(resources: Vec<String>, base_dir: &str) -> Self {
fn get_available_locales(base_dir: &str) -> Result<Vec<LanguageIdentifier>, io::Error> {
let mut locales = vec![];
let res_dir = fs::read_dir(base_dir)?;
for entry in res_dir {
if let Ok(entry) = entry {
let path = entry.path();
if path.is_dir() {
if let Some(name) = path.file_name() {
if let Some(name) = name.to_str() {
let langid: LanguageIdentifier =
name.parse().expect("Parsing failed.");
locales.push(langid);
}
}
}
}
}
Ok(locales)
}
let default_locale: LanguageIdentifier =
"en-US".parse().expect("failed to parse default locale");
let current_locale = Application::get_locale()
.parse()
.unwrap_or_else(|_| default_locale.clone());
let locales = get_available_locales(base_dir).unwrap_or_default();
debug!(
"available locales {}, current {}",
PrintLocales(&locales),
current_locale,
);
let mut path_scheme = base_dir.to_string();
path_scheme.push_str("/{locale}/{res_id}");
let mut res_mgr = ResourceManager {
resources: HashMap::new(),
path_scheme,
default_locale,
locales,
};
let current_bundle = res_mgr.get_bundle(¤t_locale, &resources);
L10nManager {
res_mgr,
current_bundle,
resources,
current_locale,
}
}
/// Fetch a localized string from the current bundle by key.
///
/// In general, this should not be used directly; [`LocalizedString`]
/// should be used for localization, and you should call
/// [`LocalizedString::resolve`] to update the string as required.
///
///[`LocalizedString`]: struct.LocalizedString.html
///[`LocalizedString::resolve`]: struct.LocalizedString.html#method.resolve
pub fn localize<'args>(
&'args self,
key: &str,
args: impl Into<Option<&'args FluentArgs<'args>>>,
) -> Option<String> {
let args = args.into();
let value = match self
.current_bundle
.get_message(key)
.and_then(|msg| msg.value)
{
Some(v) => v,
None => return None,
};
let mut errs = Vec::new();
let result = self
.current_bundle
.format_pattern(key, value, args, &mut errs);
for err in errs {
warn!("localization error {:?}", err);
}
// fluent inserts bidi controls when interpolating, and they can
// cause rendering issues; for now we just strip them.
// https://www.w3.org/International/questions/qa-bidi-unicode-controls#basedirection
const START_ISOLATE: char = '\u{2068}';
const END_ISOLATE: char = '\u{2069}';
if args.is_some() && result.chars().any(|c| c == START_ISOLATE) {
Some(
result
.chars()
.filter(|c| c != &START_ISOLATE && c != &END_ISOLATE)
.collect(),
)
} else {
Some(result)
}
}
//TODO: handle locale change
}
*/ ////
impl<T> LocalizedString<T> {
/// Create a new `LocalizedString` with the given key.
pub const fn new(key: &'static str) -> Self {
LocalizedString {
key,
args: None,
placeholder: None,
resolved: None,
////resolved_lang: None,
}
}
/* ////
/// Add a placeholder value. This will be used if localization fails.
///
/// This is intended for use during prototyping.
pub fn with_placeholder(mut self, placeholder: String) -> Self {
self.placeholder = Some(placeholder);
self
}
*/ ////
/// Return the localized value for this string, or the placeholder, if
/// the localization is missing, or the key if there is no placeholder.
pub fn localized_str(&self) -> &str {
//cortex_m::asm::bkpt(); ////
self.resolved ////
.as_ref()
.expect("not resolved")
.as_str()
/* ////
self.resolved
.as_ref()
.map(|s| s.as_str())
.or_else(|| self.placeholder.as_ref().map(String::as_ref))
.unwrap_or(self.key)
*/ ////
}
}
impl<T: Data> LocalizedString<T> {
/// Add a named argument and a corresponding [`ArgClosure`]. This closure
/// is a function that will return a value for the given key from the current
/// environment and data.
///
/// [`ArgClosure`]: type.ArgClosure.html
pub fn with_arg(
mut self,
key: &'static str,
f: fn(&T, &Env) -> ArgValue, ////
////f: impl Fn(&T, &Env) -> FluentValue<'static> + 'static,
) -> Self {
self.args
.get_or_insert(Vec::new())
.push((key, ArgSource(f)))
.expect("with arg failed"); ////
////.push((key, ArgSource(Arc::new(f))));
self
}
/// Lazily compute the localized value for this string based on the provided
/// environment and data.
///
/// Returns `true` if the current value of the string has changed.
pub fn resolve<'a>(&'a mut self, data: &T, env: &Env) -> bool {
//TODO: this recomputes the string if either the language has changed,
//or *anytime* we have arguments. Ideally we would be using a lens
//to only recompute when our actual data has changed.
if self.args.is_some()
////|| self.resolved_lang.as_ref() != Some(&env.localization_manager().current_locale)
{
//// Resolve all args
let mut args = ArgValues::new();
for arg in self.args.as_ref().expect("resolve fail") {
let (k, v) = arg;
let argvalue = (v.0)(data, env);
args.insert(k, argvalue.clone())
.expect("resolve fail");
//// Convert the first arg to text and exit
self.resolved = Some(argvalue.to_string());
//cortex_m::asm::bkpt(); ////
return true;
}
//// No args to resolve
false
/* ////
let args: Option<FluentArgs> = self
.args
.as_ref()
.map(|a| a.iter().map(|(k, v)| (*k, (v.0)(data, env))).collect());
self.resolved_lang = Some(env.localization_manager().current_locale.clone());
let next = env.localization_manager().localize(self.key, args.as_ref());
let result = next != self.resolved;
self.resolved = next;
result
*/ ////
} else | {
false
} | conditional_block |
|
localization.rs | on things like `Menu`.
/////// A closure that generates a localization value.
type ArgClosure<T> = fn(&T, &Env) -> ArgValue; ////
////type ArgClosure<T> = Arc<dyn Fn(&T, &Env) -> FluentValue<'static> + 'static>;
/// Wraps a closure that generates an argument for localization.
#[derive(Clone)]
struct ArgSource<T>(ArgClosure<T>);
/// A string that can be localized based on the current locale.
///
/// At its simplest, a `LocalizedString` is a key that can be resolved
/// against a map of localized strings for a given locale.
#[derive(Clone)]////
////#[derive(Debug, Clone)]
pub struct LocalizedString<T> {
pub(crate) key: &'static str,
placeholder: Option<String>,
args: Option<Vec<(&'static str, ArgSource<T>)>>,
resolved: Option<String>,
////resolved_lang: Option<LanguageIdentifier>,
}
/* ////
/// A stack of localization resources, used for fallback.
struct BundleStack(Vec<FluentBundle<Arc<FluentResource>>>);
impl BundleStack {
fn get_message(&self, id: &str) -> Option<FluentMessage> {
self.0.iter().flat_map(|b| b.get_message(id)).next()
}
fn format_pattern(
&self,
id: &str,
pattern: &FluentPattern,
args: Option<&FluentArgs>,
errors: &mut Vec<FluentError>,
) -> String {
for bundle in self.0.iter() {
if bundle.has_message(id) {
return bundle.format_pattern(pattern, args, errors).to_string();
}
}
format!("localization failed for key '{}'", id)
}
}
//NOTE: much of this is adapted from https://github.com/projectfluent/fluent-rs/blob/master/fluent-resmgr/src/resource_manager.rs
impl ResourceManager {
/// Loads a new localization resource from disk, as needed.
fn get_resource(&mut self, res_id: &str, locale: &str) -> Arc<FluentResource> {
let path = self
.path_scheme
.replace("{locale}", locale)
.replace("{res_id}", res_id);
if let Some(res) = self.resources.get(&path) {
res.clone()
} else {
let string = fs::read_to_string(&path).unwrap_or_else(|_| {
if (res_id, locale) == ("builtin.ftl", "en-US") {
FALLBACK_STRINGS.to_string()
} else {
error!("missing resouce {}/{}", locale, res_id);
String::new()
}
});
let res = match FluentResource::try_new(string) {
Ok(res) => Arc::new(res),
Err((res, _err)) => Arc::new(res),
};
self.resources.insert(path, res.clone());
res
}
}
/// Return the best localization bundle for the provided `LanguageIdentifier`.
fn get_bundle(&mut self, locale: &LanguageIdentifier, resource_ids: &[String]) -> BundleStack {
let resolved_locales = self.resolve_locales(locale.clone());
debug!("resolved: {}", PrintLocales(resolved_locales.as_slice()));
let mut stack = Vec::new();
for locale in &resolved_locales {
let mut bundle = FluentBundle::new(&resolved_locales);
for res_id in resource_ids {
let res = self.get_resource(&res_id, &locale.to_string());
bundle.add_resource(res).unwrap();
}
stack.push(bundle);
}
BundleStack(stack)
}
/// Given a locale, returns the best set of available locales.
pub(crate) fn resolve_locales(&self, locale: LanguageIdentifier) -> Vec<LanguageIdentifier> {
negotiate_languages(
&[locale],
&self.locales,
Some(&self.default_locale),
NegotiationStrategy::Filtering,
)
.into_iter()
.map(|l| l.to_owned())
.collect()
}
}
impl L10nManager {
/// Create a new localization manager.
///
/// `resources` is a list of file names that contain strings. `base_dir`
/// is a path to a directory that includes per-locale subdirectories.
///
/// This directory should be of the structure `base_dir/{locale}/{resource}`,
/// where '{locale}' is a valid BCP47 language tag, and {resource} is a `.ftl`
/// included in `resources`.
pub fn new(resources: Vec<String>, base_dir: &str) -> Self {
fn get_available_locales(base_dir: &str) -> Result<Vec<LanguageIdentifier>, io::Error> {
let mut locales = vec![];
let res_dir = fs::read_dir(base_dir)?;
for entry in res_dir {
if let Ok(entry) = entry {
let path = entry.path();
if path.is_dir() {
if let Some(name) = path.file_name() {
if let Some(name) = name.to_str() {
let langid: LanguageIdentifier =
name.parse().expect("Parsing failed.");
locales.push(langid);
}
}
}
}
}
Ok(locales)
}
let default_locale: LanguageIdentifier =
"en-US".parse().expect("failed to parse default locale");
let current_locale = Application::get_locale()
.parse()
.unwrap_or_else(|_| default_locale.clone());
let locales = get_available_locales(base_dir).unwrap_or_default();
debug!(
"available locales {}, current {}",
PrintLocales(&locales),
current_locale,
);
let mut path_scheme = base_dir.to_string();
path_scheme.push_str("/{locale}/{res_id}");
let mut res_mgr = ResourceManager {
resources: HashMap::new(),
path_scheme,
default_locale,
locales,
};
let current_bundle = res_mgr.get_bundle(¤t_locale, &resources);
L10nManager {
res_mgr,
current_bundle,
resources,
current_locale,
}
}
/// Fetch a localized string from the current bundle by key.
///
/// In general, this should not be used directly; [`LocalizedString`]
/// should be used for localization, and you should call
/// [`LocalizedString::resolve`] to update the string as required.
///
///[`LocalizedString`]: struct.LocalizedString.html
///[`LocalizedString::resolve`]: struct.LocalizedString.html#method.resolve
pub fn localize<'args>(
&'args self,
key: &str,
args: impl Into<Option<&'args FluentArgs<'args>>>,
) -> Option<String> {
let args = args.into();
let value = match self
.current_bundle
.get_message(key)
.and_then(|msg| msg.value)
{
Some(v) => v,
None => return None,
};
let mut errs = Vec::new();
let result = self
.current_bundle
.format_pattern(key, value, args, &mut errs);
for err in errs {
warn!("localization error {:?}", err);
}
// fluent inserts bidi controls when interpolating, and they can
// cause rendering issues; for now we just strip them.
// https://www.w3.org/International/questions/qa-bidi-unicode-controls#basedirection
const START_ISOLATE: char = '\u{2068}';
const END_ISOLATE: char = '\u{2069}';
if args.is_some() && result.chars().any(|c| c == START_ISOLATE) {
Some(
result
.chars()
.filter(|c| c != &START_ISOLATE && c != &END_ISOLATE)
.collect(),
)
} else {
Some(result)
}
}
//TODO: handle locale change
}
*/ ////
impl<T> LocalizedString<T> {
/// Create a new `LocalizedString` with the given key.
pub const fn new(key: &'static str) -> Self {
LocalizedString {
key,
args: None,
placeholder: None,
resolved: None,
////resolved_lang: None,
}
}
/* ////
/// Add a placeholder value. This will be used if localization fails.
///
/// This is intended for use during prototyping.
pub fn with_placeholder(mut self, placeholder: String) -> Self {
self.placeholder = Some(placeholder);
self
}
*/ ////
/// Return the localized value for this string, or the placeholder, if
/// the localization is missing, or the key if there is no placeholder.
pub fn localized_str(&self) -> &str {
//cortex_m::asm::bkpt(); ////
self.resolved ////
.as_ref()
.expect("not resolved")
.as_str()
/* ////
self.resolved
.as_ref()
.map(|s| s.as_str())
.or_else(|| self.placeholder.as_ref().map(String::as_ref))
.unwrap_or(self.key)
*/ ////
}
}
impl<T: Data> LocalizedString<T> {
/// Add a named argument and a corresponding [`ArgClosure`]. This closure
/// is a function that will return a value for the given key from the current
/// environment and data.
///
/// [`ArgClosure`]: type.ArgClosure.html
pub fn | with_arg | identifier_name |
|
localization.rs | Arg = heapless::consts::U2; //// Max number of localized args
type Vec<T> = heapless::Vec::<T, MaxLocalizedArg>; ///
//////NOTE: instead of a closure, at some point we can use something like a lens for this.
//////TODO: this is an Arc so that it can be clone, which is a bound on things like `Menu`.
/////// A closure that generates a localization value.
type ArgClosure<T> = fn(&T, &Env) -> ArgValue; ////
////type ArgClosure<T> = Arc<dyn Fn(&T, &Env) -> FluentValue<'static> + 'static>;
/// Wraps a closure that generates an argument for localization.
#[derive(Clone)]
struct ArgSource<T>(ArgClosure<T>);
/// A string that can be localized based on the current locale.
///
/// At its simplest, a `LocalizedString` is a key that can be resolved
/// against a map of localized strings for a given locale.
#[derive(Clone)]////
////#[derive(Debug, Clone)]
pub struct LocalizedString<T> {
pub(crate) key: &'static str,
placeholder: Option<String>,
args: Option<Vec<(&'static str, ArgSource<T>)>>,
resolved: Option<String>,
////resolved_lang: Option<LanguageIdentifier>,
}
/* ////
/// A stack of localization resources, used for fallback.
struct BundleStack(Vec<FluentBundle<Arc<FluentResource>>>);
impl BundleStack {
fn get_message(&self, id: &str) -> Option<FluentMessage> {
self.0.iter().flat_map(|b| b.get_message(id)).next()
}
fn format_pattern(
&self,
id: &str,
pattern: &FluentPattern,
args: Option<&FluentArgs>,
errors: &mut Vec<FluentError>,
) -> String {
for bundle in self.0.iter() {
if bundle.has_message(id) {
return bundle.format_pattern(pattern, args, errors).to_string();
}
}
format!("localization failed for key '{}'", id)
}
}
//NOTE: much of this is adapted from https://github.com/projectfluent/fluent-rs/blob/master/fluent-resmgr/src/resource_manager.rs
impl ResourceManager {
/// Loads a new localization resource from disk, as needed.
fn get_resource(&mut self, res_id: &str, locale: &str) -> Arc<FluentResource> {
let path = self
.path_scheme
.replace("{locale}", locale)
.replace("{res_id}", res_id);
if let Some(res) = self.resources.get(&path) {
res.clone()
} else {
let string = fs::read_to_string(&path).unwrap_or_else(|_| {
if (res_id, locale) == ("builtin.ftl", "en-US") {
FALLBACK_STRINGS.to_string()
} else {
error!("missing resouce {}/{}", locale, res_id);
String::new()
}
});
let res = match FluentResource::try_new(string) {
Ok(res) => Arc::new(res),
Err((res, _err)) => Arc::new(res),
};
self.resources.insert(path, res.clone());
res
}
}
/// Return the best localization bundle for the provided `LanguageIdentifier`.
fn get_bundle(&mut self, locale: &LanguageIdentifier, resource_ids: &[String]) -> BundleStack {
let resolved_locales = self.resolve_locales(locale.clone());
debug!("resolved: {}", PrintLocales(resolved_locales.as_slice()));
let mut stack = Vec::new();
for locale in &resolved_locales {
let mut bundle = FluentBundle::new(&resolved_locales);
for res_id in resource_ids {
let res = self.get_resource(&res_id, &locale.to_string());
bundle.add_resource(res).unwrap();
}
stack.push(bundle);
}
BundleStack(stack)
}
/// Given a locale, returns the best set of available locales.
pub(crate) fn resolve_locales(&self, locale: LanguageIdentifier) -> Vec<LanguageIdentifier> {
negotiate_languages(
&[locale],
&self.locales,
Some(&self.default_locale),
NegotiationStrategy::Filtering,
)
.into_iter()
.map(|l| l.to_owned())
.collect()
}
}
impl L10nManager {
/// Create a new localization manager.
///
/// `resources` is a list of file names that contain strings. `base_dir`
/// is a path to a directory that includes per-locale subdirectories.
///
/// This directory should be of the structure `base_dir/{locale}/{resource}`,
/// where '{locale}' is a valid BCP47 language tag, and {resource} is a `.ftl`
/// included in `resources`.
pub fn new(resources: Vec<String>, base_dir: &str) -> Self {
fn get_available_locales(base_dir: &str) -> Result<Vec<LanguageIdentifier>, io::Error> {
let mut locales = vec![];
let res_dir = fs::read_dir(base_dir)?;
for entry in res_dir {
if let Ok(entry) = entry {
let path = entry.path();
if path.is_dir() {
if let Some(name) = path.file_name() {
if let Some(name) = name.to_str() {
let langid: LanguageIdentifier =
name.parse().expect("Parsing failed.");
locales.push(langid);
}
}
}
}
}
Ok(locales)
}
let default_locale: LanguageIdentifier =
"en-US".parse().expect("failed to parse default locale");
let current_locale = Application::get_locale()
.parse()
.unwrap_or_else(|_| default_locale.clone());
let locales = get_available_locales(base_dir).unwrap_or_default();
debug!(
"available locales {}, current {}",
PrintLocales(&locales),
current_locale,
);
let mut path_scheme = base_dir.to_string();
path_scheme.push_str("/{locale}/{res_id}");
let mut res_mgr = ResourceManager {
resources: HashMap::new(),
path_scheme,
default_locale,
locales,
};
let current_bundle = res_mgr.get_bundle(¤t_locale, &resources);
L10nManager {
res_mgr,
current_bundle,
resources,
current_locale,
}
}
/// Fetch a localized string from the current bundle by key.
///
/// In general, this should not be used directly; [`LocalizedString`]
/// should be used for localization, and you should call
/// [`LocalizedString::resolve`] to update the string as required.
///
///[`LocalizedString`]: struct.LocalizedString.html
///[`LocalizedString::resolve`]: struct.LocalizedString.html#method.resolve
pub fn localize<'args>(
&'args self,
key: &str,
args: impl Into<Option<&'args FluentArgs<'args>>>,
) -> Option<String> {
let args = args.into();
let value = match self
.current_bundle
.get_message(key)
.and_then(|msg| msg.value)
{
Some(v) => v,
None => return None,
};
let mut errs = Vec::new();
let result = self
.current_bundle
.format_pattern(key, value, args, &mut errs);
for err in errs {
warn!("localization error {:?}", err);
}
// fluent inserts bidi controls when interpolating, and they can
// cause rendering issues; for now we just strip them.
// https://www.w3.org/International/questions/qa-bidi-unicode-controls#basedirection
const START_ISOLATE: char = '\u{2068}';
const END_ISOLATE: char = '\u{2069}';
if args.is_some() && result.chars().any(|c| c == START_ISOLATE) {
Some(
result
.chars()
.filter(|c| c != &START_ISOLATE && c != &END_ISOLATE)
.collect(),
)
} else {
Some(result)
}
}
//TODO: handle locale change
}
*/ ////
impl<T> LocalizedString<T> {
/// Create a new `LocalizedString` with the given key.
pub const fn new(key: &'static str) -> Self {
LocalizedString {
key,
args: None,
placeholder: None,
resolved: None,
////resolved_lang: None,
}
}
/* ////
/// Add a placeholder value. This will be used if localization fails.
///
/// This is intended for use during prototyping.
pub fn with_placeholder(mut self, placeholder: String) -> Self {
self.placeholder = Some(placeholder);
self
}
*/ ////
/// Return the localized value for this string, or the placeholder, if
/// the localization is missing, or the key if there is no placeholder.
pub fn localized_str(&self) -> &str {
//cortex_m::asm::bkpt(); ////
self.resolved ////
.as_ref()
.expect("not resolved")
.as_str()
/* ////
self.resolved
.as_ref()
.map(|s| s.as_str())
.or_else(|| self.placeholder.as_ref().map(String::as_ref))
.unwrap_or(self.key) | */ //// | random_line_split |
|
localization.rs | "en-US".parse().expect("failed to parse default locale");
let current_locale = Application::get_locale()
.parse()
.unwrap_or_else(|_| default_locale.clone());
let locales = get_available_locales(base_dir).unwrap_or_default();
debug!(
"available locales {}, current {}",
PrintLocales(&locales),
current_locale,
);
let mut path_scheme = base_dir.to_string();
path_scheme.push_str("/{locale}/{res_id}");
let mut res_mgr = ResourceManager {
resources: HashMap::new(),
path_scheme,
default_locale,
locales,
};
let current_bundle = res_mgr.get_bundle(¤t_locale, &resources);
L10nManager {
res_mgr,
current_bundle,
resources,
current_locale,
}
}
/// Fetch a localized string from the current bundle by key.
///
/// In general, this should not be used directly; [`LocalizedString`]
/// should be used for localization, and you should call
/// [`LocalizedString::resolve`] to update the string as required.
///
///[`LocalizedString`]: struct.LocalizedString.html
///[`LocalizedString::resolve`]: struct.LocalizedString.html#method.resolve
pub fn localize<'args>(
&'args self,
key: &str,
args: impl Into<Option<&'args FluentArgs<'args>>>,
) -> Option<String> {
let args = args.into();
let value = match self
.current_bundle
.get_message(key)
.and_then(|msg| msg.value)
{
Some(v) => v,
None => return None,
};
let mut errs = Vec::new();
let result = self
.current_bundle
.format_pattern(key, value, args, &mut errs);
for err in errs {
warn!("localization error {:?}", err);
}
// fluent inserts bidi controls when interpolating, and they can
// cause rendering issues; for now we just strip them.
// https://www.w3.org/International/questions/qa-bidi-unicode-controls#basedirection
const START_ISOLATE: char = '\u{2068}';
const END_ISOLATE: char = '\u{2069}';
if args.is_some() && result.chars().any(|c| c == START_ISOLATE) {
Some(
result
.chars()
.filter(|c| c != &START_ISOLATE && c != &END_ISOLATE)
.collect(),
)
} else {
Some(result)
}
}
//TODO: handle locale change
}
*/ ////
impl<T> LocalizedString<T> {
/// Create a new `LocalizedString` with the given key.
pub const fn new(key: &'static str) -> Self {
LocalizedString {
key,
args: None,
placeholder: None,
resolved: None,
////resolved_lang: None,
}
}
/* ////
/// Add a placeholder value. This will be used if localization fails.
///
/// This is intended for use during prototyping.
pub fn with_placeholder(mut self, placeholder: String) -> Self {
self.placeholder = Some(placeholder);
self
}
*/ ////
/// Return the localized value for this string, or the placeholder, if
/// the localization is missing, or the key if there is no placeholder.
pub fn localized_str(&self) -> &str {
//cortex_m::asm::bkpt(); ////
self.resolved ////
.as_ref()
.expect("not resolved")
.as_str()
/* ////
self.resolved
.as_ref()
.map(|s| s.as_str())
.or_else(|| self.placeholder.as_ref().map(String::as_ref))
.unwrap_or(self.key)
*/ ////
}
}
impl<T: Data> LocalizedString<T> {
/// Add a named argument and a corresponding [`ArgClosure`]. This closure
/// is a function that will return a value for the given key from the current
/// environment and data.
///
/// [`ArgClosure`]: type.ArgClosure.html
pub fn with_arg(
mut self,
key: &'static str,
f: fn(&T, &Env) -> ArgValue, ////
////f: impl Fn(&T, &Env) -> FluentValue<'static> + 'static,
) -> Self {
self.args
.get_or_insert(Vec::new())
.push((key, ArgSource(f)))
.expect("with arg failed"); ////
////.push((key, ArgSource(Arc::new(f))));
self
}
/// Lazily compute the localized value for this string based on the provided
/// environment and data.
///
/// Returns `true` if the current value of the string has changed.
pub fn resolve<'a>(&'a mut self, data: &T, env: &Env) -> bool {
//TODO: this recomputes the string if either the language has changed,
//or *anytime* we have arguments. Ideally we would be using a lens
//to only recompute when our actual data has changed.
if self.args.is_some()
////|| self.resolved_lang.as_ref() != Some(&env.localization_manager().current_locale)
{
//// Resolve all args
let mut args = ArgValues::new();
for arg in self.args.as_ref().expect("resolve fail") {
let (k, v) = arg;
let argvalue = (v.0)(data, env);
args.insert(k, argvalue.clone())
.expect("resolve fail");
//// Convert the first arg to text and exit
self.resolved = Some(argvalue.to_string());
//cortex_m::asm::bkpt(); ////
return true;
}
//// No args to resolve
false
/* ////
let args: Option<FluentArgs> = self
.args
.as_ref()
.map(|a| a.iter().map(|(k, v)| (*k, (v.0)(data, env))).collect());
self.resolved_lang = Some(env.localization_manager().current_locale.clone());
let next = env.localization_manager().localize(self.key, args.as_ref());
let result = next != self.resolved;
self.resolved = next;
result
*/ ////
} else {
false
}
}
}
/* ////
impl<T> std::fmt::Debug for ArgSource<T> {
fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result {
write!(f, "Arg Resolver {:p}", self.0)
}
}
/// Helper to impl display for slices of displayable things.
struct PrintLocales<'a, T>(&'a [T]);
impl<'a, T: std::fmt::Display> std::fmt::Display for PrintLocales<'a, T> {
fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result {
write!(f, "[")?;
let mut prev = false;
for l in self.0 {
if prev {
write!(f, ", ")?;
}
prev = true;
write!(f, "{}", l)?;
}
write!(f, "]")
}
}
#[cfg(test)]
mod tests {
use super::*;
#[test]
fn resolve() {
let en_us: LanguageIdentifier = "en-US".parse().unwrap();
let en_ca: LanguageIdentifier = "en-CA".parse().unwrap();
let en_gb: LanguageIdentifier = "en-GB".parse().unwrap();
let fr_fr: LanguageIdentifier = "fr-FR".parse().unwrap();
let pt_pt: LanguageIdentifier = "pt-PT".parse().unwrap();
let resmgr = ResourceManager {
resources: HashMap::new(),
locales: vec![en_us.clone(), en_ca.clone(), en_gb.clone(), fr_fr.clone()],
default_locale: en_us.clone(),
path_scheme: String::new(),
};
let en_za: LanguageIdentifier = "en-GB".parse().unwrap();
let cn_hk: LanguageIdentifier = "cn-HK".parse().unwrap();
let fr_ca: LanguageIdentifier = "fr-CA".parse().unwrap();
assert_eq!(
resmgr.resolve_locales(en_ca.clone()),
vec![en_ca.clone(), en_us.clone(), en_gb.clone()]
);
assert_eq!(
resmgr.resolve_locales(en_za.clone()),
vec![en_gb.clone(), en_us.clone(), en_ca.clone()]
);
assert_eq!(
resmgr.resolve_locales(fr_ca.clone()),
vec![fr_fr.clone(), en_us.clone()]
);
assert_eq!(
resmgr.resolve_locales(fr_fr.clone()),
vec![fr_fr.clone(), en_us.clone()]
);
assert_eq!(resmgr.resolve_locales(cn_hk), vec![en_us.clone()]);
assert_eq!(resmgr.resolve_locales(pt_pt), vec![en_us.clone()]);
}
}
*/ ////
/// Implement formatted output for ArgSource
impl<T> core::fmt::Debug for ArgSource<T> { ////
fn fmt(&self, _fmt: &mut ::core::fmt::Formatter<'_>) -> ::core::fmt::Result | {
// TODO
Ok(())
} | identifier_body |
|
cardano_wallet.d.ts | {Uint8Array} salt
* @param {Uint8Array} nonce
* @param {Uint8Array} data
* @returns {any}
*/
export function password_encrypt(password: string, salt: Uint8Array, nonce: Uint8Array, data: Uint8Array): any;
/**
* decrypt the data with the password
*
* @param {string} password
* @param {Uint8Array} encrypted_data
* @returns {any}
*/
export function password_decrypt(password: string, encrypted_data: Uint8Array): any;
/**
*/
export class AccountIndex {
free(): void;
static new(index: number): AccountIndex;
}
/**
*/
export class Address {
free(): void;
to_base58(): string;
static from_base58(s: string): Address;
}
/**
*/
export class AddressKeyIndex {
free(): void;
static new(index: number): AddressKeyIndex;
}
/**
*/
export class Bip44AccountPrivate {
free(): void;
static new(key: PrivateKey, derivation_scheme: DerivationScheme): Bip44AccountPrivate;
public(): Bip44AccountPublic;
address_key(internal: boolean, index: AddressKeyIndex): PrivateKey;
key(): PrivateKey;
}
/**
*/
export class Bip44AccountPublic {
free(): void;
static new(key: PublicKey, derivation_scheme: DerivationScheme): Bip44AccountPublic;
address_key(internal: boolean, index: AddressKeyIndex): PublicKey;
key(): PublicKey;
}
/**
* Root Private Key of a BIP44 HD Wallet
*/
export class Bip44RootPrivateKey {
free(): void;
static new(key: PrivateKey, derivation_scheme: DerivationScheme): Bip44RootPrivateKey;
static recover(entropy: Entropy, password: string): Bip44RootPrivateKey;
bip44_account(index: AccountIndex): Bip44AccountPrivate;
key(): PrivateKey;
}
/**
* setting of the blockchain
*
* This includes the `ProtocolMagic` a discriminant value to differentiate
* different instances of the cardano blockchain (Mainnet, Testnet... ).
*/
export class BlockchainSettings {
free(): void;
to_json(): any;
static from_json(value: any): BlockchainSettings;
static mainnet(): BlockchainSettings;
}
/**
*/
export class Coin {
free(): void;
constructor();
static from_str(s: string): Coin;
to_str(): string;
static from(ada: number, lovelace: number): Coin;
ada(): number;
lovelace(): number;
add(other: Coin): Coin;
}
/**
*/
export class CoinDiff {
free(): void;
is_zero(): boolean;
is_negative(): boolean;
is_positive(): boolean;
value(): Coin;
}
/**
*/
export class DaedalusAddressChecker {
free(): void;
static new(wallet: DaedalusWallet): DaedalusAddressChecker;
check_address(address: Address): DaedalusCheckedAddress;
}
/**
* result value of the check_address function of the DaedalusAddressChecker.
*
* If the address passed to check_address was recognised by the daedalus wallet
* then this object will contain the private key associated to this wallet
* private key necessary to sign transactions
*/
export class DaedalusCheckedAddress {
free(): void;
is_checked(): boolean;
private_key(): PrivateKey;
}
/**
*/
export class DaedalusWallet {
free(): void;
static new(key: PrivateKey): DaedalusWallet;
static recover(entropy: Entropy): DaedalusWallet;
}
/**
* There is a special function to use when deriving Addresses. This function
* has been revised to offer stronger properties. This is why there is a
* V2 derivation scheme. The V1 being the legacy one still used in daedalus
* now a days.
*
* It is strongly advised to use V2 as the V1 is deprecated since April 2018.
* Its support is already provided for backward compatibility with old
* addresses.
*/
export class DerivationScheme {
free(): void;
static v1(): DerivationScheme;
static v2(): DerivationScheme;
}
/**
* the entropy associated to mnemonics. This is a bytes representation of the
* mnemonics the user has to remember how to generate the root key of an
* HD Wallet.
*
* TODO: interface to generate a new entropy
*
* # Security considerations
*
* * do not store this value without encrypting it;
* * do not leak the mnemonics;
* * make sure the user remembers the mnemonics string;
*
*/
export class Entropy {
free(): void;
static from_english_mnemonics(mnemonics: string): Entropy;
to_english_mnemonics(): string;
to_array(): any;
}
/**
*/
export class InputSelectionBuilder {
free(): void;
static first_match_first(): InputSelectionBuilder;
static largest_first(): InputSelectionBuilder;
static blackjack(dust_threshold: Coin): InputSelectionBuilder;
add_input(tx_input: TxInput): void;
add_output(output: TxOut): void;
select_inputs(fee_algorithm: LinearFeeAlgorithm, output_policy: OutputPolicy): InputSelectionResult;
}
/**
*/
export class InputSelectionResult {
free(): void;
is_input(txo_pointer: TxoPointer): boolean;
estimated_fees(): Coin;
estimated_change(): Coin;
}
/**
* This is the linear fee algorithm used buy the current cardano blockchain.
*
* However it is possible the linear fee algorithm may change its settings:
*
* It is currently a function `fee(n) = a * x + b`. `a` and `b` can be
* re-configured by a protocol update. Users of this object need to be aware
* that it may change and that they might need to update its settings.
*
*/
export class LinearFeeAlgorithm {
free(): void;
static default(): LinearFeeAlgorithm;
}
/**
* This is the Output policy for automatic Input selection.
*/
export class OutputPolicy {
free(): void;
static change_to_one_address(address: Address): OutputPolicy;
}
/**
* A given private key. You can use this key to sign transactions.
*
* # security considerations
*
* * do not store this key without encrypting it;
* * if leaked anyone can _spend_ a UTxO (Unspent Transaction Output)
* with it;
*
*/
export class PrivateKey {
free(): void;
static new(entropy: Entropy, password: string): PrivateKey;
static from_hex(hex: string): PrivateKey;
to_hex(): string;
public(): PublicKey;
sign(data: Uint8Array): Signature;
derive(derivation_scheme: DerivationScheme, index: number): PrivateKey;
}
/**
*/
export class PrivateRedeemKey {
free(): void;
static from_bytes(bytes: Uint8Array): PrivateRedeemKey;
static from_hex(hex: string): PrivateRedeemKey;
to_hex(): string;
public(): PublicRedeemKey;
sign(data: Uint8Array): RedeemSignature;
}
/**
* The public key associated to a given private key.
*
* It is not possible to sign (and then spend) with a private key.
* However it is possible to verify a Signature.
*
* # Security Consideration
*
* * it is rather harmless to leak a public key, in the worst case
* only the privacy is leaked;
*
*/
export class PublicKey {
free(): void;
static from_hex(hex: string): PublicKey;
to_hex(): string;
verify(data: Uint8Array, signature: Signature): boolean;
derive(derivation_scheme: DerivationScheme, index: number): PublicKey;
bootstrap_era_address(blockchain_settings: BlockchainSettings): Address;
}
/**
*/
export class | {
free(): void;
static from_hex(hex: string): PublicRedeemKey;
to_hex(): string;
verify(data: Uint8Array, signature: RedeemSignature): boolean;
address(settings: BlockchainSettings): Address;
}
/**
*/
export class RedeemSignature {
free(): void;
static from_hex(hex: string): RedeemSignature;
to_hex(): string;
}
/**
*/
export class Signature {
free(): void;
static from_hex(hex: string): Signature;
to_hex(): string;
}
/**
* a signed transaction, ready to be sent to the network.
*/
export class SignedTransaction {
free(): void;
id(): string;
to_json(): any;
static from_bytes(bytes: Uint8Array): SignedTransaction;
to_hex(): string;
}
/**
* a transaction type, this is not ready for sending to the network. It is only an
* intermediate type to use between the transaction builder and the transaction
* finalizer. It allows separation of concerns:
*
* 1. build the transaction on one side/thread/machine/...;
* 2. sign the transaction on the other/thread/machines/cold-wallet...;
*
*/
export class Transaction {
free(): void;
id(): TransactionId;
to_json(): any;
to_hex(): string;
}
/**
* The transaction builder provides a set of tools to help build
* a valid Transaction.
*/
export class TransactionBuilder {
free(): void;
constructor();
add_input(txo_pointer: TxoPointer, value: Coin): void;
get_input_total(): Coin;
add_output(output: TxOut): void;
apply_output_policy(fee_algorithm: LinearFeeAlgorithm, policy: OutputPolicy): any;
get_output_total(): | PublicRedeemKey | identifier_name |
cardano_wallet.d.ts | {Uint8Array} salt
* @param {Uint8Array} nonce
* @param {Uint8Array} data
* @returns {any}
*/
export function password_encrypt(password: string, salt: Uint8Array, nonce: Uint8Array, data: Uint8Array): any;
/**
* decrypt the data with the password
*
* @param {string} password
* @param {Uint8Array} encrypted_data
* @returns {any}
*/
export function password_decrypt(password: string, encrypted_data: Uint8Array): any;
/**
*/
export class AccountIndex {
free(): void;
static new(index: number): AccountIndex;
}
/**
*/
export class Address {
free(): void;
to_base58(): string;
static from_base58(s: string): Address;
}
/**
*/
export class AddressKeyIndex {
free(): void;
static new(index: number): AddressKeyIndex;
}
/**
*/
export class Bip44AccountPrivate {
free(): void;
static new(key: PrivateKey, derivation_scheme: DerivationScheme): Bip44AccountPrivate;
public(): Bip44AccountPublic;
address_key(internal: boolean, index: AddressKeyIndex): PrivateKey;
key(): PrivateKey;
}
/**
*/
export class Bip44AccountPublic {
free(): void;
static new(key: PublicKey, derivation_scheme: DerivationScheme): Bip44AccountPublic;
address_key(internal: boolean, index: AddressKeyIndex): PublicKey;
key(): PublicKey;
}
/**
* Root Private Key of a BIP44 HD Wallet
*/
export class Bip44RootPrivateKey {
free(): void;
static new(key: PrivateKey, derivation_scheme: DerivationScheme): Bip44RootPrivateKey;
static recover(entropy: Entropy, password: string): Bip44RootPrivateKey;
bip44_account(index: AccountIndex): Bip44AccountPrivate;
key(): PrivateKey;
}
/**
* setting of the blockchain
*
* This includes the `ProtocolMagic` a discriminant value to differentiate
* different instances of the cardano blockchain (Mainnet, Testnet... ).
*/
export class BlockchainSettings {
free(): void;
to_json(): any;
static from_json(value: any): BlockchainSettings;
static mainnet(): BlockchainSettings;
}
/**
*/
export class Coin {
free(): void;
constructor();
static from_str(s: string): Coin;
to_str(): string;
static from(ada: number, lovelace: number): Coin;
ada(): number;
lovelace(): number;
add(other: Coin): Coin;
}
/**
*/
export class CoinDiff {
free(): void;
is_zero(): boolean;
is_negative(): boolean;
is_positive(): boolean;
value(): Coin;
}
/**
*/
export class DaedalusAddressChecker {
free(): void;
static new(wallet: DaedalusWallet): DaedalusAddressChecker;
check_address(address: Address): DaedalusCheckedAddress;
}
/**
* result value of the check_address function of the DaedalusAddressChecker.
*
* If the address passed to check_address was recognised by the daedalus wallet
* then this object will contain the private key associated to this wallet
* private key necessary to sign transactions
*/
export class DaedalusCheckedAddress {
free(): void;
is_checked(): boolean;
private_key(): PrivateKey;
}
/**
*/
export class DaedalusWallet {
free(): void;
static new(key: PrivateKey): DaedalusWallet;
static recover(entropy: Entropy): DaedalusWallet;
}
/**
* There is a special function to use when deriving Addresses. This function
* has been revised to offer stronger properties. This is why there is a
* V2 derivation scheme. The V1 being the legacy one still used in daedalus
* now a days.
*
* It is strongly advised to use V2 as the V1 is deprecated since April 2018.
* Its support is already provided for backward compatibility with old
* addresses.
*/
export class DerivationScheme {
free(): void;
static v1(): DerivationScheme;
static v2(): DerivationScheme;
}
/**
* the entropy associated to mnemonics. This is a bytes representation of the
* mnemonics the user has to remember how to generate the root key of an
* HD Wallet.
*
* TODO: interface to generate a new entropy
*
* # Security considerations
*
* * do not store this value without encrypting it;
* * do not leak the mnemonics;
* * make sure the user remembers the mnemonics string;
* | export class Entropy {
free(): void;
static from_english_mnemonics(mnemonics: string): Entropy;
to_english_mnemonics(): string;
to_array(): any;
}
/**
*/
export class InputSelectionBuilder {
free(): void;
static first_match_first(): InputSelectionBuilder;
static largest_first(): InputSelectionBuilder;
static blackjack(dust_threshold: Coin): InputSelectionBuilder;
add_input(tx_input: TxInput): void;
add_output(output: TxOut): void;
select_inputs(fee_algorithm: LinearFeeAlgorithm, output_policy: OutputPolicy): InputSelectionResult;
}
/**
*/
export class InputSelectionResult {
free(): void;
is_input(txo_pointer: TxoPointer): boolean;
estimated_fees(): Coin;
estimated_change(): Coin;
}
/**
* This is the linear fee algorithm used buy the current cardano blockchain.
*
* However it is possible the linear fee algorithm may change its settings:
*
* It is currently a function `fee(n) = a * x + b`. `a` and `b` can be
* re-configured by a protocol update. Users of this object need to be aware
* that it may change and that they might need to update its settings.
*
*/
export class LinearFeeAlgorithm {
free(): void;
static default(): LinearFeeAlgorithm;
}
/**
* This is the Output policy for automatic Input selection.
*/
export class OutputPolicy {
free(): void;
static change_to_one_address(address: Address): OutputPolicy;
}
/**
* A given private key. You can use this key to sign transactions.
*
* # security considerations
*
* * do not store this key without encrypting it;
* * if leaked anyone can _spend_ a UTxO (Unspent Transaction Output)
* with it;
*
*/
export class PrivateKey {
free(): void;
static new(entropy: Entropy, password: string): PrivateKey;
static from_hex(hex: string): PrivateKey;
to_hex(): string;
public(): PublicKey;
sign(data: Uint8Array): Signature;
derive(derivation_scheme: DerivationScheme, index: number): PrivateKey;
}
/**
*/
export class PrivateRedeemKey {
free(): void;
static from_bytes(bytes: Uint8Array): PrivateRedeemKey;
static from_hex(hex: string): PrivateRedeemKey;
to_hex(): string;
public(): PublicRedeemKey;
sign(data: Uint8Array): RedeemSignature;
}
/**
* The public key associated to a given private key.
*
* It is not possible to sign (and then spend) with a private key.
* However it is possible to verify a Signature.
*
* # Security Consideration
*
* * it is rather harmless to leak a public key, in the worst case
* only the privacy is leaked;
*
*/
export class PublicKey {
free(): void;
static from_hex(hex: string): PublicKey;
to_hex(): string;
verify(data: Uint8Array, signature: Signature): boolean;
derive(derivation_scheme: DerivationScheme, index: number): PublicKey;
bootstrap_era_address(blockchain_settings: BlockchainSettings): Address;
}
/**
*/
export class PublicRedeemKey {
free(): void;
static from_hex(hex: string): PublicRedeemKey;
to_hex(): string;
verify(data: Uint8Array, signature: RedeemSignature): boolean;
address(settings: BlockchainSettings): Address;
}
/**
*/
export class RedeemSignature {
free(): void;
static from_hex(hex: string): RedeemSignature;
to_hex(): string;
}
/**
*/
export class Signature {
free(): void;
static from_hex(hex: string): Signature;
to_hex(): string;
}
/**
* a signed transaction, ready to be sent to the network.
*/
export class SignedTransaction {
free(): void;
id(): string;
to_json(): any;
static from_bytes(bytes: Uint8Array): SignedTransaction;
to_hex(): string;
}
/**
* a transaction type, this is not ready for sending to the network. It is only an
* intermediate type to use between the transaction builder and the transaction
* finalizer. It allows separation of concerns:
*
* 1. build the transaction on one side/thread/machine/...;
* 2. sign the transaction on the other/thread/machines/cold-wallet...;
*
*/
export class Transaction {
free(): void;
id(): TransactionId;
to_json(): any;
to_hex(): string;
}
/**
* The transaction builder provides a set of tools to help build
* a valid Transaction.
*/
export class TransactionBuilder {
free(): void;
constructor();
add_input(txo_pointer: TxoPointer, value: Coin): void;
get_input_total(): Coin;
add_output(output: TxOut): void;
apply_output_policy(fee_algorithm: LinearFeeAlgorithm, policy: OutputPolicy): any;
get_output_total(): Coin | */ | random_line_split |
Upload.js | 500);
}
async readFile(inputFile) {
let readTheData = (ifile, method)=>{
return new Promise((res, rej)=>{
let reader = new FileReader();
reader.onload = (e)=>{
res(e.target.result);
}
reader.onerror = (e)=>{
rej(e);
}
reader.onabort = ()=>{
rej();
}
reader[method](ifile);
});
};
let dims = (ifile)=>{
return new Promise((res,rej)=>{
const img = new Image();
img.onload = function () {
res([this.width, this.height]);
};
img.onerror = function() {
res([100,100]); // ???
};
img.src = ifile.dataURL;
});
};
try {
let ab = await readTheData(inputFile, 'readAsArrayBuffer');
let filename = (''+inputFile.name) || 'undefined';
let ext = (filename.lastIndexOf('.') != -1) ? (filename.substr(filename.lastIndexOf('.') + 1)).toLowerCase() : '';
let fileData = {
ab: ab,
name: inputFile.name,
size: inputFile.size,
filename: filename,
ext: ext,
type: 'doc',
aspectRatio: 1,
};
// nice one. https://stackoverflow.com/a/20732091/1119169 thanks Andrew!
const sizeI = Math.floor( Math.log(fileData.size) / Math.log(1024) );
fileData.sizeHuman = ( fileData.size / Math.pow(1024, sizeI) ).toFixed(2) * 1 + ' ' + ['B', 'kB', 'MB', 'GB', 'TB'][sizeI];
fileData.color = ((''+ext+' ').charCodeAt(0) % 8 + 1);
fileData.random = (''+Math.random()).split('.').join('');
let photoExts = ['png', 'jpg', 'jpeg'];
let videoExts = ['mp4', 'mpeg', 'avi', 'mov'];
if (photoExts.indexOf(ext) != -1 && this._uploadingMedia) {
fileData.type = 'photo';
} else if (videoExts.indexOf(ext) != -1 && this._uploadingMedia) {
fileData.type = 'video';
}
if (fileData.type == 'photo') {
// let dataURL = await readTheData(inputFile, 'readAsDataURL'); // ? URL.createObjectURL(file);
fileData.dataURL = URL.createObjectURL(inputFile);
[fileData.width, fileData.height] = await dims(fileData); // parallelize?
} else if (fileData.type == 'video') {
fileData.width = 100;
fileData.height = 100;
}
this._files.push(fileData);
return fileData;
} catch(e) {
// console.log(e);
return false;
}
}
loaded() {
// console.error(this._files);
this.updateTitle();
this.generateUploadPreview();
}
updateTitle() {
let suffix = {
doc: 'File',
video: 'Video',
photo: 'Photo',
};
let title = 'Send '+( (this._files.length > 1) ? this._files.length : '' ) + ' ' + suffix[this._files[0].type] + ( (this._files.length > 1) ? 's' : '' );
this.$('.uploadTitle').innerHTML = title;
}
onAlbumClick(e) {
const base = this.$();
let closest = event.target.closest('.upiRemove');
if (closest && base.contains(closest)) {
const el = closest.parentElement;
if (el) {
const id = el.id.split('uploadPreviewItem_').join('');
let ri = null;
for (let i = 0; i < this._files.length; i++) {
if (this._files[i].random == id) {
ri = i;
}
}
if (ri !== null) {
this._files.splice(ri, 1);
el.remove();
if (!this._files.length) {
this.hide();
} else {
this.doCalcs();
this.updateTitle();
}
}
}
}
}
doCalcs() {
this._layouter = new Layouter(this._files, {maxWidth: 384});
this._files = this._layouter.layout();
for (let file of this._files) {
const el = this.$('#uploadPreviewItem_'+file.random);
if (el) |
}
this.$('#uploadPreviewAlbum').style.height = this._layouter._height+'px';
this.initScrollBarOn(this.$('.uploadPreview'));
}
swapItems(id1, id2) {
if (id1 == id2) {
return;
}
id1 = id1.split('uploadPreviewItem_').join('');
id2 = id2.split('uploadPreviewItem_').join('');
let toMove1 = null;
let i1 = null;
let toMove2 = null;
let i2 = null;
// let insertBeforeI = null;
for (let i = 0; i < this._files.length; i++) {
if (this._files[i].random == id1) {
toMove1 = this._files[i];
i1 = i;
}
if (this._files[i].random == id2) {
toMove2 = this._files[i];
i2 = i;
}
}
this._files[i1] = toMove2;
this._files[i2] = toMove1;
this.doCalcs();
}
async generateVideoPreview(file, tryToSliceLength) {
let respAB = file.ab;
if (tryToSliceLength) {
respAB = file.ab.slice(0, tryToSliceLength);
}
let response = new Response(
respAB,
{
status: 206,
statusText: 'Partial Content',
headers: [
['Content-Type', 'video/mp4'],
['Content-Length', file.ab.byteLength],
['Content-Range', '0-'+(respAB.byteLength - 1)+'/'+file.ab.byteLength ]]
});
let blob = await response.blob();
let blobUrl = URL.createObjectURL(blob);
let promise = new Promise((res, rej)=>{
let video = document.createElement('video');
video.addEventListener('error', function(event) {
rej();
}, true);
video.onloadeddata = (event) => {
let width = video.videoWidth;
let height = video.videoHeight;
let canvas = document.createElement('canvas');
canvas.width = width;
canvas.height = height;
file.aspectRatio = (height ? (width/height) : 1);
file.width = width;
file.height = height;
this.doCalcs();
let ctx = canvas.getContext("2d");
ctx.imageSmoothingEnabled = true;
ctx.drawImage(video, 0, 0, width, height);
let blankCanvas = document.createElement('canvas');
blankCanvas.width = width;
blankCanvas.height = height;
if (canvas.toDataURL() == blankCanvas.toDataURL()) {
rej();
}
canvas.toBlob((canvasBlob)=>{
let canvasBlobUrl = URL.createObjectURL(canvasBlob);
res(canvasBlobUrl);
});
};
video.preload = 'metadata';
video.src = blobUrl;
// Load video in Safari / IE11
video.muted = true;
video.playsInline = true;
video.play();
});
return await promise;
}
generateUploadPreviewForFile(file) {
const closeHTML = this._components.CloseIcon.render({noDOM: true});
if (file.type == 'photo') {
return '<div class="uploadPreviewItem" draggable="true" id="uploadPreviewItem_'+file.random+'" style="background-image: url(\''+file.dataURL+'\'); '+file.style+'"><div class="upiRemove">'+closeHTML+'</div></div>';
} else if (file.type == 'video') {
return '<div class="uploadPreviewItem" draggable="true" id="uploadPreviewItem_'+file.random+'" style="border: 1px solid #eee;"><div class="upiRemove">'+closeHTML+'</div></div>';
} else if (file.type == 'doc') {
return `<div class="uploadDocItem">
<div class="rsDoc">
<div class="rsDocIcon avatarC${file.color}">${file.ext}</div>
<div class="rsDocName">${file.filename}</div>
<div class="rsDocMeta">${file.sizeHuman}</div>
</div>
</div>`;
}
| {
el.style.left = '' + file.pos.left + 'px';
el.style.top = '' + file.pos.top + 'px';
el.style.width = '' + file.pos.width + 'px';
el.style.height = '' + file.pos.height + 'px';
} | conditional_block |
Upload.js | 500);
}
async readFile(inputFile) {
let readTheData = (ifile, method)=>{
return new Promise((res, rej)=>{
let reader = new FileReader();
reader.onload = (e)=>{
res(e.target.result);
}
reader.onerror = (e)=>{
rej(e);
}
reader.onabort = ()=>{
rej();
}
reader[method](ifile);
});
};
let dims = (ifile)=>{
return new Promise((res,rej)=>{
const img = new Image();
img.onload = function () {
res([this.width, this.height]);
};
img.onerror = function() {
res([100,100]); // ???
};
img.src = ifile.dataURL;
});
};
try {
let ab = await readTheData(inputFile, 'readAsArrayBuffer');
let filename = (''+inputFile.name) || 'undefined';
let ext = (filename.lastIndexOf('.') != -1) ? (filename.substr(filename.lastIndexOf('.') + 1)).toLowerCase() : '';
let fileData = {
ab: ab,
name: inputFile.name,
size: inputFile.size,
filename: filename,
ext: ext,
type: 'doc',
aspectRatio: 1,
};
// nice one. https://stackoverflow.com/a/20732091/1119169 thanks Andrew!
const sizeI = Math.floor( Math.log(fileData.size) / Math.log(1024) );
fileData.sizeHuman = ( fileData.size / Math.pow(1024, sizeI) ).toFixed(2) * 1 + ' ' + ['B', 'kB', 'MB', 'GB', 'TB'][sizeI];
fileData.color = ((''+ext+' ').charCodeAt(0) % 8 + 1);
fileData.random = (''+Math.random()).split('.').join('');
let photoExts = ['png', 'jpg', 'jpeg'];
let videoExts = ['mp4', 'mpeg', 'avi', 'mov'];
if (photoExts.indexOf(ext) != -1 && this._uploadingMedia) {
fileData.type = 'photo';
} else if (videoExts.indexOf(ext) != -1 && this._uploadingMedia) {
fileData.type = 'video';
}
if (fileData.type == 'photo') {
// let dataURL = await readTheData(inputFile, 'readAsDataURL'); // ? URL.createObjectURL(file);
fileData.dataURL = URL.createObjectURL(inputFile);
[fileData.width, fileData.height] = await dims(fileData); // parallelize?
} else if (fileData.type == 'video') {
fileData.width = 100;
fileData.height = 100;
}
this._files.push(fileData);
return fileData;
} catch(e) {
// console.log(e);
return false;
}
}
loaded() {
// console.error(this._files);
this.updateTitle();
this.generateUploadPreview();
}
updateTitle() {
let suffix = {
doc: 'File',
video: 'Video',
photo: 'Photo', | let title = 'Send '+( (this._files.length > 1) ? this._files.length : '' ) + ' ' + suffix[this._files[0].type] + ( (this._files.length > 1) ? 's' : '' );
this.$('.uploadTitle').innerHTML = title;
}
onAlbumClick(e) {
const base = this.$();
let closest = event.target.closest('.upiRemove');
if (closest && base.contains(closest)) {
const el = closest.parentElement;
if (el) {
const id = el.id.split('uploadPreviewItem_').join('');
let ri = null;
for (let i = 0; i < this._files.length; i++) {
if (this._files[i].random == id) {
ri = i;
}
}
if (ri !== null) {
this._files.splice(ri, 1);
el.remove();
if (!this._files.length) {
this.hide();
} else {
this.doCalcs();
this.updateTitle();
}
}
}
}
}
doCalcs() {
this._layouter = new Layouter(this._files, {maxWidth: 384});
this._files = this._layouter.layout();
for (let file of this._files) {
const el = this.$('#uploadPreviewItem_'+file.random);
if (el) {
el.style.left = '' + file.pos.left + 'px';
el.style.top = '' + file.pos.top + 'px';
el.style.width = '' + file.pos.width + 'px';
el.style.height = '' + file.pos.height + 'px';
}
}
this.$('#uploadPreviewAlbum').style.height = this._layouter._height+'px';
this.initScrollBarOn(this.$('.uploadPreview'));
}
swapItems(id1, id2) {
if (id1 == id2) {
return;
}
id1 = id1.split('uploadPreviewItem_').join('');
id2 = id2.split('uploadPreviewItem_').join('');
let toMove1 = null;
let i1 = null;
let toMove2 = null;
let i2 = null;
// let insertBeforeI = null;
for (let i = 0; i < this._files.length; i++) {
if (this._files[i].random == id1) {
toMove1 = this._files[i];
i1 = i;
}
if (this._files[i].random == id2) {
toMove2 = this._files[i];
i2 = i;
}
}
this._files[i1] = toMove2;
this._files[i2] = toMove1;
this.doCalcs();
}
async generateVideoPreview(file, tryToSliceLength) {
let respAB = file.ab;
if (tryToSliceLength) {
respAB = file.ab.slice(0, tryToSliceLength);
}
let response = new Response(
respAB,
{
status: 206,
statusText: 'Partial Content',
headers: [
['Content-Type', 'video/mp4'],
['Content-Length', file.ab.byteLength],
['Content-Range', '0-'+(respAB.byteLength - 1)+'/'+file.ab.byteLength ]]
});
let blob = await response.blob();
let blobUrl = URL.createObjectURL(blob);
let promise = new Promise((res, rej)=>{
let video = document.createElement('video');
video.addEventListener('error', function(event) {
rej();
}, true);
video.onloadeddata = (event) => {
let width = video.videoWidth;
let height = video.videoHeight;
let canvas = document.createElement('canvas');
canvas.width = width;
canvas.height = height;
file.aspectRatio = (height ? (width/height) : 1);
file.width = width;
file.height = height;
this.doCalcs();
let ctx = canvas.getContext("2d");
ctx.imageSmoothingEnabled = true;
ctx.drawImage(video, 0, 0, width, height);
let blankCanvas = document.createElement('canvas');
blankCanvas.width = width;
blankCanvas.height = height;
if (canvas.toDataURL() == blankCanvas.toDataURL()) {
rej();
}
canvas.toBlob((canvasBlob)=>{
let canvasBlobUrl = URL.createObjectURL(canvasBlob);
res(canvasBlobUrl);
});
};
video.preload = 'metadata';
video.src = blobUrl;
// Load video in Safari / IE11
video.muted = true;
video.playsInline = true;
video.play();
});
return await promise;
}
generateUploadPreviewForFile(file) {
const closeHTML = this._components.CloseIcon.render({noDOM: true});
if (file.type == 'photo') {
return '<div class="uploadPreviewItem" draggable="true" id="uploadPreviewItem_'+file.random+'" style="background-image: url(\''+file.dataURL+'\'); '+file.style+'"><div class="upiRemove">'+closeHTML+'</div></div>';
} else if (file.type == 'video') {
return '<div class="uploadPreviewItem" draggable="true" id="uploadPreviewItem_'+file.random+'" style="border: 1px solid #eee;"><div class="upiRemove">'+closeHTML+'</div></div>';
} else if (file.type == 'doc') {
return `<div class="uploadDocItem">
<div class="rsDoc">
<div class="rsDocIcon avatarC${file.color}">${file.ext}</div>
<div class="rsDocName">${file.filename}</div>
<div class="rsDocMeta">${file.sizeHuman}</div>
</div>
</div>`;
}
}
| };
| random_line_split |
Upload.js | () {
this.$('#uploadTop').style.display = 'block';
this.$('.popupOverlay').classList.add('active');
this.loaded();
this.initScrollBarOn(this.$('.uploadPreview'));
}
hide() {
this.$('.popupOverlay').classList.remove('active');
this.$('.popupOverlay').classList.add('fading');
setTimeout(()=>{
this.$('.popupOverlay').classList.remove('fading');
this.$('#uploadTop').style.display = 'none';
this.buttonLoading(false);
}, 500);
}
async readFile(inputFile) {
let readTheData = (ifile, method)=>{
return new Promise((res, rej)=>{
let reader = new FileReader();
reader.onload = (e)=>{
res(e.target.result);
}
reader.onerror = (e)=>{
rej(e);
}
reader.onabort = ()=>{
rej();
}
reader[method](ifile);
});
};
let dims = (ifile)=>{
return new Promise((res,rej)=>{
const img = new Image();
img.onload = function () {
res([this.width, this.height]);
};
img.onerror = function() {
res([100,100]); // ???
};
img.src = ifile.dataURL;
});
};
try {
let ab = await readTheData(inputFile, 'readAsArrayBuffer');
let filename = (''+inputFile.name) || 'undefined';
let ext = (filename.lastIndexOf('.') != -1) ? (filename.substr(filename.lastIndexOf('.') + 1)).toLowerCase() : '';
let fileData = {
ab: ab,
name: inputFile.name,
size: inputFile.size,
filename: filename,
ext: ext,
type: 'doc',
aspectRatio: 1,
};
// nice one. https://stackoverflow.com/a/20732091/1119169 thanks Andrew!
const sizeI = Math.floor( Math.log(fileData.size) / Math.log(1024) );
fileData.sizeHuman = ( fileData.size / Math.pow(1024, sizeI) ).toFixed(2) * 1 + ' ' + ['B', 'kB', 'MB', 'GB', 'TB'][sizeI];
fileData.color = ((''+ext+' ').charCodeAt(0) % 8 + 1);
fileData.random = (''+Math.random()).split('.').join('');
let photoExts = ['png', 'jpg', 'jpeg'];
let videoExts = ['mp4', 'mpeg', 'avi', 'mov'];
if (photoExts.indexOf(ext) != -1 && this._uploadingMedia) {
fileData.type = 'photo';
} else if (videoExts.indexOf(ext) != -1 && this._uploadingMedia) {
fileData.type = 'video';
}
if (fileData.type == 'photo') {
// let dataURL = await readTheData(inputFile, 'readAsDataURL'); // ? URL.createObjectURL(file);
fileData.dataURL = URL.createObjectURL(inputFile);
[fileData.width, fileData.height] = await dims(fileData); // parallelize?
} else if (fileData.type == 'video') {
fileData.width = 100;
fileData.height = 100;
}
this._files.push(fileData);
return fileData;
} catch(e) {
// console.log(e);
return false;
}
}
loaded() {
// console.error(this._files);
this.updateTitle();
this.generateUploadPreview();
}
updateTitle() {
let suffix = {
doc: 'File',
video: 'Video',
photo: 'Photo',
};
let title = 'Send '+( (this._files.length > 1) ? this._files.length : '' ) + ' ' + suffix[this._files[0].type] + ( (this._files.length > 1) ? 's' : '' );
this.$('.uploadTitle').innerHTML = title;
}
onAlbumClick(e) {
const base = this.$();
let closest = event.target.closest('.upiRemove');
if (closest && base.contains(closest)) {
const el = closest.parentElement;
if (el) {
const id = el.id.split('uploadPreviewItem_').join('');
let ri = null;
for (let i = 0; i < this._files.length; i++) {
if (this._files[i].random == id) {
ri = i;
}
}
if (ri !== null) {
this._files.splice(ri, 1);
el.remove();
if (!this._files.length) {
this.hide();
} else {
this.doCalcs();
this.updateTitle();
}
}
}
}
}
doCalcs() {
this._layouter = new Layouter(this._files, {maxWidth: 384});
this._files = this._layouter.layout();
for (let file of this._files) {
const el = this.$('#uploadPreviewItem_'+file.random);
if (el) {
el.style.left = '' + file.pos.left + 'px';
el.style.top = '' + file.pos.top + 'px';
el.style.width = '' + file.pos.width + 'px';
el.style.height = '' + file.pos.height + 'px';
}
}
this.$('#uploadPreviewAlbum').style.height = this._layouter._height+'px';
this.initScrollBarOn(this.$('.uploadPreview'));
}
swapItems(id1, id2) {
if (id1 == id2) {
return;
}
id1 = id1.split('uploadPreviewItem_').join('');
id2 = id2.split('uploadPreviewItem_').join('');
let toMove1 = null;
let i1 = null;
let toMove2 = null;
let i2 = null;
// let insertBeforeI = null;
for (let i = 0; i < this._files.length; i++) {
if (this._files[i].random == id1) {
toMove1 = this._files[i];
i1 = i;
}
if (this._files[i].random == id2) {
toMove2 = this._files[i];
i2 = i;
}
}
this._files[i1] = toMove2;
this._files[i2] = toMove1;
this.doCalcs();
}
async generateVideoPreview(file, tryToSliceLength) {
let respAB = file.ab;
if (tryToSliceLength) {
respAB = file.ab.slice(0, tryToSliceLength);
}
let response = new Response(
respAB,
{
status: 206,
statusText: 'Partial Content',
headers: [
['Content-Type', 'video/mp4'],
['Content-Length', file.ab.byteLength],
['Content-Range', '0-'+(respAB.byteLength - 1)+'/'+file.ab.byteLength ]]
});
let blob = await response.blob();
let blobUrl = URL.createObjectURL(blob);
let promise = new Promise((res, rej)=>{
let video = document.createElement('video');
video.addEventListener('error', function(event) {
rej();
}, true);
video.onloadeddata = (event) => {
let width = video.videoWidth;
let height = video.videoHeight;
let canvas = document.createElement('canvas');
canvas.width = width;
canvas.height = height;
file.aspectRatio = (height ? (width/height) : 1);
file.width = width;
file.height = height;
this.doCalcs();
let ctx = canvas.getContext("2d");
ctx.imageSmoothingEnabled = true;
ctx.drawImage(video, 0, 0, width, height);
let blankCanvas = document.createElement('canvas');
blankCanvas.width = width;
blankCanvas.height = height;
if (canvas.toDataURL() == blankCanvas.toDataURL()) {
rej();
}
canvas.toBlob((canvasBlob)=>{
let canvasBlobUrl = URL.createObjectURL(canvasBlob);
res(canvasBlobUrl);
});
};
video.preload = 'metadata';
video.src = blobUrl;
// Load video in Safari / IE11
video.muted = true;
video.playsInline = true;
video.play();
});
return await promise;
}
generateUploadPreviewForFile(file) {
const closeHTML = this._components.CloseIcon.render({noDOM: true});
if (file.type == 'photo') {
return '<div class="uploadPreviewItem" draggable="true" id="uploadPreviewItem_'+file.random+'" style="background-image: url(\''+file.dataURL+'\'); '+file.style+'"><div class="upiRemove">'+closeHTML+'</div></div>';
} else if (file.type == 'video') {
return '<div class="uploadPreviewItem" draggable="true" id="uploadPreviewItem_'+file.random+'" style="border: 1px solid #eee;">< | show | identifier_name |
|
image_utils.py | n_channels == 1:
curr_image = np.reshape(curr_image, [height, width])
s = io.BytesIO()
matplotlib_pyplot().imsave(s, curr_image, format="png")
img_sum = tf.Summary.Image(encoded_image_string=s.getvalue(),
height=height, width=width,
colorspace=n_channels)
return tf.Summary.Value(tag=tag, image=img_sum)
def convert_predictions_to_image_summaries(hook_args):
"""Optionally converts images from hooks_args to image summaries.
Args:
hook_args: DecodeHookArgs namedtuple
Returns:
summaries: list of tf.Summary values if hook_args.decode_hpara
"""
decode_hparams = hook_args.decode_hparams
if not decode_hparams.display_decoded_images:
return []
predictions = hook_args.predictions[0]
# Display ten random inputs and outputs so that tensorboard does not hang.
all_summaries = []
rand_predictions = np.random.choice(predictions, size=10)
for ind, prediction in enumerate(rand_predictions):
output_summary = image_to_tf_summary_value(
prediction["outputs"], tag="%d_output" % ind)
input_summary = image_to_tf_summary_value(
prediction["inputs"], tag="%d_input" % ind)
all_summaries.append(input_summary)
all_summaries.append(output_summary)
return all_summaries
def resize_by_area(img, size):
"""image resize function used by quite a few image problems."""
return tf.to_int64(
tf.image.resize_images(img, [size, size], tf.image.ResizeMethod.AREA))
def make_multiscale(image, resolutions,
resize_method=tf.image.ResizeMethod.BICUBIC,
num_channels=3):
"""Returns list of scaled images, one for each resolution.
Args:
image: Tensor of shape [height, height, num_channels].
resolutions: List of heights that image's height is resized to.
resize_method: tf.image.ResizeMethod.
num_channels: Number of channels in image.
Returns:
List of Tensors, one for each resolution with shape given by
[resolutions[i], resolutions[i], num_channels].
"""
scaled_images = []
for height in resolutions:
scaled_image = tf.image.resize_images(
image,
size=[height, height], # assuming that height = width
method=resize_method)
scaled_image = tf.to_int64(scaled_image)
scaled_image.set_shape([height, height, num_channels])
scaled_images.append(scaled_image)
return scaled_images
def make_multiscale_dilated(image, resolutions, num_channels=3):
"""Returns list of scaled images, one for each resolution.
Resizes by skipping every nth pixel.
Args:
image: Tensor of shape [height, height, num_channels].
resolutions: List of heights that image's height is resized to. The function
assumes VALID padding, so the original image's height must be divisible
by each resolution's height to return the exact resolution size.
num_channels: Number of channels in image.
Returns:
List of Tensors, one for each resolution with shape given by
[resolutions[i], resolutions[i], num_channels] if resolutions properly
divide the original image's height; otherwise shape height and width is up
to valid skips.
"""
image_height = common_layers.shape_list(image)[0]
scaled_images = []
for height in resolutions:
dilation_rate = image_height // height # assuming height = width
scaled_image = image[::dilation_rate, ::dilation_rate]
scaled_image = tf.to_int64(scaled_image)
scaled_image.set_shape([None, None, num_channels])
scaled_images.append(scaled_image)
return scaled_images
class ImageProblem(problem.Problem):
"""Base class for problems with images."""
@property
def num_channels(self):
"""Number of color channels."""
return 3
@property
def vocab_size(self):
"""Number of pixel values."""
return 256
def example_reading_spec(self):
data_fields = {
"image/encoded": tf.FixedLenFeature((), tf.string),
"image/format": tf.FixedLenFeature((), tf.string),
}
data_items_to_decoders = {
"inputs":
tf.contrib.slim.tfexample_decoder.Image(
image_key="image/encoded",
format_key="image/format",
channels=self.num_channels),
}
return data_fields, data_items_to_decoders
def preprocess_example(self, example, mode, hparams):
if not self._was_reversed:
example["inputs"] = tf.image.per_image_standardization(example["inputs"])
return example
def eval_metrics(self):
eval_metrics = [
metrics.Metrics.ACC, metrics.Metrics.ACC_TOP5,
metrics.Metrics.ACC_PER_SEQ, metrics.Metrics.NEG_LOG_PERPLEXITY
]
if self._was_reversed:
eval_metrics += [metrics.Metrics.IMAGE_SUMMARY]
return eval_metrics
@property
def decode_hooks(self):
return [convert_predictions_to_image_summaries]
class Image2ClassProblem(ImageProblem):
"""Base class for image classification problems."""
@property
def is_small(self):
raise NotImplementedError()
@property
def num_classes(self):
raise NotImplementedError()
@property
def train_shards(self):
raise NotImplementedError()
@property
def dev_shards(self):
return 1
@property
def class_labels(self):
return ["ID_%d" % i for i in range(self.num_classes)]
def feature_encoders(self, data_dir):
del data_dir
return {
"inputs": text_encoder.ImageEncoder(channels=self.num_channels),
"targets": text_encoder.ClassLabelEncoder(self.class_labels)
}
def generator(self, data_dir, tmp_dir, is_training):
raise NotImplementedError()
def example_reading_spec(self):
label_key = "image/class/label"
data_fields, data_items_to_decoders = (
super(Image2ClassProblem, self).example_reading_spec())
data_fields[label_key] = tf.FixedLenFeature((1,), tf.int64)
data_items_to_decoders[
"targets"] = tf.contrib.slim.tfexample_decoder.Tensor(label_key)
return data_fields, data_items_to_decoders
def hparams(self, defaults, unused_model_hparams):
p = defaults
p.modality = {"inputs": modalities.ModalityType.IMAGE,
"targets": modalities.ModalityType.CLASS_LABEL}
p.vocab_size = {"inputs": 256,
"targets": self.num_classes}
p.batch_size_multiplier = 4 if self.is_small else 256
p.loss_multiplier = 3.0 if self.is_small else 1.0
if self._was_reversed:
p.loss_multiplier = 1.0
p.input_space_id = problem.SpaceID.IMAGE
p.target_space_id = problem.SpaceID.IMAGE_LABEL
def generate_data(self, data_dir, tmp_dir, task_id=-1):
generator_utils.generate_dataset_and_shuffle(
self.generator(data_dir, tmp_dir, True),
self.training_filepaths(data_dir, self.train_shards, shuffled=False),
self.generator(data_dir, tmp_dir, False),
self.dev_filepaths(data_dir, self.dev_shards, shuffled=False))
def encode_images_as_png(images):
"""Yield images encoded as pngs."""
if tf.executing_eagerly():
for image in images:
yield tf.image.encode_png(image).numpy()
else:
(height, width, channels) = images[0].shape
with tf.Graph().as_default():
image_t = tf.placeholder(dtype=tf.uint8, shape=(height, width, channels))
encoded_image_t = tf.image.encode_png(image_t)
with tf.Session() as sess:
for image in images:
enc_string = sess.run(encoded_image_t, feed_dict={image_t: image})
yield enc_string
def image_generator(images, labels):
"""Generator for images that takes image and labels lists and creates pngs.
Args:
images: list of images given as [width x height x channels] numpy arrays.
labels: list of ints, same length as images.
Yields:
A dictionary representing the images with the following fields:
* image/encoded: the string encoding the image as PNG,
* image/format: the string "png" representing image format,
* image/class/label: an integer representing the label,
* image/height: an integer representing the height,
* image/width: an integer representing the width.
Every field is actually a singleton list of the corresponding type.
Raises:
ValueError: if images is an empty list.
"""
if not images:
raise ValueError("Must provide some images for the generator.")
width, height, _ = images[0].shape
for (enc_image, label) in zip(encode_images_as_png(images), labels):
|
class Image2TextProblem(ImageProblem):
"""Base class for image-to-text problems."""
@property
def is_character_level(self):
raise NotImplementedError()
@property
def vocab_problem(self):
raise NotImplementedError() # Not needed if self.is_character_level.
@property
def target_space_id | yield {
"image/encoded": [enc_image],
"image/format": ["png"],
"image/class/label": [int(label)],
"image/height": [height],
"image/width": [width]
} | conditional_block |
image_utils.py | n_channels == 1:
curr_image = np.reshape(curr_image, [height, width])
s = io.BytesIO()
matplotlib_pyplot().imsave(s, curr_image, format="png")
img_sum = tf.Summary.Image(encoded_image_string=s.getvalue(),
height=height, width=width,
colorspace=n_channels)
return tf.Summary.Value(tag=tag, image=img_sum)
def convert_predictions_to_image_summaries(hook_args):
"""Optionally converts images from hooks_args to image summaries.
Args:
hook_args: DecodeHookArgs namedtuple
Returns:
summaries: list of tf.Summary values if hook_args.decode_hpara
"""
decode_hparams = hook_args.decode_hparams
if not decode_hparams.display_decoded_images:
return []
predictions = hook_args.predictions[0]
# Display ten random inputs and outputs so that tensorboard does not hang.
all_summaries = []
rand_predictions = np.random.choice(predictions, size=10)
for ind, prediction in enumerate(rand_predictions):
output_summary = image_to_tf_summary_value(
prediction["outputs"], tag="%d_output" % ind)
input_summary = image_to_tf_summary_value(
prediction["inputs"], tag="%d_input" % ind)
all_summaries.append(input_summary)
all_summaries.append(output_summary)
return all_summaries
def resize_by_area(img, size):
"""image resize function used by quite a few image problems."""
return tf.to_int64(
tf.image.resize_images(img, [size, size], tf.image.ResizeMethod.AREA))
def make_multiscale(image, resolutions,
resize_method=tf.image.ResizeMethod.BICUBIC,
num_channels=3):
"""Returns list of scaled images, one for each resolution.
Args:
image: Tensor of shape [height, height, num_channels].
resolutions: List of heights that image's height is resized to.
resize_method: tf.image.ResizeMethod.
num_channels: Number of channels in image.
Returns:
List of Tensors, one for each resolution with shape given by
[resolutions[i], resolutions[i], num_channels].
"""
scaled_images = []
for height in resolutions:
scaled_image = tf.image.resize_images(
image,
size=[height, height], # assuming that height = width
method=resize_method)
scaled_image = tf.to_int64(scaled_image)
scaled_image.set_shape([height, height, num_channels])
scaled_images.append(scaled_image)
return scaled_images
def make_multiscale_dilated(image, resolutions, num_channels=3):
"""Returns list of scaled images, one for each resolution.
Resizes by skipping every nth pixel.
Args:
image: Tensor of shape [height, height, num_channels].
resolutions: List of heights that image's height is resized to. The function
assumes VALID padding, so the original image's height must be divisible
by each resolution's height to return the exact resolution size.
num_channels: Number of channels in image.
Returns:
List of Tensors, one for each resolution with shape given by
[resolutions[i], resolutions[i], num_channels] if resolutions properly
divide the original image's height; otherwise shape height and width is up
to valid skips.
"""
image_height = common_layers.shape_list(image)[0]
scaled_images = []
for height in resolutions:
dilation_rate = image_height // height # assuming height = width
scaled_image = image[::dilation_rate, ::dilation_rate]
scaled_image = tf.to_int64(scaled_image)
scaled_image.set_shape([None, None, num_channels])
scaled_images.append(scaled_image)
return scaled_images
class ImageProblem(problem.Problem):
"""Base class for problems with images."""
@property
def num_channels(self):
"""Number of color channels."""
return 3
@property
def vocab_size(self):
"""Number of pixel values."""
return 256
def example_reading_spec(self):
data_fields = {
"image/encoded": tf.FixedLenFeature((), tf.string),
"image/format": tf.FixedLenFeature((), tf.string),
}
data_items_to_decoders = {
"inputs":
tf.contrib.slim.tfexample_decoder.Image(
image_key="image/encoded",
format_key="image/format",
channels=self.num_channels),
}
return data_fields, data_items_to_decoders
def preprocess_example(self, example, mode, hparams):
if not self._was_reversed:
example["inputs"] = tf.image.per_image_standardization(example["inputs"])
return example
def eval_metrics(self):
eval_metrics = [
metrics.Metrics.ACC, metrics.Metrics.ACC_TOP5,
metrics.Metrics.ACC_PER_SEQ, metrics.Metrics.NEG_LOG_PERPLEXITY
]
if self._was_reversed:
eval_metrics += [metrics.Metrics.IMAGE_SUMMARY]
return eval_metrics
@property
def decode_hooks(self):
return [convert_predictions_to_image_summaries]
class Image2ClassProblem(ImageProblem):
"""Base class for image classification problems."""
@property
def is_small(self):
raise NotImplementedError()
@property
def num_classes(self):
raise NotImplementedError()
@property
def train_shards(self):
raise NotImplementedError()
@property
def dev_shards(self):
return 1
@property
def class_labels(self):
return ["ID_%d" % i for i in range(self.num_classes)]
def feature_encoders(self, data_dir):
del data_dir
return {
"inputs": text_encoder.ImageEncoder(channels=self.num_channels),
"targets": text_encoder.ClassLabelEncoder(self.class_labels)
}
def generator(self, data_dir, tmp_dir, is_training):
raise NotImplementedError()
def example_reading_spec(self):
label_key = "image/class/label"
data_fields, data_items_to_decoders = (
super(Image2ClassProblem, self).example_reading_spec())
data_fields[label_key] = tf.FixedLenFeature((1,), tf.int64)
data_items_to_decoders[
"targets"] = tf.contrib.slim.tfexample_decoder.Tensor(label_key)
return data_fields, data_items_to_decoders
def hparams(self, defaults, unused_model_hparams):
p = defaults
p.modality = {"inputs": modalities.ModalityType.IMAGE,
"targets": modalities.ModalityType.CLASS_LABEL}
p.vocab_size = {"inputs": 256,
"targets": self.num_classes}
p.batch_size_multiplier = 4 if self.is_small else 256
p.loss_multiplier = 3.0 if self.is_small else 1.0
if self._was_reversed:
p.loss_multiplier = 1.0
p.input_space_id = problem.SpaceID.IMAGE
p.target_space_id = problem.SpaceID.IMAGE_LABEL
def generate_data(self, data_dir, tmp_dir, task_id=-1):
generator_utils.generate_dataset_and_shuffle(
self.generator(data_dir, tmp_dir, True),
self.training_filepaths(data_dir, self.train_shards, shuffled=False),
self.generator(data_dir, tmp_dir, False),
self.dev_filepaths(data_dir, self.dev_shards, shuffled=False))
def encode_images_as_png(images):
"""Yield images encoded as pngs."""
if tf.executing_eagerly():
for image in images:
yield tf.image.encode_png(image).numpy()
else:
(height, width, channels) = images[0].shape
with tf.Graph().as_default():
image_t = tf.placeholder(dtype=tf.uint8, shape=(height, width, channels))
encoded_image_t = tf.image.encode_png(image_t)
with tf.Session() as sess:
for image in images:
enc_string = sess.run(encoded_image_t, feed_dict={image_t: image})
yield enc_string
def image_generator(images, labels):
"""Generator for images that takes image and labels lists and creates pngs.
Args:
images: list of images given as [width x height x channels] numpy arrays.
labels: list of ints, same length as images.
Yields:
A dictionary representing the images with the following fields:
* image/encoded: the string encoding the image as PNG,
* image/format: the string "png" representing image format,
* image/class/label: an integer representing the label,
* image/height: an integer representing the height,
* image/width: an integer representing the width.
Every field is actually a singleton list of the corresponding type.
Raises:
ValueError: if images is an empty list.
"""
if not images:
raise ValueError("Must provide some images for the generator.")
width, height, _ = images[0].shape
for (enc_image, label) in zip(encode_images_as_png(images), labels):
yield {
"image/encoded": [enc_image],
"image/format": ["png"],
"image/class/label": [int(label)],
"image/height": [height],
"image/width": [width]
}
class Image2TextProblem(ImageProblem):
"""Base class for image-to-text problems."""
@property
def is_character_level(self):
raise NotImplementedError()
@property
def vocab_problem(self):
raise NotImplementedError() # Not needed if self.is_character_level.
@property
def | target_space_id | identifier_name |
|
image_utils.py | n_channels == 1:
curr_image = np.reshape(curr_image, [height, width])
s = io.BytesIO()
matplotlib_pyplot().imsave(s, curr_image, format="png")
img_sum = tf.Summary.Image(encoded_image_string=s.getvalue(),
height=height, width=width,
colorspace=n_channels)
return tf.Summary.Value(tag=tag, image=img_sum)
def convert_predictions_to_image_summaries(hook_args):
"""Optionally converts images from hooks_args to image summaries.
Args:
hook_args: DecodeHookArgs namedtuple
Returns:
summaries: list of tf.Summary values if hook_args.decode_hpara
"""
decode_hparams = hook_args.decode_hparams
if not decode_hparams.display_decoded_images:
return []
predictions = hook_args.predictions[0]
# Display ten random inputs and outputs so that tensorboard does not hang.
all_summaries = []
rand_predictions = np.random.choice(predictions, size=10)
for ind, prediction in enumerate(rand_predictions):
output_summary = image_to_tf_summary_value(
prediction["outputs"], tag="%d_output" % ind)
input_summary = image_to_tf_summary_value(
prediction["inputs"], tag="%d_input" % ind)
all_summaries.append(input_summary)
all_summaries.append(output_summary)
return all_summaries
def resize_by_area(img, size):
"""image resize function used by quite a few image problems."""
return tf.to_int64(
tf.image.resize_images(img, [size, size], tf.image.ResizeMethod.AREA))
def make_multiscale(image, resolutions,
resize_method=tf.image.ResizeMethod.BICUBIC,
num_channels=3):
"""Returns list of scaled images, one for each resolution.
Args:
image: Tensor of shape [height, height, num_channels].
resolutions: List of heights that image's height is resized to.
resize_method: tf.image.ResizeMethod.
num_channels: Number of channels in image.
Returns:
List of Tensors, one for each resolution with shape given by
[resolutions[i], resolutions[i], num_channels].
"""
scaled_images = []
for height in resolutions:
scaled_image = tf.image.resize_images(
image,
size=[height, height], # assuming that height = width
method=resize_method)
scaled_image = tf.to_int64(scaled_image)
scaled_image.set_shape([height, height, num_channels])
scaled_images.append(scaled_image)
return scaled_images
def make_multiscale_dilated(image, resolutions, num_channels=3):
"""Returns list of scaled images, one for each resolution.
Resizes by skipping every nth pixel.
Args: | assumes VALID padding, so the original image's height must be divisible
by each resolution's height to return the exact resolution size.
num_channels: Number of channels in image.
Returns:
List of Tensors, one for each resolution with shape given by
[resolutions[i], resolutions[i], num_channels] if resolutions properly
divide the original image's height; otherwise shape height and width is up
to valid skips.
"""
image_height = common_layers.shape_list(image)[0]
scaled_images = []
for height in resolutions:
dilation_rate = image_height // height # assuming height = width
scaled_image = image[::dilation_rate, ::dilation_rate]
scaled_image = tf.to_int64(scaled_image)
scaled_image.set_shape([None, None, num_channels])
scaled_images.append(scaled_image)
return scaled_images
class ImageProblem(problem.Problem):
"""Base class for problems with images."""
@property
def num_channels(self):
"""Number of color channels."""
return 3
@property
def vocab_size(self):
"""Number of pixel values."""
return 256
def example_reading_spec(self):
data_fields = {
"image/encoded": tf.FixedLenFeature((), tf.string),
"image/format": tf.FixedLenFeature((), tf.string),
}
data_items_to_decoders = {
"inputs":
tf.contrib.slim.tfexample_decoder.Image(
image_key="image/encoded",
format_key="image/format",
channels=self.num_channels),
}
return data_fields, data_items_to_decoders
def preprocess_example(self, example, mode, hparams):
if not self._was_reversed:
example["inputs"] = tf.image.per_image_standardization(example["inputs"])
return example
def eval_metrics(self):
eval_metrics = [
metrics.Metrics.ACC, metrics.Metrics.ACC_TOP5,
metrics.Metrics.ACC_PER_SEQ, metrics.Metrics.NEG_LOG_PERPLEXITY
]
if self._was_reversed:
eval_metrics += [metrics.Metrics.IMAGE_SUMMARY]
return eval_metrics
@property
def decode_hooks(self):
return [convert_predictions_to_image_summaries]
class Image2ClassProblem(ImageProblem):
"""Base class for image classification problems."""
@property
def is_small(self):
raise NotImplementedError()
@property
def num_classes(self):
raise NotImplementedError()
@property
def train_shards(self):
raise NotImplementedError()
@property
def dev_shards(self):
return 1
@property
def class_labels(self):
return ["ID_%d" % i for i in range(self.num_classes)]
def feature_encoders(self, data_dir):
del data_dir
return {
"inputs": text_encoder.ImageEncoder(channels=self.num_channels),
"targets": text_encoder.ClassLabelEncoder(self.class_labels)
}
def generator(self, data_dir, tmp_dir, is_training):
raise NotImplementedError()
def example_reading_spec(self):
label_key = "image/class/label"
data_fields, data_items_to_decoders = (
super(Image2ClassProblem, self).example_reading_spec())
data_fields[label_key] = tf.FixedLenFeature((1,), tf.int64)
data_items_to_decoders[
"targets"] = tf.contrib.slim.tfexample_decoder.Tensor(label_key)
return data_fields, data_items_to_decoders
def hparams(self, defaults, unused_model_hparams):
p = defaults
p.modality = {"inputs": modalities.ModalityType.IMAGE,
"targets": modalities.ModalityType.CLASS_LABEL}
p.vocab_size = {"inputs": 256,
"targets": self.num_classes}
p.batch_size_multiplier = 4 if self.is_small else 256
p.loss_multiplier = 3.0 if self.is_small else 1.0
if self._was_reversed:
p.loss_multiplier = 1.0
p.input_space_id = problem.SpaceID.IMAGE
p.target_space_id = problem.SpaceID.IMAGE_LABEL
def generate_data(self, data_dir, tmp_dir, task_id=-1):
generator_utils.generate_dataset_and_shuffle(
self.generator(data_dir, tmp_dir, True),
self.training_filepaths(data_dir, self.train_shards, shuffled=False),
self.generator(data_dir, tmp_dir, False),
self.dev_filepaths(data_dir, self.dev_shards, shuffled=False))
def encode_images_as_png(images):
"""Yield images encoded as pngs."""
if tf.executing_eagerly():
for image in images:
yield tf.image.encode_png(image).numpy()
else:
(height, width, channels) = images[0].shape
with tf.Graph().as_default():
image_t = tf.placeholder(dtype=tf.uint8, shape=(height, width, channels))
encoded_image_t = tf.image.encode_png(image_t)
with tf.Session() as sess:
for image in images:
enc_string = sess.run(encoded_image_t, feed_dict={image_t: image})
yield enc_string
def image_generator(images, labels):
"""Generator for images that takes image and labels lists and creates pngs.
Args:
images: list of images given as [width x height x channels] numpy arrays.
labels: list of ints, same length as images.
Yields:
A dictionary representing the images with the following fields:
* image/encoded: the string encoding the image as PNG,
* image/format: the string "png" representing image format,
* image/class/label: an integer representing the label,
* image/height: an integer representing the height,
* image/width: an integer representing the width.
Every field is actually a singleton list of the corresponding type.
Raises:
ValueError: if images is an empty list.
"""
if not images:
raise ValueError("Must provide some images for the generator.")
width, height, _ = images[0].shape
for (enc_image, label) in zip(encode_images_as_png(images), labels):
yield {
"image/encoded": [enc_image],
"image/format": ["png"],
"image/class/label": [int(label)],
"image/height": [height],
"image/width": [width]
}
class Image2TextProblem(ImageProblem):
"""Base class for image-to-text problems."""
@property
def is_character_level(self):
raise NotImplementedError()
@property
def vocab_problem(self):
raise NotImplementedError() # Not needed if self.is_character_level.
@property
def target_space_id(self):
| image: Tensor of shape [height, height, num_channels].
resolutions: List of heights that image's height is resized to. The function | random_line_split |
image_utils.py | n_channels == 1:
curr_image = np.reshape(curr_image, [height, width])
s = io.BytesIO()
matplotlib_pyplot().imsave(s, curr_image, format="png")
img_sum = tf.Summary.Image(encoded_image_string=s.getvalue(),
height=height, width=width,
colorspace=n_channels)
return tf.Summary.Value(tag=tag, image=img_sum)
def convert_predictions_to_image_summaries(hook_args):
"""Optionally converts images from hooks_args to image summaries.
Args:
hook_args: DecodeHookArgs namedtuple
Returns:
summaries: list of tf.Summary values if hook_args.decode_hpara
"""
decode_hparams = hook_args.decode_hparams
if not decode_hparams.display_decoded_images:
return []
predictions = hook_args.predictions[0]
# Display ten random inputs and outputs so that tensorboard does not hang.
all_summaries = []
rand_predictions = np.random.choice(predictions, size=10)
for ind, prediction in enumerate(rand_predictions):
output_summary = image_to_tf_summary_value(
prediction["outputs"], tag="%d_output" % ind)
input_summary = image_to_tf_summary_value(
prediction["inputs"], tag="%d_input" % ind)
all_summaries.append(input_summary)
all_summaries.append(output_summary)
return all_summaries
def resize_by_area(img, size):
"""image resize function used by quite a few image problems."""
return tf.to_int64(
tf.image.resize_images(img, [size, size], tf.image.ResizeMethod.AREA))
def make_multiscale(image, resolutions,
resize_method=tf.image.ResizeMethod.BICUBIC,
num_channels=3):
"""Returns list of scaled images, one for each resolution.
Args:
image: Tensor of shape [height, height, num_channels].
resolutions: List of heights that image's height is resized to.
resize_method: tf.image.ResizeMethod.
num_channels: Number of channels in image.
Returns:
List of Tensors, one for each resolution with shape given by
[resolutions[i], resolutions[i], num_channels].
"""
scaled_images = []
for height in resolutions:
scaled_image = tf.image.resize_images(
image,
size=[height, height], # assuming that height = width
method=resize_method)
scaled_image = tf.to_int64(scaled_image)
scaled_image.set_shape([height, height, num_channels])
scaled_images.append(scaled_image)
return scaled_images
def make_multiscale_dilated(image, resolutions, num_channels=3):
"""Returns list of scaled images, one for each resolution.
Resizes by skipping every nth pixel.
Args:
image: Tensor of shape [height, height, num_channels].
resolutions: List of heights that image's height is resized to. The function
assumes VALID padding, so the original image's height must be divisible
by each resolution's height to return the exact resolution size.
num_channels: Number of channels in image.
Returns:
List of Tensors, one for each resolution with shape given by
[resolutions[i], resolutions[i], num_channels] if resolutions properly
divide the original image's height; otherwise shape height and width is up
to valid skips.
"""
image_height = common_layers.shape_list(image)[0]
scaled_images = []
for height in resolutions:
dilation_rate = image_height // height # assuming height = width
scaled_image = image[::dilation_rate, ::dilation_rate]
scaled_image = tf.to_int64(scaled_image)
scaled_image.set_shape([None, None, num_channels])
scaled_images.append(scaled_image)
return scaled_images
class ImageProblem(problem.Problem):
"""Base class for problems with images."""
@property
def num_channels(self):
"""Number of color channels."""
return 3
@property
def vocab_size(self):
"""Number of pixel values."""
return 256
def example_reading_spec(self):
data_fields = {
"image/encoded": tf.FixedLenFeature((), tf.string),
"image/format": tf.FixedLenFeature((), tf.string),
}
data_items_to_decoders = {
"inputs":
tf.contrib.slim.tfexample_decoder.Image(
image_key="image/encoded",
format_key="image/format",
channels=self.num_channels),
}
return data_fields, data_items_to_decoders
def preprocess_example(self, example, mode, hparams):
if not self._was_reversed:
example["inputs"] = tf.image.per_image_standardization(example["inputs"])
return example
def eval_metrics(self):
eval_metrics = [
metrics.Metrics.ACC, metrics.Metrics.ACC_TOP5,
metrics.Metrics.ACC_PER_SEQ, metrics.Metrics.NEG_LOG_PERPLEXITY
]
if self._was_reversed:
eval_metrics += [metrics.Metrics.IMAGE_SUMMARY]
return eval_metrics
@property
def decode_hooks(self):
return [convert_predictions_to_image_summaries]
class Image2ClassProblem(ImageProblem):
"""Base class for image classification problems."""
@property
def is_small(self):
raise NotImplementedError()
@property
def num_classes(self):
raise NotImplementedError()
@property
def train_shards(self):
raise NotImplementedError()
@property
def dev_shards(self):
return 1
@property
def class_labels(self):
return ["ID_%d" % i for i in range(self.num_classes)]
def feature_encoders(self, data_dir):
del data_dir
return {
"inputs": text_encoder.ImageEncoder(channels=self.num_channels),
"targets": text_encoder.ClassLabelEncoder(self.class_labels)
}
def generator(self, data_dir, tmp_dir, is_training):
raise NotImplementedError()
def example_reading_spec(self):
label_key = "image/class/label"
data_fields, data_items_to_decoders = (
super(Image2ClassProblem, self).example_reading_spec())
data_fields[label_key] = tf.FixedLenFeature((1,), tf.int64)
data_items_to_decoders[
"targets"] = tf.contrib.slim.tfexample_decoder.Tensor(label_key)
return data_fields, data_items_to_decoders
def hparams(self, defaults, unused_model_hparams):
p = defaults
p.modality = {"inputs": modalities.ModalityType.IMAGE,
"targets": modalities.ModalityType.CLASS_LABEL}
p.vocab_size = {"inputs": 256,
"targets": self.num_classes}
p.batch_size_multiplier = 4 if self.is_small else 256
p.loss_multiplier = 3.0 if self.is_small else 1.0
if self._was_reversed:
p.loss_multiplier = 1.0
p.input_space_id = problem.SpaceID.IMAGE
p.target_space_id = problem.SpaceID.IMAGE_LABEL
def generate_data(self, data_dir, tmp_dir, task_id=-1):
generator_utils.generate_dataset_and_shuffle(
self.generator(data_dir, tmp_dir, True),
self.training_filepaths(data_dir, self.train_shards, shuffled=False),
self.generator(data_dir, tmp_dir, False),
self.dev_filepaths(data_dir, self.dev_shards, shuffled=False))
def encode_images_as_png(images):
|
def image_generator(images, labels):
"""Generator for images that takes image and labels lists and creates pngs.
Args:
images: list of images given as [width x height x channels] numpy arrays.
labels: list of ints, same length as images.
Yields:
A dictionary representing the images with the following fields:
* image/encoded: the string encoding the image as PNG,
* image/format: the string "png" representing image format,
* image/class/label: an integer representing the label,
* image/height: an integer representing the height,
* image/width: an integer representing the width.
Every field is actually a singleton list of the corresponding type.
Raises:
ValueError: if images is an empty list.
"""
if not images:
raise ValueError("Must provide some images for the generator.")
width, height, _ = images[0].shape
for (enc_image, label) in zip(encode_images_as_png(images), labels):
yield {
"image/encoded": [enc_image],
"image/format": ["png"],
"image/class/label": [int(label)],
"image/height": [height],
"image/width": [width]
}
class Image2TextProblem(ImageProblem):
"""Base class for image-to-text problems."""
@property
def is_character_level(self):
raise NotImplementedError()
@property
def vocab_problem(self):
raise NotImplementedError() # Not needed if self.is_character_level.
@property
def target_space_id(self | """Yield images encoded as pngs."""
if tf.executing_eagerly():
for image in images:
yield tf.image.encode_png(image).numpy()
else:
(height, width, channels) = images[0].shape
with tf.Graph().as_default():
image_t = tf.placeholder(dtype=tf.uint8, shape=(height, width, channels))
encoded_image_t = tf.image.encode_png(image_t)
with tf.Session() as sess:
for image in images:
enc_string = sess.run(encoded_image_t, feed_dict={image_t: image})
yield enc_string | identifier_body |
congestion.go | ) {
}
func (cc *CongestionControllerDummy) onAckReceived(acks ackRanges, delay time.Duration) {
}
func (cc *CongestionControllerDummy) bytesAllowedToSend() int {
/* return the the maximum int value */
return int(^uint(0) >> 1)
}
func (cc *CongestionControllerDummy) setLostPacketHandler(handler func(pn uint64)) {
}
func (cc *CongestionControllerDummy) rto() time.Duration {
return kMinRTOTimeout
}
/*
* draft-ietf-quic-recovery congestion controller
*/
type CongestionControllerIetf struct {
// Congestion control related
bytesInFlight int
congestionWindow int
endOfRecovery uint64
sstresh int
// Loss detection related
lossDetectionAlarm int //TODO([email protected]) set this to the right type
handshakeCount int
tlpCount int
rtoCount int
largestSendBeforeRto uint64
timeOfLastSentPacket time.Time
largestSendPacket uint64
largestAckedPacket uint64
maxAckDelay time.Duration
minRtt time.Duration
// largestRtt time.Duration
smoothedRtt time.Duration
rttVar time.Duration
smoothedRttTcp time.Duration
rttVarTcp time.Duration
reorderingThreshold int
timeReorderingFraction float32
lossTime time.Time
sentPackets map[uint64]packetEntry
// others
lostPacketHandler func(pn uint64)
conn *Connection
}
type packetEntry struct {
pn uint64
txTime time.Time
bytes int
ackOnly bool
}
func (cc *CongestionControllerIetf) onPacketSent(pn uint64, isAckOnly bool, sentBytes int) {
cc.timeOfLastSentPacket = time.Now()
cc.largestSendPacket = pn
packetData := packetEntry{pn, time.Now(), 0, isAckOnly}
cc.conn.log(logTypeCongestion, "Packet send pn: %d len:%d ackonly: %v\n", pn, sentBytes, isAckOnly)
if !isAckOnly {
cc.onPacketSentCC(sentBytes)
packetData.bytes = sentBytes
cc.setLossDetectionAlarm()
}
cc.sentPackets[pn] = packetData
}
// acks is received to be a sorted list, where the largest packet numbers are at the beginning
func (cc *CongestionControllerIetf) onAckReceived(acks ackRanges, ackDelay time.Duration) {
// keep track of largest packet acked overall
if acks[0].lastPacket > cc.largestAckedPacket {
cc.largestAckedPacket = acks[0].lastPacket
}
// If the largest acked is newly acked update rtt
lastPacket, present := cc.sentPackets[acks[0].lastPacket]
if present {
latestRtt := time.Since(cc.sentPackets[acks[0].lastPacket].txTime)
cc.conn.log(logTypeCongestion, "latestRtt: %v, ackDelay: %v", latestRtt, ackDelay)
cc.updateRttTcp(latestRtt)
// Update the minRtt, but ignore ackDelay.
if latestRtt < cc.minRtt {
cc.minRtt = latestRtt
}
// Now reduce by ackDelay if it doesn't reduce the RTT below the minimum.
if latestRtt-cc.minRtt > ackDelay {
latestRtt -= ackDelay
// And update the maximum observed ACK delay.
if !lastPacket.ackOnly && ackDelay > cc.maxAckDelay {
cc.maxAckDelay = ackDelay
}
}
cc.updateRtt(latestRtt)
}
// find and proccess newly acked packets
for _, ackBlock := range acks {
for pn := ackBlock.lastPacket; pn > (ackBlock.lastPacket - ackBlock.count); pn-- {
cc.conn.log(logTypeCongestion, "Ack for pn %d received", pn)
_, present := cc.sentPackets[pn]
if present {
cc.conn.log(logTypeCongestion, "First ack for pn %d received", pn)
cc.onPacketAcked(pn)
}
}
}
cc.detectLostPackets()
cc.setLossDetectionAlarm()
}
func (cc *CongestionControllerIetf) setLostPacketHandler(handler func(pn uint64)) {
cc.lostPacketHandler = handler
}
func (cc *CongestionControllerIetf) updateRtt(latestRtt time.Duration) {
if cc.smoothedRtt == 0 {
cc.smoothedRtt = latestRtt
cc.rttVar = time.Duration(int64(latestRtt) / 2) | } else {
rttDelta := cc.smoothedRtt - latestRtt
if rttDelta < 0 {
rttDelta = -rttDelta
}
cc.rttVar = time.Duration(int64(cc.rttVar)*3/4 + int64(rttDelta)*1/4)
cc.smoothedRtt = time.Duration(int64(cc.smoothedRtt)*7/8 + int64(latestRtt)*1/8)
}
cc.conn.log(logTypeCongestion, "New RTT estimate: %v, variance: %v", cc.smoothedRtt, cc.rttVar)
}
func (cc *CongestionControllerIetf) updateRttTcp(latestRtt time.Duration) {
if cc.smoothedRttTcp == 0 {
cc.smoothedRttTcp = latestRtt
cc.rttVarTcp = time.Duration(int64(latestRtt) / 2)
} else {
rttDelta := cc.smoothedRttTcp - latestRtt
if rttDelta < 0 {
rttDelta = -rttDelta
}
cc.rttVarTcp = time.Duration(int64(cc.rttVarTcp)*3/4 + int64(rttDelta)*3/4)
cc.smoothedRttTcp = time.Duration(int64(cc.smoothedRttTcp)*7/8 + int64(latestRtt)*1/8)
}
cc.conn.log(logTypeCongestion, "New RTT(TCP) estimate: %v, variance: %v", cc.smoothedRttTcp, cc.rttVarTcp)
}
func (cc *CongestionControllerIetf) rto() time.Duration {
// max(SRTT + 4*RTTVAR + MaxAckDelay, minRTO)
rto := cc.smoothedRtt + 4*cc.rttVar + cc.maxAckDelay
if rto < kMinRTOTimeout {
return kMinRTOTimeout
}
return rto
}
func (cc *CongestionControllerIetf) onPacketAcked(pn uint64) {
cc.onPacketAckedCC(pn)
//TODO([email protected]) some RTO stuff here
delete(cc.sentPackets, pn)
}
func (cc *CongestionControllerIetf) setLossDetectionAlarm() {
//TODO([email protected])
}
func (cc *CongestionControllerIetf) onLossDetectionAlarm() {
//TODO([email protected])
}
func (cc *CongestionControllerIetf) detectLostPackets() {
var lostPackets []packetEntry
//TODO([email protected]) implement loss detection different from reorderingThreshold
for _, packet := range cc.sentPackets {
if (cc.largestAckedPacket > packet.pn) &&
(cc.largestAckedPacket-packet.pn > uint64(cc.reorderingThreshold)) {
lostPackets = append(lostPackets, packet)
}
}
if len(lostPackets) > 0 {
cc.onPacketsLost(lostPackets)
}
for _, packet := range lostPackets {
delete(cc.sentPackets, packet.pn)
}
}
func (cc *CongestionControllerIetf) onPacketSentCC(bytes_sent int) {
cc.bytesInFlight += bytes_sent
cc.conn.log(logTypeCongestion, "%d bytes added to bytesInFlight", bytes_sent)
}
func (cc *CongestionControllerIetf) onPacketAckedCC(pn uint64) {
cc.bytesInFlight -= cc.sentPackets[pn].bytes
cc.conn.log(logTypeCongestion, "%d bytes from packet %d removed from bytesInFlight", cc.sentPackets[pn].bytes, pn)
if pn < cc.endOfRecovery {
// Do not increase window size during recovery
return
}
if cc.congestionWindow < cc.sstresh {
// Slow start
cc.congestionWindow += cc.sentPackets[pn].bytes
cc.conn.log(logTypeCongestion, "PDV Slow Start: increasing window size with %d bytes to %d",
cc.sentPackets[pn].bytes, cc.congestionWindow)
} else {
// Congestion avoidance | random_line_split |
|
congestion.go | {
}
func (cc *CongestionControllerDummy) onAckReceived(acks ackRanges, delay time.Duration) {
}
func (cc *CongestionControllerDummy) bytesAllowedToSend() int {
/* return the the maximum int value */
return int(^uint(0) >> 1)
}
func (cc *CongestionControllerDummy) setLostPacketHandler(handler func(pn uint64)) {
}
func (cc *CongestionControllerDummy) rto() time.Duration |
/*
* draft-ietf-quic-recovery congestion controller
*/
type CongestionControllerIetf struct {
// Congestion control related
bytesInFlight int
congestionWindow int
endOfRecovery uint64
sstresh int
// Loss detection related
lossDetectionAlarm int //TODO([email protected]) set this to the right type
handshakeCount int
tlpCount int
rtoCount int
largestSendBeforeRto uint64
timeOfLastSentPacket time.Time
largestSendPacket uint64
largestAckedPacket uint64
maxAckDelay time.Duration
minRtt time.Duration
// largestRtt time.Duration
smoothedRtt time.Duration
rttVar time.Duration
smoothedRttTcp time.Duration
rttVarTcp time.Duration
reorderingThreshold int
timeReorderingFraction float32
lossTime time.Time
sentPackets map[uint64]packetEntry
// others
lostPacketHandler func(pn uint64)
conn *Connection
}
type packetEntry struct {
pn uint64
txTime time.Time
bytes int
ackOnly bool
}
func (cc *CongestionControllerIetf) onPacketSent(pn uint64, isAckOnly bool, sentBytes int) {
cc.timeOfLastSentPacket = time.Now()
cc.largestSendPacket = pn
packetData := packetEntry{pn, time.Now(), 0, isAckOnly}
cc.conn.log(logTypeCongestion, "Packet send pn: %d len:%d ackonly: %v\n", pn, sentBytes, isAckOnly)
if !isAckOnly {
cc.onPacketSentCC(sentBytes)
packetData.bytes = sentBytes
cc.setLossDetectionAlarm()
}
cc.sentPackets[pn] = packetData
}
// acks is received to be a sorted list, where the largest packet numbers are at the beginning
func (cc *CongestionControllerIetf) onAckReceived(acks ackRanges, ackDelay time.Duration) {
// keep track of largest packet acked overall
if acks[0].lastPacket > cc.largestAckedPacket {
cc.largestAckedPacket = acks[0].lastPacket
}
// If the largest acked is newly acked update rtt
lastPacket, present := cc.sentPackets[acks[0].lastPacket]
if present {
latestRtt := time.Since(cc.sentPackets[acks[0].lastPacket].txTime)
cc.conn.log(logTypeCongestion, "latestRtt: %v, ackDelay: %v", latestRtt, ackDelay)
cc.updateRttTcp(latestRtt)
// Update the minRtt, but ignore ackDelay.
if latestRtt < cc.minRtt {
cc.minRtt = latestRtt
}
// Now reduce by ackDelay if it doesn't reduce the RTT below the minimum.
if latestRtt-cc.minRtt > ackDelay {
latestRtt -= ackDelay
// And update the maximum observed ACK delay.
if !lastPacket.ackOnly && ackDelay > cc.maxAckDelay {
cc.maxAckDelay = ackDelay
}
}
cc.updateRtt(latestRtt)
}
// find and proccess newly acked packets
for _, ackBlock := range acks {
for pn := ackBlock.lastPacket; pn > (ackBlock.lastPacket - ackBlock.count); pn-- {
cc.conn.log(logTypeCongestion, "Ack for pn %d received", pn)
_, present := cc.sentPackets[pn]
if present {
cc.conn.log(logTypeCongestion, "First ack for pn %d received", pn)
cc.onPacketAcked(pn)
}
}
}
cc.detectLostPackets()
cc.setLossDetectionAlarm()
}
func (cc *CongestionControllerIetf) setLostPacketHandler(handler func(pn uint64)) {
cc.lostPacketHandler = handler
}
func (cc *CongestionControllerIetf) updateRtt(latestRtt time.Duration) {
if cc.smoothedRtt == 0 {
cc.smoothedRtt = latestRtt
cc.rttVar = time.Duration(int64(latestRtt) / 2)
} else {
rttDelta := cc.smoothedRtt - latestRtt
if rttDelta < 0 {
rttDelta = -rttDelta
}
cc.rttVar = time.Duration(int64(cc.rttVar)*3/4 + int64(rttDelta)*1/4)
cc.smoothedRtt = time.Duration(int64(cc.smoothedRtt)*7/8 + int64(latestRtt)*1/8)
}
cc.conn.log(logTypeCongestion, "New RTT estimate: %v, variance: %v", cc.smoothedRtt, cc.rttVar)
}
func (cc *CongestionControllerIetf) updateRttTcp(latestRtt time.Duration) {
if cc.smoothedRttTcp == 0 {
cc.smoothedRttTcp = latestRtt
cc.rttVarTcp = time.Duration(int64(latestRtt) / 2)
} else {
rttDelta := cc.smoothedRttTcp - latestRtt
if rttDelta < 0 {
rttDelta = -rttDelta
}
cc.rttVarTcp = time.Duration(int64(cc.rttVarTcp)*3/4 + int64(rttDelta)*3/4)
cc.smoothedRttTcp = time.Duration(int64(cc.smoothedRttTcp)*7/8 + int64(latestRtt)*1/8)
}
cc.conn.log(logTypeCongestion, "New RTT(TCP) estimate: %v, variance: %v", cc.smoothedRttTcp, cc.rttVarTcp)
}
func (cc *CongestionControllerIetf) rto() time.Duration {
// max(SRTT + 4*RTTVAR + MaxAckDelay, minRTO)
rto := cc.smoothedRtt + 4*cc.rttVar + cc.maxAckDelay
if rto < kMinRTOTimeout {
return kMinRTOTimeout
}
return rto
}
func (cc *CongestionControllerIetf) onPacketAcked(pn uint64) {
cc.onPacketAckedCC(pn)
//TODO([email protected]) some RTO stuff here
delete(cc.sentPackets, pn)
}
func (cc *CongestionControllerIetf) setLossDetectionAlarm() {
//TODO([email protected])
}
func (cc *CongestionControllerIetf) onLossDetectionAlarm() {
//TODO([email protected])
}
func (cc *CongestionControllerIetf) detectLostPackets() {
var lostPackets []packetEntry
//TODO([email protected]) implement loss detection different from reorderingThreshold
for _, packet := range cc.sentPackets {
if (cc.largestAckedPacket > packet.pn) &&
(cc.largestAckedPacket-packet.pn > uint64(cc.reorderingThreshold)) {
lostPackets = append(lostPackets, packet)
}
}
if len(lostPackets) > 0 {
cc.onPacketsLost(lostPackets)
}
for _, packet := range lostPackets {
delete(cc.sentPackets, packet.pn)
}
}
func (cc *CongestionControllerIetf) onPacketSentCC(bytes_sent int) {
cc.bytesInFlight += bytes_sent
cc.conn.log(logTypeCongestion, "%d bytes added to bytesInFlight", bytes_sent)
}
func (cc *CongestionControllerIetf) onPacketAckedCC(pn uint64) {
cc.bytesInFlight -= cc.sentPackets[pn].bytes
cc.conn.log(logTypeCongestion, "%d bytes from packet %d removed from bytesInFlight", cc.sentPackets[pn].bytes, pn)
if pn < cc.endOfRecovery {
// Do not increase window size during recovery
return
}
if cc.congestionWindow < cc.sstresh {
// Slow start
cc.congestionWindow += cc.sentPackets[pn].bytes
cc.conn.log(logTypeCongestion, "PDV Slow Start: increasing window size with %d bytes to %d",
cc.sentPackets[pn].bytes, cc.congestionWindow)
} else {
// Congestion | {
return kMinRTOTimeout
} | identifier_body |
congestion.go | ) {
}
func (cc *CongestionControllerDummy) onAckReceived(acks ackRanges, delay time.Duration) {
}
func (cc *CongestionControllerDummy) bytesAllowedToSend() int {
/* return the the maximum int value */
return int(^uint(0) >> 1)
}
func (cc *CongestionControllerDummy) setLostPacketHandler(handler func(pn uint64)) {
}
func (cc *CongestionControllerDummy) rto() time.Duration {
return kMinRTOTimeout
}
/*
* draft-ietf-quic-recovery congestion controller
*/
type CongestionControllerIetf struct {
// Congestion control related
bytesInFlight int
congestionWindow int
endOfRecovery uint64
sstresh int
// Loss detection related
lossDetectionAlarm int //TODO([email protected]) set this to the right type
handshakeCount int
tlpCount int
rtoCount int
largestSendBeforeRto uint64
timeOfLastSentPacket time.Time
largestSendPacket uint64
largestAckedPacket uint64
maxAckDelay time.Duration
minRtt time.Duration
// largestRtt time.Duration
smoothedRtt time.Duration
rttVar time.Duration
smoothedRttTcp time.Duration
rttVarTcp time.Duration
reorderingThreshold int
timeReorderingFraction float32
lossTime time.Time
sentPackets map[uint64]packetEntry
// others
lostPacketHandler func(pn uint64)
conn *Connection
}
type packetEntry struct {
pn uint64
txTime time.Time
bytes int
ackOnly bool
}
func (cc *CongestionControllerIetf) onPacketSent(pn uint64, isAckOnly bool, sentBytes int) {
cc.timeOfLastSentPacket = time.Now()
cc.largestSendPacket = pn
packetData := packetEntry{pn, time.Now(), 0, isAckOnly}
cc.conn.log(logTypeCongestion, "Packet send pn: %d len:%d ackonly: %v\n", pn, sentBytes, isAckOnly)
if !isAckOnly {
cc.onPacketSentCC(sentBytes)
packetData.bytes = sentBytes
cc.setLossDetectionAlarm()
}
cc.sentPackets[pn] = packetData
}
// acks is received to be a sorted list, where the largest packet numbers are at the beginning
func (cc *CongestionControllerIetf) onAckReceived(acks ackRanges, ackDelay time.Duration) {
// keep track of largest packet acked overall
if acks[0].lastPacket > cc.largestAckedPacket {
cc.largestAckedPacket = acks[0].lastPacket
}
// If the largest acked is newly acked update rtt
lastPacket, present := cc.sentPackets[acks[0].lastPacket]
if present {
latestRtt := time.Since(cc.sentPackets[acks[0].lastPacket].txTime)
cc.conn.log(logTypeCongestion, "latestRtt: %v, ackDelay: %v", latestRtt, ackDelay)
cc.updateRttTcp(latestRtt)
// Update the minRtt, but ignore ackDelay.
if latestRtt < cc.minRtt {
cc.minRtt = latestRtt
}
// Now reduce by ackDelay if it doesn't reduce the RTT below the minimum.
if latestRtt-cc.minRtt > ackDelay {
latestRtt -= ackDelay
// And update the maximum observed ACK delay.
if !lastPacket.ackOnly && ackDelay > cc.maxAckDelay {
cc.maxAckDelay = ackDelay
}
}
cc.updateRtt(latestRtt)
}
// find and proccess newly acked packets
for _, ackBlock := range acks {
for pn := ackBlock.lastPacket; pn > (ackBlock.lastPacket - ackBlock.count); pn-- {
cc.conn.log(logTypeCongestion, "Ack for pn %d received", pn)
_, present := cc.sentPackets[pn]
if present {
cc.conn.log(logTypeCongestion, "First ack for pn %d received", pn)
cc.onPacketAcked(pn)
}
}
}
cc.detectLostPackets()
cc.setLossDetectionAlarm()
}
func (cc *CongestionControllerIetf) setLostPacketHandler(handler func(pn uint64)) {
cc.lostPacketHandler = handler
}
func (cc *CongestionControllerIetf) updateRtt(latestRtt time.Duration) {
if cc.smoothedRtt == 0 {
cc.smoothedRtt = latestRtt
cc.rttVar = time.Duration(int64(latestRtt) / 2)
} else {
rttDelta := cc.smoothedRtt - latestRtt
if rttDelta < 0 {
rttDelta = -rttDelta
}
cc.rttVar = time.Duration(int64(cc.rttVar)*3/4 + int64(rttDelta)*1/4)
cc.smoothedRtt = time.Duration(int64(cc.smoothedRtt)*7/8 + int64(latestRtt)*1/8)
}
cc.conn.log(logTypeCongestion, "New RTT estimate: %v, variance: %v", cc.smoothedRtt, cc.rttVar)
}
func (cc *CongestionControllerIetf) updateRttTcp(latestRtt time.Duration) {
if cc.smoothedRttTcp == 0 {
cc.smoothedRttTcp = latestRtt
cc.rttVarTcp = time.Duration(int64(latestRtt) / 2)
} else {
rttDelta := cc.smoothedRttTcp - latestRtt
if rttDelta < 0 {
rttDelta = -rttDelta
}
cc.rttVarTcp = time.Duration(int64(cc.rttVarTcp)*3/4 + int64(rttDelta)*3/4)
cc.smoothedRttTcp = time.Duration(int64(cc.smoothedRttTcp)*7/8 + int64(latestRtt)*1/8)
}
cc.conn.log(logTypeCongestion, "New RTT(TCP) estimate: %v, variance: %v", cc.smoothedRttTcp, cc.rttVarTcp)
}
func (cc *CongestionControllerIetf) rto() time.Duration {
// max(SRTT + 4*RTTVAR + MaxAckDelay, minRTO)
rto := cc.smoothedRtt + 4*cc.rttVar + cc.maxAckDelay
if rto < kMinRTOTimeout {
return kMinRTOTimeout
}
return rto
}
func (cc *CongestionControllerIetf) onPacketAcked(pn uint64) {
cc.onPacketAckedCC(pn)
//TODO([email protected]) some RTO stuff here
delete(cc.sentPackets, pn)
}
func (cc *CongestionControllerIetf) setLossDetectionAlarm() {
//TODO([email protected])
}
func (cc *CongestionControllerIetf) | () {
//TODO([email protected])
}
func (cc *CongestionControllerIetf) detectLostPackets() {
var lostPackets []packetEntry
//TODO([email protected]) implement loss detection different from reorderingThreshold
for _, packet := range cc.sentPackets {
if (cc.largestAckedPacket > packet.pn) &&
(cc.largestAckedPacket-packet.pn > uint64(cc.reorderingThreshold)) {
lostPackets = append(lostPackets, packet)
}
}
if len(lostPackets) > 0 {
cc.onPacketsLost(lostPackets)
}
for _, packet := range lostPackets {
delete(cc.sentPackets, packet.pn)
}
}
func (cc *CongestionControllerIetf) onPacketSentCC(bytes_sent int) {
cc.bytesInFlight += bytes_sent
cc.conn.log(logTypeCongestion, "%d bytes added to bytesInFlight", bytes_sent)
}
func (cc *CongestionControllerIetf) onPacketAckedCC(pn uint64) {
cc.bytesInFlight -= cc.sentPackets[pn].bytes
cc.conn.log(logTypeCongestion, "%d bytes from packet %d removed from bytesInFlight", cc.sentPackets[pn].bytes, pn)
if pn < cc.endOfRecovery {
// Do not increase window size during recovery
return
}
if cc.congestionWindow < cc.sstresh {
// Slow start
cc.congestionWindow += cc.sentPackets[pn].bytes
cc.conn.log(logTypeCongestion, "PDV Slow Start: increasing window size with %d bytes to %d",
cc.sentPackets[pn].bytes, cc.congestionWindow)
} else {
// Congestion | onLossDetectionAlarm | identifier_name |
congestion.go | ) {
}
func (cc *CongestionControllerDummy) onAckReceived(acks ackRanges, delay time.Duration) {
}
func (cc *CongestionControllerDummy) bytesAllowedToSend() int {
/* return the the maximum int value */
return int(^uint(0) >> 1)
}
func (cc *CongestionControllerDummy) setLostPacketHandler(handler func(pn uint64)) {
}
func (cc *CongestionControllerDummy) rto() time.Duration {
return kMinRTOTimeout
}
/*
* draft-ietf-quic-recovery congestion controller
*/
type CongestionControllerIetf struct {
// Congestion control related
bytesInFlight int
congestionWindow int
endOfRecovery uint64
sstresh int
// Loss detection related
lossDetectionAlarm int //TODO([email protected]) set this to the right type
handshakeCount int
tlpCount int
rtoCount int
largestSendBeforeRto uint64
timeOfLastSentPacket time.Time
largestSendPacket uint64
largestAckedPacket uint64
maxAckDelay time.Duration
minRtt time.Duration
// largestRtt time.Duration
smoothedRtt time.Duration
rttVar time.Duration
smoothedRttTcp time.Duration
rttVarTcp time.Duration
reorderingThreshold int
timeReorderingFraction float32
lossTime time.Time
sentPackets map[uint64]packetEntry
// others
lostPacketHandler func(pn uint64)
conn *Connection
}
type packetEntry struct {
pn uint64
txTime time.Time
bytes int
ackOnly bool
}
func (cc *CongestionControllerIetf) onPacketSent(pn uint64, isAckOnly bool, sentBytes int) {
cc.timeOfLastSentPacket = time.Now()
cc.largestSendPacket = pn
packetData := packetEntry{pn, time.Now(), 0, isAckOnly}
cc.conn.log(logTypeCongestion, "Packet send pn: %d len:%d ackonly: %v\n", pn, sentBytes, isAckOnly)
if !isAckOnly {
cc.onPacketSentCC(sentBytes)
packetData.bytes = sentBytes
cc.setLossDetectionAlarm()
}
cc.sentPackets[pn] = packetData
}
// acks is received to be a sorted list, where the largest packet numbers are at the beginning
func (cc *CongestionControllerIetf) onAckReceived(acks ackRanges, ackDelay time.Duration) {
// keep track of largest packet acked overall
if acks[0].lastPacket > cc.largestAckedPacket {
cc.largestAckedPacket = acks[0].lastPacket
}
// If the largest acked is newly acked update rtt
lastPacket, present := cc.sentPackets[acks[0].lastPacket]
if present {
latestRtt := time.Since(cc.sentPackets[acks[0].lastPacket].txTime)
cc.conn.log(logTypeCongestion, "latestRtt: %v, ackDelay: %v", latestRtt, ackDelay)
cc.updateRttTcp(latestRtt)
// Update the minRtt, but ignore ackDelay.
if latestRtt < cc.minRtt {
cc.minRtt = latestRtt
}
// Now reduce by ackDelay if it doesn't reduce the RTT below the minimum.
if latestRtt-cc.minRtt > ackDelay {
latestRtt -= ackDelay
// And update the maximum observed ACK delay.
if !lastPacket.ackOnly && ackDelay > cc.maxAckDelay {
cc.maxAckDelay = ackDelay
}
}
cc.updateRtt(latestRtt)
}
// find and proccess newly acked packets
for _, ackBlock := range acks {
for pn := ackBlock.lastPacket; pn > (ackBlock.lastPacket - ackBlock.count); pn-- {
cc.conn.log(logTypeCongestion, "Ack for pn %d received", pn)
_, present := cc.sentPackets[pn]
if present {
cc.conn.log(logTypeCongestion, "First ack for pn %d received", pn)
cc.onPacketAcked(pn)
}
}
}
cc.detectLostPackets()
cc.setLossDetectionAlarm()
}
func (cc *CongestionControllerIetf) setLostPacketHandler(handler func(pn uint64)) {
cc.lostPacketHandler = handler
}
func (cc *CongestionControllerIetf) updateRtt(latestRtt time.Duration) {
if cc.smoothedRtt == 0 | else {
rttDelta := cc.smoothedRtt - latestRtt
if rttDelta < 0 {
rttDelta = -rttDelta
}
cc.rttVar = time.Duration(int64(cc.rttVar)*3/4 + int64(rttDelta)*1/4)
cc.smoothedRtt = time.Duration(int64(cc.smoothedRtt)*7/8 + int64(latestRtt)*1/8)
}
cc.conn.log(logTypeCongestion, "New RTT estimate: %v, variance: %v", cc.smoothedRtt, cc.rttVar)
}
func (cc *CongestionControllerIetf) updateRttTcp(latestRtt time.Duration) {
if cc.smoothedRttTcp == 0 {
cc.smoothedRttTcp = latestRtt
cc.rttVarTcp = time.Duration(int64(latestRtt) / 2)
} else {
rttDelta := cc.smoothedRttTcp - latestRtt
if rttDelta < 0 {
rttDelta = -rttDelta
}
cc.rttVarTcp = time.Duration(int64(cc.rttVarTcp)*3/4 + int64(rttDelta)*3/4)
cc.smoothedRttTcp = time.Duration(int64(cc.smoothedRttTcp)*7/8 + int64(latestRtt)*1/8)
}
cc.conn.log(logTypeCongestion, "New RTT(TCP) estimate: %v, variance: %v", cc.smoothedRttTcp, cc.rttVarTcp)
}
func (cc *CongestionControllerIetf) rto() time.Duration {
// max(SRTT + 4*RTTVAR + MaxAckDelay, minRTO)
rto := cc.smoothedRtt + 4*cc.rttVar + cc.maxAckDelay
if rto < kMinRTOTimeout {
return kMinRTOTimeout
}
return rto
}
func (cc *CongestionControllerIetf) onPacketAcked(pn uint64) {
cc.onPacketAckedCC(pn)
//TODO([email protected]) some RTO stuff here
delete(cc.sentPackets, pn)
}
func (cc *CongestionControllerIetf) setLossDetectionAlarm() {
//TODO([email protected])
}
func (cc *CongestionControllerIetf) onLossDetectionAlarm() {
//TODO([email protected])
}
func (cc *CongestionControllerIetf) detectLostPackets() {
var lostPackets []packetEntry
//TODO([email protected]) implement loss detection different from reorderingThreshold
for _, packet := range cc.sentPackets {
if (cc.largestAckedPacket > packet.pn) &&
(cc.largestAckedPacket-packet.pn > uint64(cc.reorderingThreshold)) {
lostPackets = append(lostPackets, packet)
}
}
if len(lostPackets) > 0 {
cc.onPacketsLost(lostPackets)
}
for _, packet := range lostPackets {
delete(cc.sentPackets, packet.pn)
}
}
func (cc *CongestionControllerIetf) onPacketSentCC(bytes_sent int) {
cc.bytesInFlight += bytes_sent
cc.conn.log(logTypeCongestion, "%d bytes added to bytesInFlight", bytes_sent)
}
func (cc *CongestionControllerIetf) onPacketAckedCC(pn uint64) {
cc.bytesInFlight -= cc.sentPackets[pn].bytes
cc.conn.log(logTypeCongestion, "%d bytes from packet %d removed from bytesInFlight", cc.sentPackets[pn].bytes, pn)
if pn < cc.endOfRecovery {
// Do not increase window size during recovery
return
}
if cc.congestionWindow < cc.sstresh {
// Slow start
cc.congestionWindow += cc.sentPackets[pn].bytes
cc.conn.log(logTypeCongestion, "PDV Slow Start: increasing window size with %d bytes to %d",
cc.sentPackets[pn].bytes, cc.congestionWindow)
} else {
// Congestion | {
cc.smoothedRtt = latestRtt
cc.rttVar = time.Duration(int64(latestRtt) / 2)
} | conditional_block |
Leanote4MD.py | ['Ok']:
return True
else:
print json_object['Msg']
return False
else:
return True
def req_get(url, param = '', type = 'json', token = True):
if token:
if param:
param.update({'token': leanote_token})
else:
param={'token': leanote_token}
s = requests.Session()
if leanote_host.startswith('https'):
s.mount('https://', SSLAdapter(ssl.PROTOCOL_TLSv1))
r = s.get(leanote_host + '/api/' + url, params = param, verify=False)
if r.status_code == requests.codes.ok:
if type=='json':
if is_ok(r.text):
rj = json.loads(r.text)
# if 'Msg' in rj:
# rj=rj['Msg']
return rj
else:
print '[Err] requests to url %s fail' %(r.url)
return None
elif type=='image':
i = Image.open(StringIO(r.content))
return i
else:
print '[Err] connect to url %s fail, error code %d ' %(r.url, r.status_cde)
return None
def req_post(url, param = '', type = 'json', token = True):
if token:
if param:
param.update({'token': leanote_token})
else:
param={'token': leanote_token}
s = requests.Session()
if leanote_host.startswith('https'):
s.mount('https://', SSLAdapter(ssl.PROTOCOL_TLSv1))
r = s.post(leanote_host + '/api/' + url, data = param, verify=False)
if r.status_code == requests.codes.ok:
if type=='json':
if is_ok(r.text):
rj = json.loads(r.text)
# if 'Msg' in rj:
# rj=rj['Msg']
return rj
else:
print '[Err] requests to url %s fail' %(r.url)
return None
else:
print '[Err] connect to url %s fail, error code %d ' %(r.url, r.status_cde)
return None
#ret leanote_token
def login(email, pwd):
param = {
'email': email,
'pwd': pwd,
}
r = req_get('auth/login', param, token=False)
if r:
print 'Login success! Welcome %s (%s)' %(r['Username'], r['Email'])
return r['Token']
else:
print 'Login fail! Start again.'
exit()
def logout():
return req_get('auth/logout')
#ret dict(notebookId: type.Notebook}
def getNotebooks(includeTrash = False):
r = req_get('notebook/getNotebooks')
if r:
if includeTrash:
return {notebook['NotebookId'] : notebook for notebook in r}
else:
return {notebook['NotebookId'] : notebook for notebook in r if not notebook['IsDeleted']}
else:
return none
#ret [type.Note], which contains noteId, and note meta data
def getNotesMeta(notebookId):
param = {
'notebookId': notebookId,
} | def getNoteDetail(noteId):
param = {
'noteId': noteId,
}
return req_get('note/getNoteAndContent', param)
def getImage(fileId):
param = {
'fileId': fileId,
}
return req_get('file/getImage', param, type = 'image')
def addNotebook(title='Import', parentId='', seq=-1):
param = {
'title': title,
'parentNotebookId': parentId,
'seq' : seq
}
return req_post('notebook/addNotebook', param)
def addNote(NotebookId, Title, Content, Tags=[], IsMarkdown = True, Abstract= '', Files=[]):
param = {
'NotebookId': NotebookId,
'Title': Title,
'Content': Content,
'Tags[]': Tags,
'IsMarkdown': IsMarkdown,
'Abstract': Abstract,
#'Files' : seq
}
return req_post('note/addNote', param)
def readFromFile(filename):
import yaml
with open (filename) as file:
file_meta = ''
file_content = ''
meta_flag=False
for line in file:
#print line
if meta_flag:
file_content += line
else:
if line.find('---')>-1:
meta_flag = True
else:
file_meta += line
#print meta
if not meta_flag:
file_content = file_meta
file_meta = ''
if meta_flag:
meta = yaml.load(file_meta)
else:
meta = {}
return file_content, meta
def saveToFile(notes, noteBooks, path = '.'):
unique_noteTitle = set()
for note in notes:
if note['Title'] == '':
filename = note['NoteId']
else:
filename = note['Title']
if filename in unique_noteTitle:
filename='%s_%s' %(filename, note['NoteId'])
else:
unique_noteTitle.add(filename)
if note['IsMarkdown']:
filename += '.md'
else:
filename += '.txt'
try:
with open(path + '/' + filename, 'w') as file:
print 'write file: %s' %filename
file.write('title: %s\n' %note['Title'].encode('utf-8'))
date = dateutil.parser.parse(note['CreatedTime'])
file.write('date: %s\n' %datetime.strftime(date.astimezone(local_zone), '%Y/%m/%d %H:%M:%S'))
date = dateutil.parser.parse(note['UpdatedTime'])
file.write('updated: %s\n' %datetime.strftime(date.astimezone(local_zone), '%Y/%m/%d %H:%M:%S'))
if note['Tags']:
if len(note['Tags']) == 1:
if note['Tags'][0]:
file.write('tags:\n')
for tag in note['Tags']:
file.write('- %s\n' %tag.encode('utf-8'))
category = []
current_notebook = note['NotebookId']
category.append(noteBooks[current_notebook]['Title'])
while noteBooks[current_notebook]['ParentNotebookId'] != '':
category.append(noteBooks[noteBooks[current_notebook]['ParentNotebookId']]['Title'])
current_notebook = noteBooks[current_notebook]['ParentNotebookId']
file.write('categories:\n')
category.reverse()
for cat in category:
file.write('- %s\n' %cat.encode('utf-8'))
file.write('---\n')
file.write('%s' %note['Content'].encode('utf-8'))
file.close()
if note['Files']:
if len(note['Files']) > 0:
for attach in note['Files']:
if not attach['IsAttach']:
i = getImage(attach['FileId'])
print 'saving its image: %s.%s' %(attach['FileId'], i.format)
i.save(attach['FileId'] + '.' + i.format)
except Exception as e:
logging.exception(e)
print "error: ", filename
def LeanoteExportToMD(path = '.'):
print 'Reading your notebooks...'
noteBooks = getNotebooks()
#get not deleted notes list
notes=[]
for notebook in noteBooks.values():
if not notebook['IsDeleted']:
notesMeta = getNotesMeta(notebook['NotebookId'])
for noteMeta in notesMeta:
if not noteMeta['IsTrash']:
note = getNoteDetail(noteMeta['NoteId'])
notes.append(note)
print 'found %d notes' %len(notes)
#write file
saveToFile(notes, noteBooks, path = path)
logout()
print 'all done, bye~'
def LeanoteImportFromMD(path = '.'):
filelist = os.listdir(path)
filelist = [file for file in filelist if file.find('.md')>-1 or file.find('.txt')>-1]
importedNotebookTitleMapID = {}
ret = addNotebook(title='imported_note', parentId='', seq=-1)
if ret:
print 'imporing into a new notebook: %s' %ret['Title']
importedNotebookTitleMapID['import'] = ret['NotebookId']
for filename in filelist:
content, meta = readFromFile(path + '/' + filename)
parentTitle='import'
currentTitle=''
if not meta.get('categories'):
categories=['import']
else:
categories= meta.get('categories')
for cat in categories:
currentTitle=cat
if currentTitle in importedNotebookTitleMapID.keys():
parentTitle=currentTitle
else:
ret = addNotebook(title = currentTitle, parentId = importedNotebookTitleMapID[parentTitle])
importedNotebookTitleMapID[currentTitle] = ret['NotebookId']
parentTitle=currentTitle
if not meta.get('title'):
meta['title'] = filename.replace('.md','').replace('.txt','')
importedNote = addNote(NotebookId=importedNotebookTitleMapID[currentTitle | return req_get('note/getNotes', param)
#ret type.NoteContent | random_line_split |
Leanote4MD.py | Ok']:
return True
else:
print json_object['Msg']
return False
else:
return True
def req_get(url, param = '', type = 'json', token = True):
if token:
if param:
param.update({'token': leanote_token})
else:
param={'token': leanote_token}
s = requests.Session()
if leanote_host.startswith('https'):
s.mount('https://', SSLAdapter(ssl.PROTOCOL_TLSv1))
r = s.get(leanote_host + '/api/' + url, params = param, verify=False)
if r.status_code == requests.codes.ok:
if type=='json':
if is_ok(r.text):
rj = json.loads(r.text)
# if 'Msg' in rj:
# rj=rj['Msg']
return rj
else:
print '[Err] requests to url %s fail' %(r.url)
return None
elif type=='image':
i = Image.open(StringIO(r.content))
return i
else:
print '[Err] connect to url %s fail, error code %d ' %(r.url, r.status_cde)
return None
def req_post(url, param = '', type = 'json', token = True):
if token:
if param:
param.update({'token': leanote_token})
else:
param={'token': leanote_token}
s = requests.Session()
if leanote_host.startswith('https'):
s.mount('https://', SSLAdapter(ssl.PROTOCOL_TLSv1))
r = s.post(leanote_host + '/api/' + url, data = param, verify=False)
if r.status_code == requests.codes.ok:
if type=='json':
if is_ok(r.text):
rj = json.loads(r.text)
# if 'Msg' in rj:
# rj=rj['Msg']
return rj
else:
print '[Err] requests to url %s fail' %(r.url)
return None
else:
print '[Err] connect to url %s fail, error code %d ' %(r.url, r.status_cde)
return None
#ret leanote_token
def login(email, pwd):
param = {
'email': email,
'pwd': pwd,
}
r = req_get('auth/login', param, token=False)
if r:
print 'Login success! Welcome %s (%s)' %(r['Username'], r['Email'])
return r['Token']
else:
print 'Login fail! Start again.'
exit()
def logout():
return req_get('auth/logout')
#ret dict(notebookId: type.Notebook}
def getNotebooks(includeTrash = False):
r = req_get('notebook/getNotebooks')
if r:
if includeTrash:
return {notebook['NotebookId'] : notebook for notebook in r}
else:
return {notebook['NotebookId'] : notebook for notebook in r if not notebook['IsDeleted']}
else:
return none
#ret [type.Note], which contains noteId, and note meta data
def getNotesMeta(notebookId):
param = {
'notebookId': notebookId,
}
return req_get('note/getNotes', param)
#ret type.NoteContent
def getNoteDetail(noteId):
param = {
'noteId': noteId,
}
return req_get('note/getNoteAndContent', param)
def getImage(fileId):
param = {
'fileId': fileId,
}
return req_get('file/getImage', param, type = 'image')
def addNotebook(title='Import', parentId='', seq=-1):
param = {
'title': title,
'parentNotebookId': parentId,
'seq' : seq
}
return req_post('notebook/addNotebook', param)
def addNote(NotebookId, Title, Content, Tags=[], IsMarkdown = True, Abstract= '', Files=[]):
param = {
'NotebookId': NotebookId,
'Title': Title,
'Content': Content,
'Tags[]': Tags,
'IsMarkdown': IsMarkdown,
'Abstract': Abstract,
#'Files' : seq
}
return req_post('note/addNote', param)
def readFromFile(filename):
import yaml
with open (filename) as file:
file_meta = ''
file_content = ''
meta_flag=False
for line in file:
#print line
|
if not meta_flag:
file_content = file_meta
file_meta = ''
if meta_flag:
meta = yaml.load(file_meta)
else:
meta = {}
return file_content, meta
def saveToFile(notes, noteBooks, path = '.'):
unique_noteTitle = set()
for note in notes:
if note['Title'] == '':
filename = note['NoteId']
else:
filename = note['Title']
if filename in unique_noteTitle:
filename='%s_%s' %(filename, note['NoteId'])
else:
unique_noteTitle.add(filename)
if note['IsMarkdown']:
filename += '.md'
else:
filename += '.txt'
try:
with open(path + '/' + filename, 'w') as file:
print 'write file: %s' %filename
file.write('title: %s\n' %note['Title'].encode('utf-8'))
date = dateutil.parser.parse(note['CreatedTime'])
file.write('date: %s\n' %datetime.strftime(date.astimezone(local_zone), '%Y/%m/%d %H:%M:%S'))
date = dateutil.parser.parse(note['UpdatedTime'])
file.write('updated: %s\n' %datetime.strftime(date.astimezone(local_zone), '%Y/%m/%d %H:%M:%S'))
if note['Tags']:
if len(note['Tags']) == 1:
if note['Tags'][0]:
file.write('tags:\n')
for tag in note['Tags']:
file.write('- %s\n' %tag.encode('utf-8'))
category = []
current_notebook = note['NotebookId']
category.append(noteBooks[current_notebook]['Title'])
while noteBooks[current_notebook]['ParentNotebookId'] != '':
category.append(noteBooks[noteBooks[current_notebook]['ParentNotebookId']]['Title'])
current_notebook = noteBooks[current_notebook]['ParentNotebookId']
file.write('categories:\n')
category.reverse()
for cat in category:
file.write('- %s\n' %cat.encode('utf-8'))
file.write('---\n')
file.write('%s' %note['Content'].encode('utf-8'))
file.close()
if note['Files']:
if len(note['Files']) > 0:
for attach in note['Files']:
if not attach['IsAttach']:
i = getImage(attach['FileId'])
print 'saving its image: %s.%s' %(attach['FileId'], i.format)
i.save(attach['FileId'] + '.' + i.format)
except Exception as e:
logging.exception(e)
print "error: ", filename
def LeanoteExportToMD(path = '.'):
print 'Reading your notebooks...'
noteBooks = getNotebooks()
#get not deleted notes list
notes=[]
for notebook in noteBooks.values():
if not notebook['IsDeleted']:
notesMeta = getNotesMeta(notebook['NotebookId'])
for noteMeta in notesMeta:
if not noteMeta['IsTrash']:
note = getNoteDetail(noteMeta['NoteId'])
notes.append(note)
print 'found %d notes' %len(notes)
#write file
saveToFile(notes, noteBooks, path = path)
logout()
print 'all done, bye~'
def LeanoteImportFromMD(path = '.'):
filelist = os.listdir(path)
filelist = [file for file in filelist if file.find('.md')>-1 or file.find('.txt')>-1]
importedNotebookTitleMapID = {}
ret = addNotebook(title='imported_note', parentId='', seq=-1)
if ret:
print 'imporing into a new notebook: %s' %ret['Title']
importedNotebookTitleMapID['import'] = ret['NotebookId']
for filename in filelist:
content, meta = readFromFile(path + '/' + filename)
parentTitle='import'
currentTitle=''
if not meta.get('categories'):
categories=['import']
else:
categories= meta.get('categories')
for cat in categories:
currentTitle=cat
if currentTitle in importedNotebookTitleMapID.keys():
parentTitle=currentTitle
else:
ret = addNotebook(title = currentTitle, parentId = importedNotebookTitleMapID[parentTitle])
importedNotebookTitleMapID[currentTitle] = ret['NotebookId']
parentTitle=currentTitle
if not meta.get('title'):
meta['title'] = filename.replace('.md','').replace('.txt','')
importedNote = addNote(NotebookId=importedNotebookTitleMapID[currentTitle | if meta_flag:
file_content += line
else:
if line.find('---')>-1:
meta_flag = True
else:
file_meta += line
#print meta | conditional_block |
Leanote4MD.py | Ok']:
return True
else:
print json_object['Msg']
return False
else:
return True
def req_get(url, param = '', type = 'json', token = True):
if token:
if param:
param.update({'token': leanote_token})
else:
param={'token': leanote_token}
s = requests.Session()
if leanote_host.startswith('https'):
s.mount('https://', SSLAdapter(ssl.PROTOCOL_TLSv1))
r = s.get(leanote_host + '/api/' + url, params = param, verify=False)
if r.status_code == requests.codes.ok:
if type=='json':
if is_ok(r.text):
rj = json.loads(r.text)
# if 'Msg' in rj:
# rj=rj['Msg']
return rj
else:
print '[Err] requests to url %s fail' %(r.url)
return None
elif type=='image':
i = Image.open(StringIO(r.content))
return i
else:
print '[Err] connect to url %s fail, error code %d ' %(r.url, r.status_cde)
return None
def req_post(url, param = '', type = 'json', token = True):
if token:
if param:
param.update({'token': leanote_token})
else:
param={'token': leanote_token}
s = requests.Session()
if leanote_host.startswith('https'):
s.mount('https://', SSLAdapter(ssl.PROTOCOL_TLSv1))
r = s.post(leanote_host + '/api/' + url, data = param, verify=False)
if r.status_code == requests.codes.ok:
if type=='json':
if is_ok(r.text):
rj = json.loads(r.text)
# if 'Msg' in rj:
# rj=rj['Msg']
return rj
else:
print '[Err] requests to url %s fail' %(r.url)
return None
else:
print '[Err] connect to url %s fail, error code %d ' %(r.url, r.status_cde)
return None
#ret leanote_token
def login(email, pwd):
param = {
'email': email,
'pwd': pwd,
}
r = req_get('auth/login', param, token=False)
if r:
print 'Login success! Welcome %s (%s)' %(r['Username'], r['Email'])
return r['Token']
else:
print 'Login fail! Start again.'
exit()
def logout():
return req_get('auth/logout')
#ret dict(notebookId: type.Notebook}
def getNotebooks(includeTrash = False):
r = req_get('notebook/getNotebooks')
if r:
if includeTrash:
return {notebook['NotebookId'] : notebook for notebook in r}
else:
return {notebook['NotebookId'] : notebook for notebook in r if not notebook['IsDeleted']}
else:
return none
#ret [type.Note], which contains noteId, and note meta data
def getNotesMeta(notebookId):
param = {
'notebookId': notebookId,
}
return req_get('note/getNotes', param)
#ret type.NoteContent
def getNoteDetail(noteId):
param = {
'noteId': noteId,
}
return req_get('note/getNoteAndContent', param)
def getImage(fileId):
param = {
'fileId': fileId,
}
return req_get('file/getImage', param, type = 'image')
def addNotebook(title='Import', parentId='', seq=-1):
|
def addNote(NotebookId, Title, Content, Tags=[], IsMarkdown = True, Abstract= '', Files=[]):
param = {
'NotebookId': NotebookId,
'Title': Title,
'Content': Content,
'Tags[]': Tags,
'IsMarkdown': IsMarkdown,
'Abstract': Abstract,
#'Files' : seq
}
return req_post('note/addNote', param)
def readFromFile(filename):
import yaml
with open (filename) as file:
file_meta = ''
file_content = ''
meta_flag=False
for line in file:
#print line
if meta_flag:
file_content += line
else:
if line.find('---')>-1:
meta_flag = True
else:
file_meta += line
#print meta
if not meta_flag:
file_content = file_meta
file_meta = ''
if meta_flag:
meta = yaml.load(file_meta)
else:
meta = {}
return file_content, meta
def saveToFile(notes, noteBooks, path = '.'):
unique_noteTitle = set()
for note in notes:
if note['Title'] == '':
filename = note['NoteId']
else:
filename = note['Title']
if filename in unique_noteTitle:
filename='%s_%s' %(filename, note['NoteId'])
else:
unique_noteTitle.add(filename)
if note['IsMarkdown']:
filename += '.md'
else:
filename += '.txt'
try:
with open(path + '/' + filename, 'w') as file:
print 'write file: %s' %filename
file.write('title: %s\n' %note['Title'].encode('utf-8'))
date = dateutil.parser.parse(note['CreatedTime'])
file.write('date: %s\n' %datetime.strftime(date.astimezone(local_zone), '%Y/%m/%d %H:%M:%S'))
date = dateutil.parser.parse(note['UpdatedTime'])
file.write('updated: %s\n' %datetime.strftime(date.astimezone(local_zone), '%Y/%m/%d %H:%M:%S'))
if note['Tags']:
if len(note['Tags']) == 1:
if note['Tags'][0]:
file.write('tags:\n')
for tag in note['Tags']:
file.write('- %s\n' %tag.encode('utf-8'))
category = []
current_notebook = note['NotebookId']
category.append(noteBooks[current_notebook]['Title'])
while noteBooks[current_notebook]['ParentNotebookId'] != '':
category.append(noteBooks[noteBooks[current_notebook]['ParentNotebookId']]['Title'])
current_notebook = noteBooks[current_notebook]['ParentNotebookId']
file.write('categories:\n')
category.reverse()
for cat in category:
file.write('- %s\n' %cat.encode('utf-8'))
file.write('---\n')
file.write('%s' %note['Content'].encode('utf-8'))
file.close()
if note['Files']:
if len(note['Files']) > 0:
for attach in note['Files']:
if not attach['IsAttach']:
i = getImage(attach['FileId'])
print 'saving its image: %s.%s' %(attach['FileId'], i.format)
i.save(attach['FileId'] + '.' + i.format)
except Exception as e:
logging.exception(e)
print "error: ", filename
def LeanoteExportToMD(path = '.'):
print 'Reading your notebooks...'
noteBooks = getNotebooks()
#get not deleted notes list
notes=[]
for notebook in noteBooks.values():
if not notebook['IsDeleted']:
notesMeta = getNotesMeta(notebook['NotebookId'])
for noteMeta in notesMeta:
if not noteMeta['IsTrash']:
note = getNoteDetail(noteMeta['NoteId'])
notes.append(note)
print 'found %d notes' %len(notes)
#write file
saveToFile(notes, noteBooks, path = path)
logout()
print 'all done, bye~'
def LeanoteImportFromMD(path = '.'):
filelist = os.listdir(path)
filelist = [file for file in filelist if file.find('.md')>-1 or file.find('.txt')>-1]
importedNotebookTitleMapID = {}
ret = addNotebook(title='imported_note', parentId='', seq=-1)
if ret:
print 'imporing into a new notebook: %s' %ret['Title']
importedNotebookTitleMapID['import'] = ret['NotebookId']
for filename in filelist:
content, meta = readFromFile(path + '/' + filename)
parentTitle='import'
currentTitle=''
if not meta.get('categories'):
categories=['import']
else:
categories= meta.get('categories')
for cat in categories:
currentTitle=cat
if currentTitle in importedNotebookTitleMapID.keys():
parentTitle=currentTitle
else:
ret = addNotebook(title = currentTitle, parentId = importedNotebookTitleMapID[parentTitle])
importedNotebookTitleMapID[currentTitle] = ret['NotebookId']
parentTitle=currentTitle
if not meta.get('title'):
meta['title'] = filename.replace('.md','').replace('.txt','')
importedNote = addNote(NotebookId=importedNotebookTitleMapID[current | param = {
'title': title,
'parentNotebookId': parentId,
'seq' : seq
}
return req_post('notebook/addNotebook', param) | identifier_body |
Leanote4MD.py | Ok']:
return True
else:
print json_object['Msg']
return False
else:
return True
def req_get(url, param = '', type = 'json', token = True):
if token:
if param:
param.update({'token': leanote_token})
else:
param={'token': leanote_token}
s = requests.Session()
if leanote_host.startswith('https'):
s.mount('https://', SSLAdapter(ssl.PROTOCOL_TLSv1))
r = s.get(leanote_host + '/api/' + url, params = param, verify=False)
if r.status_code == requests.codes.ok:
if type=='json':
if is_ok(r.text):
rj = json.loads(r.text)
# if 'Msg' in rj:
# rj=rj['Msg']
return rj
else:
print '[Err] requests to url %s fail' %(r.url)
return None
elif type=='image':
i = Image.open(StringIO(r.content))
return i
else:
print '[Err] connect to url %s fail, error code %d ' %(r.url, r.status_cde)
return None
def req_post(url, param = '', type = 'json', token = True):
if token:
if param:
param.update({'token': leanote_token})
else:
param={'token': leanote_token}
s = requests.Session()
if leanote_host.startswith('https'):
s.mount('https://', SSLAdapter(ssl.PROTOCOL_TLSv1))
r = s.post(leanote_host + '/api/' + url, data = param, verify=False)
if r.status_code == requests.codes.ok:
if type=='json':
if is_ok(r.text):
rj = json.loads(r.text)
# if 'Msg' in rj:
# rj=rj['Msg']
return rj
else:
print '[Err] requests to url %s fail' %(r.url)
return None
else:
print '[Err] connect to url %s fail, error code %d ' %(r.url, r.status_cde)
return None
#ret leanote_token
def login(email, pwd):
param = {
'email': email,
'pwd': pwd,
}
r = req_get('auth/login', param, token=False)
if r:
print 'Login success! Welcome %s (%s)' %(r['Username'], r['Email'])
return r['Token']
else:
print 'Login fail! Start again.'
exit()
def logout():
return req_get('auth/logout')
#ret dict(notebookId: type.Notebook}
def getNotebooks(includeTrash = False):
r = req_get('notebook/getNotebooks')
if r:
if includeTrash:
return {notebook['NotebookId'] : notebook for notebook in r}
else:
return {notebook['NotebookId'] : notebook for notebook in r if not notebook['IsDeleted']}
else:
return none
#ret [type.Note], which contains noteId, and note meta data
def getNotesMeta(notebookId):
param = {
'notebookId': notebookId,
}
return req_get('note/getNotes', param)
#ret type.NoteContent
def getNoteDetail(noteId):
param = {
'noteId': noteId,
}
return req_get('note/getNoteAndContent', param)
def | (fileId):
param = {
'fileId': fileId,
}
return req_get('file/getImage', param, type = 'image')
def addNotebook(title='Import', parentId='', seq=-1):
param = {
'title': title,
'parentNotebookId': parentId,
'seq' : seq
}
return req_post('notebook/addNotebook', param)
def addNote(NotebookId, Title, Content, Tags=[], IsMarkdown = True, Abstract= '', Files=[]):
param = {
'NotebookId': NotebookId,
'Title': Title,
'Content': Content,
'Tags[]': Tags,
'IsMarkdown': IsMarkdown,
'Abstract': Abstract,
#'Files' : seq
}
return req_post('note/addNote', param)
def readFromFile(filename):
import yaml
with open (filename) as file:
file_meta = ''
file_content = ''
meta_flag=False
for line in file:
#print line
if meta_flag:
file_content += line
else:
if line.find('---')>-1:
meta_flag = True
else:
file_meta += line
#print meta
if not meta_flag:
file_content = file_meta
file_meta = ''
if meta_flag:
meta = yaml.load(file_meta)
else:
meta = {}
return file_content, meta
def saveToFile(notes, noteBooks, path = '.'):
unique_noteTitle = set()
for note in notes:
if note['Title'] == '':
filename = note['NoteId']
else:
filename = note['Title']
if filename in unique_noteTitle:
filename='%s_%s' %(filename, note['NoteId'])
else:
unique_noteTitle.add(filename)
if note['IsMarkdown']:
filename += '.md'
else:
filename += '.txt'
try:
with open(path + '/' + filename, 'w') as file:
print 'write file: %s' %filename
file.write('title: %s\n' %note['Title'].encode('utf-8'))
date = dateutil.parser.parse(note['CreatedTime'])
file.write('date: %s\n' %datetime.strftime(date.astimezone(local_zone), '%Y/%m/%d %H:%M:%S'))
date = dateutil.parser.parse(note['UpdatedTime'])
file.write('updated: %s\n' %datetime.strftime(date.astimezone(local_zone), '%Y/%m/%d %H:%M:%S'))
if note['Tags']:
if len(note['Tags']) == 1:
if note['Tags'][0]:
file.write('tags:\n')
for tag in note['Tags']:
file.write('- %s\n' %tag.encode('utf-8'))
category = []
current_notebook = note['NotebookId']
category.append(noteBooks[current_notebook]['Title'])
while noteBooks[current_notebook]['ParentNotebookId'] != '':
category.append(noteBooks[noteBooks[current_notebook]['ParentNotebookId']]['Title'])
current_notebook = noteBooks[current_notebook]['ParentNotebookId']
file.write('categories:\n')
category.reverse()
for cat in category:
file.write('- %s\n' %cat.encode('utf-8'))
file.write('---\n')
file.write('%s' %note['Content'].encode('utf-8'))
file.close()
if note['Files']:
if len(note['Files']) > 0:
for attach in note['Files']:
if not attach['IsAttach']:
i = getImage(attach['FileId'])
print 'saving its image: %s.%s' %(attach['FileId'], i.format)
i.save(attach['FileId'] + '.' + i.format)
except Exception as e:
logging.exception(e)
print "error: ", filename
def LeanoteExportToMD(path = '.'):
print 'Reading your notebooks...'
noteBooks = getNotebooks()
#get not deleted notes list
notes=[]
for notebook in noteBooks.values():
if not notebook['IsDeleted']:
notesMeta = getNotesMeta(notebook['NotebookId'])
for noteMeta in notesMeta:
if not noteMeta['IsTrash']:
note = getNoteDetail(noteMeta['NoteId'])
notes.append(note)
print 'found %d notes' %len(notes)
#write file
saveToFile(notes, noteBooks, path = path)
logout()
print 'all done, bye~'
def LeanoteImportFromMD(path = '.'):
filelist = os.listdir(path)
filelist = [file for file in filelist if file.find('.md')>-1 or file.find('.txt')>-1]
importedNotebookTitleMapID = {}
ret = addNotebook(title='imported_note', parentId='', seq=-1)
if ret:
print 'imporing into a new notebook: %s' %ret['Title']
importedNotebookTitleMapID['import'] = ret['NotebookId']
for filename in filelist:
content, meta = readFromFile(path + '/' + filename)
parentTitle='import'
currentTitle=''
if not meta.get('categories'):
categories=['import']
else:
categories= meta.get('categories')
for cat in categories:
currentTitle=cat
if currentTitle in importedNotebookTitleMapID.keys():
parentTitle=currentTitle
else:
ret = addNotebook(title = currentTitle, parentId = importedNotebookTitleMapID[parentTitle])
importedNotebookTitleMapID[currentTitle] = ret['NotebookId']
parentTitle=currentTitle
if not meta.get('title'):
meta['title'] = filename.replace('.md','').replace('.txt','')
importedNote = addNote(NotebookId=importedNotebookTitleMapID[current | getImage | identifier_name |
AddDoctorForm.js | const doctorImageRef = useRef()
if(!!props.uploadRet){
if(!!props.uploadRet.success){
console.log(props.uploadRet,"props.uploadRet")
props.setImage(props.uploadRet.data)
addToast(props.uploadRet.message, {appearance: 'success', autoDismiss:true})
}else |
props.loadingImageOff()
props.uploadRetClr()
}
if(!!props.addDoctorRet){
if(!!props.addDoctorRet.success){
addToast(props.addDoctorRet.message, {appearance: 'success', autoDismiss:true})
console.log(props.addDoctorRet,"props.addDoctorRet")
props.set_user_info({
...props.prof_data,
doctors:[...props.addDoctorRet.data.doctors]
})
props.clear_data()
}else{
addToast(props.addDoctorRet.message, {appearance: 'error', autoDismiss:true})
}
props.getUserDetails()
props.addDoctorClr()
props.addDoctorLoadingOff()
}
const submitdetails = () => {
if(props.name === '' || props.department === '' ||props.designation==='' || props.experience==="" || props.education==="" || props.specialitie_chosen===" " || (props.services_chosen.length===0) ){
addToast("Enter all the details",{ appearance: 'error', autoDismiss:true })
}else if(!!!props.doctorProfileImage){
addToast("Please provide a profile image",{ appearance: 'error', autoDismiss:true })
}else{
props.submitdetails({
name:props.name,
designation:props.designation,
department:props.department,
experience:props.experience,
education:props.education,
services_chosen:props.services_chosen,
specialitie_chosen:props.specialitie_chosen,
doctorProfileImage:props.doctorProfileImage,
doctorId:get_url_params('id')?get_url_params('id'):undefined
})
}
}
const handleImageClick = ()=>{
let element = document.getElementById('doctorImageInput')
element.click()
}
const handleUploadImage = (e) => {
e.preventDefault();
e.stopPropagation()
var reader = new FileReader();
var file = e.target.files[0];
if(!!file){
if (file.size > 2 * 1024 * 1024) {
addToast('File size should be less than 2MB', {appearance: 'error', autoDismiss:true})
} else {
props.upload({ file: file, field: 'file' })
reader.onloadend = () => {
reader.readAsDataURL(file);
}
}
}else{
addToast('No File Found', {appearance: 'error', autoDismiss:true})
}
}
console.log(props,"props in AddDoctorForm")
return (
<React.Fragment>
<div className="profile_secti">
<div style={{height:'10px',width:'10px',position:'absolute',top:'-10px'}} ref={myRef}></div>
{props.addDoctorLoading && <LoaderComponent />}
<h5 className="pfo_im">Profile Image</h5>
<div className="row">
<div className="col-lg-2 col-md-4 image_wrapper_add_doctor position-relative">
{props.laodingImage && <LoaderComponent />}
<input
style={{display:'inline',display:'none'}}
id="doctorImageInput"
type="file" accept="image/jpe ,image/png, image/jpeg"
onChange ={(e)=>handleUploadImage(e)}
ref = {doctorImageRef}
/>
<img src={!!props.doctorProfileImage?props.doctorProfileImage:'/account.svg'} className="accout"/>
<img onClick={(e)=>handleImageClick(e)} src="/camera.svg" className=" profile_camera_rish cursor-pointer" />
</div>
<div className="col-lg-3">
<h6 className="fil_nm">{!!props.doctorImageName?props.doctorImageName:'File Name'}</h6>
<button onClick={(e)=>handleImageClick(e)} className="upld common-button">Upload</button>
</div>
</div>
<form class="shake" role="form" method="post" id="contactForm" name="contact-form" data-toggle="validator">
<div class="form-group label-floating">
<label class="control-label control_label_ris" for="name">Name</label>
<input class="form-control no_padding_ris btm_in_bdr" value= {props.name} onChange={props.handleChange} id="name" type="text" name="name" required data-error="Please enter your name" />
<div class="help-block with-errors"></div>
</div>
<div class="form-group label-floating">
<label class="control-label control_label_ris" for="education">Education Qualification</label>
<input class="form-control no_padding_ris btm_in_bdr" value= {props.education} onChange={props.handleChange} id="educationqua" type="education" name="education" required data-error="Please enter your education qulification" />
<div class="help-block with-errors"></div>
</div>
<div class="form-group label-floating">
<label class="control-label control_label_ris" for="education">Department</label>
<input class="form-control no_padding_ris btm_in_bdr" value= {props.department} onChange={props.handleChange} id="department" name="department" required data-error="Please enter your education qulification" />
<div class="help-block with-errors"></div>
</div>
<div class="form-group label-floating">
<label class="control-label control_label_ris">Designation</label>
<input class="form-control no_padding_ris btm_in_bdr" id="msg_Designation" value= {props.designation} onChange={props.handleChange} type="text" name="designation" required data-error="Please enter your message Designation" />
<div class="help-block with-errors"></div>
</div>
<div className="row form-group label-floating">
<div class="col-lg-6 col-12">
<Select
options = {props.specialities}
handleChange = {props.handleSelectChange}
value = {props.specialitie_chosen}
multiple ={false}
name = "specialitie_chosen"
label = "Speciality"
placeholder = "Choose Spectiality"
/>
</div>
<div class="col-lg-6 col-12">
<Select
options = {props.services}
handleChange = {props.handleSelectChange}
value = {props.services_chosen}
name = "services_chosen"
label = "Service"
placeholder = "Choose Servives"
/>
</div>
</div>
<div class="form-group label-floating">
<label for="message" class="control-label control_label_ris">Experience</label>
<input class="form-control no_padding_ris btm_in_bdr" value= {props.experience} onChange={props.handleChange} id="msg_Experience" type="number" name="experience" required data-error="Please enter your message Experience" />
<div class="help-block with-errors"></div>
</div>
</form>
</div>
<div className="time_she">
<h3 className="abaily text-center">Availability</h3>
<div className="row text-center">
<div className="col-lg-2"><h4>All</h4></div>
<div className="col-lg-4"><h4>From - To</h4></div>
<div className="col-lg-4"><h4>From - To</h4></div>
<div className="col-lg-2"><h4>Closed</h4></div>
</div>
{props.slots.map((item,i)=>(
<div className="row text-center">
<div className="col-lg-2"><p className="m">{item.day.charAt(0).toUpperCase()}</p></div>
<div className="col-lg-4"><p><span onClick={()=>props.slotClicked(item.slots.morning,'morning','from', item)} className="time_bor cursor-pointer">{props.timeToString(item.slots.morning.from)}</span><span onClick={()=>props.slotClicked(item.slots.morning,'morning','to', item)} className="time_bor cursor-pointer">{props.timeToString(item.slots.morning.to)}</span></p></div>
<div className="col-lg-4"><p><span onClick={()=>props.slotClicked(item.slots.evening,'evening','from', item)} className="time_bor cursor-pointer">{props.timeToString(item.slots.evening.from)}</span><span onClick={()=>props.slotClicked(item.slots.evening,'evening','to', item)} className="time_bor cursor-pointer">{props.timeToString(item.slots.evening.to)}</span></p></div>
<div className="col-lg-2">
<div
onClick = {(e)=>props.handleCloseDay(item,i,e)}
className='circul_rund'>
<label className={item.closed | {
addToast(props.uploadRet.message, {appearance: 'success', autoDismiss:true})
} | conditional_block |
AddDoctorForm.js | const doctorImageRef = useRef()
if(!!props.uploadRet){
if(!!props.uploadRet.success){
console.log(props.uploadRet,"props.uploadRet")
props.setImage(props.uploadRet.data)
addToast(props.uploadRet.message, {appearance: 'success', autoDismiss:true})
}else{
addToast(props.uploadRet.message, {appearance: 'success', autoDismiss:true})
}
props.loadingImageOff()
props.uploadRetClr()
}
if(!!props.addDoctorRet){
if(!!props.addDoctorRet.success){
addToast(props.addDoctorRet.message, {appearance: 'success', autoDismiss:true})
console.log(props.addDoctorRet,"props.addDoctorRet")
props.set_user_info({
...props.prof_data,
doctors:[...props.addDoctorRet.data.doctors]
})
props.clear_data()
}else{
addToast(props.addDoctorRet.message, {appearance: 'error', autoDismiss:true})
}
props.getUserDetails()
props.addDoctorClr()
props.addDoctorLoadingOff()
}
const submitdetails = () => {
if(props.name === '' || props.department === '' ||props.designation==='' || props.experience==="" || props.education==="" || props.specialitie_chosen===" " || (props.services_chosen.length===0) ){
addToast("Enter all the details",{ appearance: 'error', autoDismiss:true })
}else if(!!!props.doctorProfileImage){
addToast("Please provide a profile image",{ appearance: 'error', autoDismiss:true })
}else{
props.submitdetails({
name:props.name,
designation:props.designation,
department:props.department,
experience:props.experience,
education:props.education,
services_chosen:props.services_chosen,
specialitie_chosen:props.specialitie_chosen,
doctorProfileImage:props.doctorProfileImage,
doctorId:get_url_params('id')?get_url_params('id'):undefined
})
}
}
const handleImageClick = ()=>{
let element = document.getElementById('doctorImageInput')
element.click()
}
const handleUploadImage = (e) => {
e.preventDefault();
e.stopPropagation()
var reader = new FileReader();
var file = e.target.files[0];
if(!!file){
if (file.size > 2 * 1024 * 1024) {
addToast('File size should be less than 2MB', {appearance: 'error', autoDismiss:true})
} else {
props.upload({ file: file, field: 'file' })
reader.onloadend = () => {
reader.readAsDataURL(file);
}
}
}else{
addToast('No File Found', {appearance: 'error', autoDismiss:true})
}
}
console.log(props,"props in AddDoctorForm")
return (
<React.Fragment>
<div className="profile_secti">
<div style={{height:'10px',width:'10px',position:'absolute',top:'-10px'}} ref={myRef}></div>
{props.addDoctorLoading && <LoaderComponent />}
<h5 className="pfo_im">Profile Image</h5>
<div className="row">
<div className="col-lg-2 col-md-4 image_wrapper_add_doctor position-relative">
{props.laodingImage && <LoaderComponent />}
<input
style={{display:'inline',display:'none'}}
id="doctorImageInput"
type="file" accept="image/jpe ,image/png, image/jpeg"
onChange ={(e)=>handleUploadImage(e)}
ref = {doctorImageRef}
/>
<img src={!!props.doctorProfileImage?props.doctorProfileImage:'/account.svg'} className="accout"/>
<img onClick={(e)=>handleImageClick(e)} src="/camera.svg" className=" profile_camera_rish cursor-pointer" />
</div>
<div className="col-lg-3">
<h6 className="fil_nm">{!!props.doctorImageName?props.doctorImageName:'File Name'}</h6>
<button onClick={(e)=>handleImageClick(e)} className="upld common-button">Upload</button>
</div>
</div>
<form class="shake" role="form" method="post" id="contactForm" name="contact-form" data-toggle="validator">
<div class="form-group label-floating">
<label class="control-label control_label_ris" for="name">Name</label>
<input class="form-control no_padding_ris btm_in_bdr" value= {props.name} onChange={props.handleChange} id="name" type="text" name="name" required data-error="Please enter your name" />
<div class="help-block with-errors"></div>
</div>
<div class="form-group label-floating">
<label class="control-label control_label_ris" for="education">Education Qualification</label>
<input class="form-control no_padding_ris btm_in_bdr" value= {props.education} onChange={props.handleChange} id="educationqua" type="education" name="education" required data-error="Please enter your education qulification" />
<div class="help-block with-errors"></div>
</div>
<div class="form-group label-floating">
<label class="control-label control_label_ris" for="education">Department</label>
<input class="form-control no_padding_ris btm_in_bdr" value= {props.department} onChange={props.handleChange} id="department" name="department" required data-error="Please enter your education qulification" />
<div class="help-block with-errors"></div>
</div>
<div class="form-group label-floating">
<label class="control-label control_label_ris">Designation</label>
<input class="form-control no_padding_ris btm_in_bdr" id="msg_Designation" value= {props.designation} onChange={props.handleChange} type="text" name="designation" required data-error="Please enter your message Designation" />
<div class="help-block with-errors"></div>
</div>
<div className="row form-group label-floating">
<div class="col-lg-6 col-12">
<Select
options = {props.specialities}
handleChange = {props.handleSelectChange}
value = {props.specialitie_chosen}
multiple ={false}
name = "specialitie_chosen"
label = "Speciality"
placeholder = "Choose Spectiality"
/>
</div>
<div class="col-lg-6 col-12">
<Select
options = {props.services}
handleChange = {props.handleSelectChange}
value = {props.services_chosen}
name = "services_chosen"
label = "Service"
placeholder = "Choose Servives"
/>
</div>
</div>
<div class="form-group label-floating">
<label for="message" class="control-label control_label_ris">Experience</label>
<input class="form-control no_padding_ris btm_in_bdr" value= {props.experience} onChange={props.handleChange} id="msg_Experience" type="number" name="experience" required data-error="Please enter your message Experience" />
<div class="help-block with-errors"></div>
</div>
</form>
</div>
<div className="time_she">
<h3 className="abaily text-center">Availability</h3>
<div className="row text-center">
<div className="col-lg-2"><h4>All</h4></div>
<div className="col-lg-4"><h4>From - To</h4></div>
<div className="col-lg-4"><h4>From - To</h4></div>
<div className="col-lg-2"><h4>Closed</h4></div>
</div>
{props.slots.map((item,i)=>(
<div className="row text-center">
<div className="col-lg-2"><p className="m">{item.day.charAt(0).toUpperCase()}</p></div>
<div className="col-lg-4"><p><span onClick={()=>props.slotClicked(item.slots.morning,'morning','from', item)} className="time_bor cursor-pointer">{props.timeToString(item.slots.morning.from)}</span><span onClick={()=>props.slotClicked(item.slots.morning,'morning','to', item)} className="time_bor cursor-pointer">{props.timeToString(item.slots.morning.to)}</span></p></div>
<div className="col-lg-4"><p><span onClick={()=>props.slotClicked(item.slots.evening,'evening','from', item)} className="time_bor cursor-pointer">{props.timeToString(item.slots.evening.from)}</span><span onClick={()=>props.slotClicked(item.slots.evening,'evening','to', item)} className="time_bor cursor-pointer">{props.timeToString(item.slots.evening.to)}</span></p></div>
<div className="col-lg-2"> | <div
onClick = {(e)=>props.handleCloseDay(item,i,e)}
className='circul_rund'>
<label className={item.closed?' | random_line_split |
|
io_file_browser_search.py | This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program; if not, write to the Free Software Foundation,
# Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
#
# ***** END GPL LICENCE BLOCK *****
bl_info = {
"name": "File Browser Search",
"author": "Jakub Zolcik",
"version": (0, 1, 1),
"blender": (2, 6, 2),
"api": 35622,
"location": "File Browser",
"description": "Allows You to find files in File Browser by name.",
"warning": "",
"wiki_url": "http://wiki.blender.org/index.php/User:Sftd/Extensions:2.6/Py/Scripts/Import-Export/File_Browser_Search",
"tracker_url": "http://projects.blender.org/tracker/?func=detail&aid=30386&group_id=153&atid=467",
"category": "Import-Export"}
"""
Usage:
Launches in File Browser
"""
import bpy
import os
import re
class FilteredFileItem(bpy.types.PropertyGroup):
name = bpy.props.StringProperty(name="File name", default="")
dname = bpy.props.StringProperty(name="Display name", default="")
def fileSearch(self, context):
# print("file")
filter = context.window_manager.filtered_search_prop
directory = context.window_manager.last_directory_prop
filecol = context.window_manager.filtered_files_prop
for fname in filecol:
filecol.remove(0)
if filter == "":
return None
pattern = ""
special = ('\\', '.', '^', '$', '*', '+', '?', '{', '}', '[', ']', '|', '(', ')')
for c in special:
filter = filter.replace(c, '\\' + c)
if ('*' in filter):
filter = filter.replace('\*', '.*')
pattern = ('^' + filter.lower() + '$')
else:
if(len(filter) < 3):
pattern = ('^' + filter.lower() + r".*\..*" + '$')
else:
pattern = ('^' + r".*" + filter.lower() + r".*\..*" + '$')
prog = re.compile(pattern)
maxf = 100
cf = 0
dlen = len(directory)
maxd = 100
cd = 0
if context.window_manager.file_searchtree:
for path, dirs, files in os.walk(directory):
cd += 1
if cd > maxd:
break
for filename in files:
filename = (os.path.join(path, filename))[dlen:]
# rfilename = (os.path.join(path, filename))[dlen:]
if prog.match(filename.lower()) != None:
p = context.window_manager.filtered_files_prop.add()
# p.name = rfilename
p.name = filename
if context.blend_data.scenes[0].file_hideextensions:
ind = filename.rfind(".")
if ind > -1:
filename = filename[0:ind]
p.dname = filename
cf += 1
if(cf >= maxf):
break
if(cf >= maxf):
break
else:
filesList = os.listdir(directory)
for filename in filesList:
if prog.match(filename.lower()) != None:
p = context.window_manager.filtered_files_prop.add()
p.name = filename
if context.blend_data.scenes[0].file_hideextensions:
ind = filename.rfind(".")
if ind > -1:
filename = filename[0:ind]
p.dname = filename
return None
def blendDataFromFile(file, part):
with bpy.data.libraries.load(file) as (data_from, data_to):
if (part == "Action"):
return data_from.actions
elif part == "Armature":
return data_from.brushes
elif part == "Brush":
return data_from.brushes
elif part == "Camera":
return data_from.cameras
elif part == "Curve":
return data_from.curves
elif part == "Font":
return data_from.fonts
elif part == "Group":
return data_from.groups
elif part == "Image":
return data_from.images
elif part == "Lamp":
return data_from.lamps
elif part == "Lattice":
return data_from.lattices
elif part == "Library":
return data_from.libraries
elif part == "FreestyleLineStyle":
return data_from.linestyles
elif part == "Mask":
return data_from.masks
elif part == "Material":
return data_from.materials
elif part == "Mesh":
return data_from.meshes
elif part == "NodeTree":
return data_from.node_groups
elif part == "Object":
return data_from.objects
elif part == "Particle":
return data_from.particles
elif part == "Scene":
return data_from.scenes
elif part == "Screen":
return data_from.screens
elif part == "Script":
return data_from.scripts
elif part == "Sound":
return data_from.sounds
elif part == "Speaker":
return data_from.speakers
elif part == "Text":
return data_from.texts
elif part == "Texture":
return data_from.textures
elif part == "World":
return data_from.worlds
else:
return None
def notFileSearch(self, context):
# print("not file")
filter = context.window_manager.filtered_search_prop
directory = context.window_manager.last_directory_prop
filecol = context.window_manager.filtered_files_prop
for fname in filecol:
filecol.remove(0)
if filter == "":
return None
ind_e = directory.find(".blend")
if(ind_e == -1):
return None
ind_e = ind_e + 6
file = directory[0:ind_e]
part = directory[ind_e + 1:-1]
if (part == ""):
return None
data = None
data = blendDataFromFile(file, part)
pattern = ""
if(len(filter) < 3):
pattern = (filter.lower() + r".*")
else:
pattern = (r".*" + filter.lower() + r".*")
prog = re.compile(pattern)
for name in data:
if prog.match(name.lower()) != None:
p = context.window_manager.filtered_files_prop.add()
p.name = name
p.dname = name
return None
def filteredSearchFunc(self, context):
if(context.active_operator.bl_idname == "WM_OT_link_append"):
return notFileSearch(self, context)
else: |
class FilteredFileSelectOperator(bpy.types.Operator):
bl_idname = "file.filtered_file_select"
bl_label = "Select File"
fname = bpy.props.StringProperty()
fexec = bpy.props.BoolProperty()
def execute(self, context):
context.space_data.params.filename = self.fname
if self.fexec:
bpy.ops.file.execute('INVOKE_DEFAULT')
return {'FINISHED'}
class FilteredSearchPanel(bpy.types.Panel):
bl_idname = "FILE_PT_filteredsearch"
bl_label = "Search:"
bl_space_type = 'FILE_BROWSER'
bl_region_type = 'CHANNELS'
@classmethod
def poll(cls, context):
return (context.space_data.params is not None)
def draw(self, context):
layout = self.layout
directory = context.space_data.params.directory
if context.window_manager.last_directory_prop != directory:
context.window_manager.last_directory_prop = directory
filteredSearchFunc(self, context)
layout.prop(context.window_manager, "filtered_search_prop", "")
box = layout.box()
length = len(context.window_manager.filtered_files_prop)
incolumn = int(length / context.blend_data.scenes[0].file_columnsnumber)
r = length % context.blend_data.scenes[0].file_columnsnumber
row = box.row()
col = row.column()
it = 0
tr = 0
for f in context.window_manager.filtered_files_prop:
op = col.operator("file.filtered_file_select", text=f.dname, emboss=False)
op.fname = f.name
op.fexec = context.blend_data.scenes[0].file_autoexecute
it += 1
if tr < r:
if it % (incolumn + 1) == 0:
tr += 1
if(it < length):
col = row.column()
else:
if (it - tr) % incolumn == 0:
if(it < length):
col = row.column()
layout.prop(context.blend_data.scenes[0], "file_autoexecute")
layout.prop(context.window_manager, "file_searchtree")
layout.prop(context.blend_data.scenes[0], "file_hideextensions")
layout.prop(context.blend_data.scenes[0 | return fileSearch(self, context) | random_line_split |
io_file_browser_search.py | This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program; if not, write to the Free Software Foundation,
# Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
#
# ***** END GPL LICENCE BLOCK *****
bl_info = {
"name": "File Browser Search",
"author": "Jakub Zolcik",
"version": (0, 1, 1),
"blender": (2, 6, 2),
"api": 35622,
"location": "File Browser",
"description": "Allows You to find files in File Browser by name.",
"warning": "",
"wiki_url": "http://wiki.blender.org/index.php/User:Sftd/Extensions:2.6/Py/Scripts/Import-Export/File_Browser_Search",
"tracker_url": "http://projects.blender.org/tracker/?func=detail&aid=30386&group_id=153&atid=467",
"category": "Import-Export"}
"""
Usage:
Launches in File Browser
"""
import bpy
import os
import re
class FilteredFileItem(bpy.types.PropertyGroup):
name = bpy.props.StringProperty(name="File name", default="")
dname = bpy.props.StringProperty(name="Display name", default="")
def fileSearch(self, context):
# print("file")
filter = context.window_manager.filtered_search_prop
directory = context.window_manager.last_directory_prop
filecol = context.window_manager.filtered_files_prop
for fname in filecol:
filecol.remove(0)
if filter == "":
return None
pattern = ""
special = ('\\', '.', '^', '$', '*', '+', '?', '{', '}', '[', ']', '|', '(', ')')
for c in special:
filter = filter.replace(c, '\\' + c)
if ('*' in filter):
filter = filter.replace('\*', '.*')
pattern = ('^' + filter.lower() + '$')
else:
if(len(filter) < 3):
pattern = ('^' + filter.lower() + r".*\..*" + '$')
else:
pattern = ('^' + r".*" + filter.lower() + r".*\..*" + '$')
prog = re.compile(pattern)
maxf = 100
cf = 0
dlen = len(directory)
maxd = 100
cd = 0
if context.window_manager.file_searchtree:
for path, dirs, files in os.walk(directory):
cd += 1
if cd > maxd:
break
for filename in files:
filename = (os.path.join(path, filename))[dlen:]
# rfilename = (os.path.join(path, filename))[dlen:]
if prog.match(filename.lower()) != None:
p = context.window_manager.filtered_files_prop.add()
# p.name = rfilename
p.name = filename
if context.blend_data.scenes[0].file_hideextensions:
ind = filename.rfind(".")
if ind > -1:
filename = filename[0:ind]
p.dname = filename
cf += 1
if(cf >= maxf):
break
if(cf >= maxf):
break
else:
filesList = os.listdir(directory)
for filename in filesList:
if prog.match(filename.lower()) != None:
p = context.window_manager.filtered_files_prop.add()
p.name = filename
if context.blend_data.scenes[0].file_hideextensions:
ind = filename.rfind(".")
if ind > -1:
filename = filename[0:ind]
p.dname = filename
return None
def blendDataFromFile(file, part):
with bpy.data.libraries.load(file) as (data_from, data_to):
if (part == "Action"):
return data_from.actions
elif part == "Armature":
return data_from.brushes
elif part == "Brush":
return data_from.brushes
elif part == "Camera":
return data_from.cameras
elif part == "Curve":
return data_from.curves
elif part == "Font":
return data_from.fonts
elif part == "Group":
return data_from.groups
elif part == "Image":
return data_from.images
elif part == "Lamp":
return data_from.lamps
elif part == "Lattice":
return data_from.lattices
elif part == "Library":
return data_from.libraries
elif part == "FreestyleLineStyle":
return data_from.linestyles
elif part == "Mask":
return data_from.masks
elif part == "Material":
return data_from.materials
elif part == "Mesh":
return data_from.meshes
elif part == "NodeTree":
return data_from.node_groups
elif part == "Object":
return data_from.objects
elif part == "Particle":
return data_from.particles
elif part == "Scene":
return data_from.scenes
elif part == "Screen":
return data_from.screens
elif part == "Script":
return data_from.scripts
elif part == "Sound":
return data_from.sounds
elif part == "Speaker":
return data_from.speakers
elif part == "Text":
return data_from.texts
elif part == "Texture":
return data_from.textures
elif part == "World":
return data_from.worlds
else:
return None
def notFileSearch(self, context):
# print("not file")
filter = context.window_manager.filtered_search_prop
directory = context.window_manager.last_directory_prop
filecol = context.window_manager.filtered_files_prop
for fname in filecol:
filecol.remove(0)
if filter == "":
return None
ind_e = directory.find(".blend")
if(ind_e == -1):
return None
ind_e = ind_e + 6
file = directory[0:ind_e]
part = directory[ind_e + 1:-1]
if (part == ""):
return None
data = None
data = blendDataFromFile(file, part)
pattern = ""
if(len(filter) < 3):
pattern = (filter.lower() + r".*")
else:
pattern = (r".*" + filter.lower() + r".*")
prog = re.compile(pattern)
for name in data:
if prog.match(name.lower()) != None:
p = context.window_manager.filtered_files_prop.add()
p.name = name
p.dname = name
return None
def filteredSearchFunc(self, context):
if(context.active_operator.bl_idname == "WM_OT_link_append"):
return notFileSearch(self, context)
else:
return fileSearch(self, context)
class | (bpy.types.Operator):
bl_idname = "file.filtered_file_select"
bl_label = "Select File"
fname = bpy.props.StringProperty()
fexec = bpy.props.BoolProperty()
def execute(self, context):
context.space_data.params.filename = self.fname
if self.fexec:
bpy.ops.file.execute('INVOKE_DEFAULT')
return {'FINISHED'}
class FilteredSearchPanel(bpy.types.Panel):
bl_idname = "FILE_PT_filteredsearch"
bl_label = "Search:"
bl_space_type = 'FILE_BROWSER'
bl_region_type = 'CHANNELS'
@classmethod
def poll(cls, context):
return (context.space_data.params is not None)
def draw(self, context):
layout = self.layout
directory = context.space_data.params.directory
if context.window_manager.last_directory_prop != directory:
context.window_manager.last_directory_prop = directory
filteredSearchFunc(self, context)
layout.prop(context.window_manager, "filtered_search_prop", "")
box = layout.box()
length = len(context.window_manager.filtered_files_prop)
incolumn = int(length / context.blend_data.scenes[0].file_columnsnumber)
r = length % context.blend_data.scenes[0].file_columnsnumber
row = box.row()
col = row.column()
it = 0
tr = 0
for f in context.window_manager.filtered_files_prop:
op = col.operator("file.filtered_file_select", text=f.dname, emboss=False)
op.fname = f.name
op.fexec = context.blend_data.scenes[0].file_autoexecute
it += 1
if tr < r:
if it % (incolumn + 1) == 0:
tr += 1
if(it < length):
col = row.column()
else:
if (it - tr) % incolumn == 0:
if(it < length):
col = row.column()
layout.prop(context.blend_data.scenes[0], "file_autoexecute")
layout.prop(context.window_manager, "file_searchtree")
layout.prop(context.blend_data.scenes[0], "file_hideextensions")
layout.prop(context.blend_data.scenes[0], | FilteredFileSelectOperator | identifier_name |
io_file_browser_search.py | "api": 35622,
"location": "File Browser",
"description": "Allows You to find files in File Browser by name.",
"warning": "",
"wiki_url": "http://wiki.blender.org/index.php/User:Sftd/Extensions:2.6/Py/Scripts/Import-Export/File_Browser_Search",
"tracker_url": "http://projects.blender.org/tracker/?func=detail&aid=30386&group_id=153&atid=467",
"category": "Import-Export"}
"""
Usage:
Launches in File Browser
"""
import bpy
import os
import re
class FilteredFileItem(bpy.types.PropertyGroup):
name = bpy.props.StringProperty(name="File name", default="")
dname = bpy.props.StringProperty(name="Display name", default="")
def fileSearch(self, context):
# print("file")
filter = context.window_manager.filtered_search_prop
directory = context.window_manager.last_directory_prop
filecol = context.window_manager.filtered_files_prop
for fname in filecol:
filecol.remove(0)
if filter == "":
return None
pattern = ""
special = ('\\', '.', '^', '$', '*', '+', '?', '{', '}', '[', ']', '|', '(', ')')
for c in special:
filter = filter.replace(c, '\\' + c)
if ('*' in filter):
filter = filter.replace('\*', '.*')
pattern = ('^' + filter.lower() + '$')
else:
if(len(filter) < 3):
pattern = ('^' + filter.lower() + r".*\..*" + '$')
else:
pattern = ('^' + r".*" + filter.lower() + r".*\..*" + '$')
prog = re.compile(pattern)
maxf = 100
cf = 0
dlen = len(directory)
maxd = 100
cd = 0
if context.window_manager.file_searchtree:
for path, dirs, files in os.walk(directory):
cd += 1
if cd > maxd:
break
for filename in files:
filename = (os.path.join(path, filename))[dlen:]
# rfilename = (os.path.join(path, filename))[dlen:]
if prog.match(filename.lower()) != None:
p = context.window_manager.filtered_files_prop.add()
# p.name = rfilename
p.name = filename
if context.blend_data.scenes[0].file_hideextensions:
ind = filename.rfind(".")
if ind > -1:
filename = filename[0:ind]
p.dname = filename
cf += 1
if(cf >= maxf):
break
if(cf >= maxf):
break
else:
filesList = os.listdir(directory)
for filename in filesList:
if prog.match(filename.lower()) != None:
p = context.window_manager.filtered_files_prop.add()
p.name = filename
if context.blend_data.scenes[0].file_hideextensions:
ind = filename.rfind(".")
if ind > -1:
filename = filename[0:ind]
p.dname = filename
return None
def blendDataFromFile(file, part):
with bpy.data.libraries.load(file) as (data_from, data_to):
if (part == "Action"):
return data_from.actions
elif part == "Armature":
return data_from.brushes
elif part == "Brush":
return data_from.brushes
elif part == "Camera":
return data_from.cameras
elif part == "Curve":
return data_from.curves
elif part == "Font":
return data_from.fonts
elif part == "Group":
return data_from.groups
elif part == "Image":
return data_from.images
elif part == "Lamp":
return data_from.lamps
elif part == "Lattice":
return data_from.lattices
elif part == "Library":
return data_from.libraries
elif part == "FreestyleLineStyle":
return data_from.linestyles
elif part == "Mask":
return data_from.masks
elif part == "Material":
return data_from.materials
elif part == "Mesh":
return data_from.meshes
elif part == "NodeTree":
return data_from.node_groups
elif part == "Object":
return data_from.objects
elif part == "Particle":
return data_from.particles
elif part == "Scene":
return data_from.scenes
elif part == "Screen":
return data_from.screens
elif part == "Script":
return data_from.scripts
elif part == "Sound":
return data_from.sounds
elif part == "Speaker":
return data_from.speakers
elif part == "Text":
return data_from.texts
elif part == "Texture":
return data_from.textures
elif part == "World":
return data_from.worlds
else:
return None
def notFileSearch(self, context):
# print("not file")
filter = context.window_manager.filtered_search_prop
directory = context.window_manager.last_directory_prop
filecol = context.window_manager.filtered_files_prop
for fname in filecol:
filecol.remove(0)
if filter == "":
return None
ind_e = directory.find(".blend")
if(ind_e == -1):
return None
ind_e = ind_e + 6
file = directory[0:ind_e]
part = directory[ind_e + 1:-1]
if (part == ""):
return None
data = None
data = blendDataFromFile(file, part)
pattern = ""
if(len(filter) < 3):
pattern = (filter.lower() + r".*")
else:
pattern = (r".*" + filter.lower() + r".*")
prog = re.compile(pattern)
for name in data:
if prog.match(name.lower()) != None:
p = context.window_manager.filtered_files_prop.add()
p.name = name
p.dname = name
return None
def filteredSearchFunc(self, context):
if(context.active_operator.bl_idname == "WM_OT_link_append"):
return notFileSearch(self, context)
else:
return fileSearch(self, context)
class FilteredFileSelectOperator(bpy.types.Operator):
bl_idname = "file.filtered_file_select"
bl_label = "Select File"
fname = bpy.props.StringProperty()
fexec = bpy.props.BoolProperty()
def execute(self, context):
context.space_data.params.filename = self.fname
if self.fexec:
bpy.ops.file.execute('INVOKE_DEFAULT')
return {'FINISHED'}
class FilteredSearchPanel(bpy.types.Panel):
bl_idname = "FILE_PT_filteredsearch"
bl_label = "Search:"
bl_space_type = 'FILE_BROWSER'
bl_region_type = 'CHANNELS'
@classmethod
def poll(cls, context):
return (context.space_data.params is not None)
def draw(self, context):
layout = self.layout
directory = context.space_data.params.directory
if context.window_manager.last_directory_prop != directory:
context.window_manager.last_directory_prop = directory
filteredSearchFunc(self, context)
layout.prop(context.window_manager, "filtered_search_prop", "")
box = layout.box()
length = len(context.window_manager.filtered_files_prop)
incolumn = int(length / context.blend_data.scenes[0].file_columnsnumber)
r = length % context.blend_data.scenes[0].file_columnsnumber
row = box.row()
col = row.column()
it = 0
tr = 0
for f in context.window_manager.filtered_files_prop:
op = col.operator("file.filtered_file_select", text=f.dname, emboss=False)
op.fname = f.name
op.fexec = context.blend_data.scenes[0].file_autoexecute
it += 1
if tr < r:
if it % (incolumn + 1) == 0:
tr += 1
if(it < length):
col = row.column()
else:
if (it - tr) % incolumn == 0:
if(it < length):
col = row.column()
layout.prop(context.blend_data.scenes[0], "file_autoexecute")
layout.prop(context.window_manager, "file_searchtree")
layout.prop(context.blend_data.scenes[0], "file_hideextensions")
layout.prop(context.blend_data.scenes[0], "file_columnsnumber")
def register():
| bpy.utils.register_module(__name__)
bpy.types.WindowManager.filtered_search_prop = bpy.props.StringProperty(update=filteredSearchFunc)
bpy.types.WindowManager.last_directory_prop = bpy.props.StringProperty()
bpy.types.Scene.file_autoexecute = bpy.props.BoolProperty(name="Open Automatically", default=True)
bpy.types.Scene.file_hideextensions = bpy.props.BoolProperty(name="Hide Extensions", update=filteredSearchFunc)
bpy.types.WindowManager.file_searchtree = bpy.props.BoolProperty(name="Search Subdirectories", update=filteredSearchFunc)
bpy.types.Scene.file_columnsnumber = bpy.props.IntProperty(name="Number of Columns", default=2, min=1, max=15, update=filteredSearchFunc)
bpy.types.WindowManager.filtered_files_prop = bpy.props.CollectionProperty(type=FilteredFileItem) | identifier_body |
|
io_file_browser_search.py | This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program; if not, write to the Free Software Foundation,
# Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
#
# ***** END GPL LICENCE BLOCK *****
bl_info = {
"name": "File Browser Search",
"author": "Jakub Zolcik",
"version": (0, 1, 1),
"blender": (2, 6, 2),
"api": 35622,
"location": "File Browser",
"description": "Allows You to find files in File Browser by name.",
"warning": "",
"wiki_url": "http://wiki.blender.org/index.php/User:Sftd/Extensions:2.6/Py/Scripts/Import-Export/File_Browser_Search",
"tracker_url": "http://projects.blender.org/tracker/?func=detail&aid=30386&group_id=153&atid=467",
"category": "Import-Export"}
"""
Usage:
Launches in File Browser
"""
import bpy
import os
import re
class FilteredFileItem(bpy.types.PropertyGroup):
name = bpy.props.StringProperty(name="File name", default="")
dname = bpy.props.StringProperty(name="Display name", default="")
def fileSearch(self, context):
# print("file")
filter = context.window_manager.filtered_search_prop
directory = context.window_manager.last_directory_prop
filecol = context.window_manager.filtered_files_prop
for fname in filecol:
filecol.remove(0)
if filter == "":
return None
pattern = ""
special = ('\\', '.', '^', '$', '*', '+', '?', '{', '}', '[', ']', '|', '(', ')')
for c in special:
filter = filter.replace(c, '\\' + c)
if ('*' in filter):
filter = filter.replace('\*', '.*')
pattern = ('^' + filter.lower() + '$')
else:
if(len(filter) < 3):
pattern = ('^' + filter.lower() + r".*\..*" + '$')
else:
pattern = ('^' + r".*" + filter.lower() + r".*\..*" + '$')
prog = re.compile(pattern)
maxf = 100
cf = 0
dlen = len(directory)
maxd = 100
cd = 0
if context.window_manager.file_searchtree:
for path, dirs, files in os.walk(directory):
cd += 1
if cd > maxd:
break
for filename in files:
filename = (os.path.join(path, filename))[dlen:]
# rfilename = (os.path.join(path, filename))[dlen:]
if prog.match(filename.lower()) != None:
p = context.window_manager.filtered_files_prop.add()
# p.name = rfilename
p.name = filename
if context.blend_data.scenes[0].file_hideextensions:
ind = filename.rfind(".")
if ind > -1:
filename = filename[0:ind]
p.dname = filename
cf += 1
if(cf >= maxf):
|
if(cf >= maxf):
break
else:
filesList = os.listdir(directory)
for filename in filesList:
if prog.match(filename.lower()) != None:
p = context.window_manager.filtered_files_prop.add()
p.name = filename
if context.blend_data.scenes[0].file_hideextensions:
ind = filename.rfind(".")
if ind > -1:
filename = filename[0:ind]
p.dname = filename
return None
def blendDataFromFile(file, part):
with bpy.data.libraries.load(file) as (data_from, data_to):
if (part == "Action"):
return data_from.actions
elif part == "Armature":
return data_from.brushes
elif part == "Brush":
return data_from.brushes
elif part == "Camera":
return data_from.cameras
elif part == "Curve":
return data_from.curves
elif part == "Font":
return data_from.fonts
elif part == "Group":
return data_from.groups
elif part == "Image":
return data_from.images
elif part == "Lamp":
return data_from.lamps
elif part == "Lattice":
return data_from.lattices
elif part == "Library":
return data_from.libraries
elif part == "FreestyleLineStyle":
return data_from.linestyles
elif part == "Mask":
return data_from.masks
elif part == "Material":
return data_from.materials
elif part == "Mesh":
return data_from.meshes
elif part == "NodeTree":
return data_from.node_groups
elif part == "Object":
return data_from.objects
elif part == "Particle":
return data_from.particles
elif part == "Scene":
return data_from.scenes
elif part == "Screen":
return data_from.screens
elif part == "Script":
return data_from.scripts
elif part == "Sound":
return data_from.sounds
elif part == "Speaker":
return data_from.speakers
elif part == "Text":
return data_from.texts
elif part == "Texture":
return data_from.textures
elif part == "World":
return data_from.worlds
else:
return None
def notFileSearch(self, context):
# print("not file")
filter = context.window_manager.filtered_search_prop
directory = context.window_manager.last_directory_prop
filecol = context.window_manager.filtered_files_prop
for fname in filecol:
filecol.remove(0)
if filter == "":
return None
ind_e = directory.find(".blend")
if(ind_e == -1):
return None
ind_e = ind_e + 6
file = directory[0:ind_e]
part = directory[ind_e + 1:-1]
if (part == ""):
return None
data = None
data = blendDataFromFile(file, part)
pattern = ""
if(len(filter) < 3):
pattern = (filter.lower() + r".*")
else:
pattern = (r".*" + filter.lower() + r".*")
prog = re.compile(pattern)
for name in data:
if prog.match(name.lower()) != None:
p = context.window_manager.filtered_files_prop.add()
p.name = name
p.dname = name
return None
def filteredSearchFunc(self, context):
if(context.active_operator.bl_idname == "WM_OT_link_append"):
return notFileSearch(self, context)
else:
return fileSearch(self, context)
class FilteredFileSelectOperator(bpy.types.Operator):
bl_idname = "file.filtered_file_select"
bl_label = "Select File"
fname = bpy.props.StringProperty()
fexec = bpy.props.BoolProperty()
def execute(self, context):
context.space_data.params.filename = self.fname
if self.fexec:
bpy.ops.file.execute('INVOKE_DEFAULT')
return {'FINISHED'}
class FilteredSearchPanel(bpy.types.Panel):
bl_idname = "FILE_PT_filteredsearch"
bl_label = "Search:"
bl_space_type = 'FILE_BROWSER'
bl_region_type = 'CHANNELS'
@classmethod
def poll(cls, context):
return (context.space_data.params is not None)
def draw(self, context):
layout = self.layout
directory = context.space_data.params.directory
if context.window_manager.last_directory_prop != directory:
context.window_manager.last_directory_prop = directory
filteredSearchFunc(self, context)
layout.prop(context.window_manager, "filtered_search_prop", "")
box = layout.box()
length = len(context.window_manager.filtered_files_prop)
incolumn = int(length / context.blend_data.scenes[0].file_columnsnumber)
r = length % context.blend_data.scenes[0].file_columnsnumber
row = box.row()
col = row.column()
it = 0
tr = 0
for f in context.window_manager.filtered_files_prop:
op = col.operator("file.filtered_file_select", text=f.dname, emboss=False)
op.fname = f.name
op.fexec = context.blend_data.scenes[0].file_autoexecute
it += 1
if tr < r:
if it % (incolumn + 1) == 0:
tr += 1
if(it < length):
col = row.column()
else:
if (it - tr) % incolumn == 0:
if(it < length):
col = row.column()
layout.prop(context.blend_data.scenes[0], "file_autoexecute")
layout.prop(context.window_manager, "file_searchtree")
layout.prop(context.blend_data.scenes[0], "file_hideextensions")
layout.prop(context.blend_data.scenes[0], | break | conditional_block |
node_dir_ops.go | err = nametransform.WriteDirIVAt(dirfd2)
syscall.Close(dirfd2)
}
if err != nil {
// Delete inconsistent directory (missing gocryptfs.diriv!)
err2 := syscallcompat.Unlinkat(dirfd, cName, unix.AT_REMOVEDIR)
if err2 != nil {
tlog.Warn.Printf("mkdirWithIv: rollback failed: %v", err2)
}
}
return err
}
// Mkdir - FUSE call. Create a directory at "newPath" with permissions "mode".
//
// Symlink-safe through use of Mkdirat().
func (n *Node) Mkdir(ctx context.Context, name string, mode uint32, out *fuse.EntryOut) (*fs.Inode, syscall.Errno) {
dirfd, cName, errno := n.prepareAtSyscall(name)
if errno != 0 {
return nil, errno
}
defer syscall.Close(dirfd)
rn := n.rootNode()
var context *fuse.Context
if rn.args.PreserveOwner {
context = toFuseCtx(ctx)
}
var st syscall.Stat_t
if rn.args.PlaintextNames {
err := syscallcompat.MkdiratUser(dirfd, cName, mode, context)
if err != nil {
return nil, fs.ToErrno(err)
}
var ust unix.Stat_t
err = syscallcompat.Fstatat(dirfd, cName, &ust, unix.AT_SYMLINK_NOFOLLOW)
if err != nil {
return nil, fs.ToErrno(err)
}
st = syscallcompat.Unix2syscall(ust)
// Create child node & return
ch := n.newChild(ctx, &st, out)
return ch, 0
}
// We need write and execute permissions to create gocryptfs.diriv.
// Also, we need read permissions to open the directory (to avoid
// race-conditions between getting and setting the mode).
origMode := mode
mode = mode | 0700
// Handle long file name
if nametransform.IsLongContent(cName) {
// Create ".name"
err := rn.nameTransform.WriteLongNameAt(dirfd, cName, name)
if err != nil {
return nil, fs.ToErrno(err)
}
// Create directory & rollback .name file on error
err = rn.mkdirWithIv(dirfd, cName, mode, context)
if err != nil {
nametransform.DeleteLongNameAt(dirfd, cName)
return nil, fs.ToErrno(err)
}
} else {
err := rn.mkdirWithIv(dirfd, cName, mode, context)
if err != nil {
return nil, fs.ToErrno(err)
}
}
// Fill `st`
fd, err := syscallcompat.Openat(dirfd, cName,
syscall.O_RDONLY|syscall.O_DIRECTORY|syscall.O_NOFOLLOW, 0)
if err != nil {
tlog.Warn.Printf("Mkdir %q: Openat failed: %v", cName, err)
return nil, fs.ToErrno(err)
}
defer syscall.Close(fd)
err = syscall.Fstat(fd, &st)
if err != nil {
tlog.Warn.Printf("Mkdir %q: Fstat failed: %v", cName, err)
return nil, fs.ToErrno(err)
}
// Fix permissions
if origMode != mode {
// Preserve SGID bit if it was set due to inheritance.
origMode = uint32(st.Mode&^0777) | origMode
err = syscall.Fchmod(fd, origMode)
if err != nil {
tlog.Warn.Printf("Mkdir %q: Fchmod %#o -> %#o failed: %v", cName, mode, origMode, err)
}
}
// Create child node & return
ch := n.newChild(ctx, &st, out)
return ch, 0
}
// Readdir - FUSE call.
//
// This function is symlink-safe through use of openBackingDir() and
// ReadDirIVAt().
func (n *Node) Readdir(ctx context.Context) (fs.DirStream, syscall.Errno) {
parentDirFd, cDirName, errno := n.prepareAtSyscallMyself()
if errno != 0 {
return nil, errno
}
defer syscall.Close(parentDirFd)
// Read ciphertext directory
fd, err := syscallcompat.Openat(parentDirFd, cDirName, syscall.O_RDONLY|syscall.O_DIRECTORY|syscall.O_NOFOLLOW, 0)
if err != nil {
return nil, fs.ToErrno(err)
}
defer syscall.Close(fd)
cipherEntries, specialEntries, err := syscallcompat.GetdentsSpecial(fd)
if err != nil {
return nil, fs.ToErrno(err)
}
// Get DirIV (stays nil if PlaintextNames is used)
var cachedIV []byte
rn := n.rootNode()
if !rn.args.PlaintextNames {
// Read the DirIV from disk
cachedIV, err = rn.nameTransform.ReadDirIVAt(fd)
if err != nil {
tlog.Warn.Printf("OpenDir %q: could not read %s: %v", cDirName, nametransform.DirIVFilename, err)
return nil, syscall.EIO
}
}
// Decrypted directory entries
var plain []fuse.DirEntry
// Add "." and ".."
plain = append(plain, specialEntries...)
// Filter and decrypt filenames
for i := range cipherEntries {
cName := cipherEntries[i].Name
if n.IsRoot() && cName == configfile.ConfDefaultName {
// silently ignore "gocryptfs.conf" in the top level dir
continue
}
if rn.args.PlaintextNames {
plain = append(plain, cipherEntries[i])
continue
}
if !rn.args.DeterministicNames && cName == nametransform.DirIVFilename {
// silently ignore "gocryptfs.diriv" everywhere if dirIV is enabled
continue
}
// Handle long file name
isLong := nametransform.LongNameNone
if rn.args.LongNames {
isLong = nametransform.NameType(cName)
}
if isLong == nametransform.LongNameContent {
cNameLong, err := nametransform.ReadLongNameAt(fd, cName)
if err != nil {
tlog.Warn.Printf("OpenDir %q: invalid entry %q: Could not read .name: %v",
cDirName, cName, err)
rn.reportMitigatedCorruption(cName)
continue
}
cName = cNameLong
} else if isLong == nametransform.LongNameFilename {
// ignore "gocryptfs.longname.*.name"
continue
}
name, err := rn.nameTransform.DecryptName(cName, cachedIV)
if err != nil {
tlog.Warn.Printf("OpenDir %q: invalid entry %q: %v",
cDirName, cName, err)
rn.reportMitigatedCorruption(cName)
continue
}
// Override the ciphertext name with the plaintext name but reuse the rest
// of the structure
cipherEntries[i].Name = name
plain = append(plain, cipherEntries[i])
}
return fs.NewListDirStream(plain), 0
}
// Rmdir - FUSE call.
//
// Symlink-safe through Unlinkat() + AT_REMOVEDIR.
func (n *Node) Rmdir(ctx context.Context, name string) (code syscall.Errno) {
rn := n.rootNode()
parentDirFd, cName, errno := n.prepareAtSyscall(name)
if errno != 0 |
defer syscall.Close(parentDirFd)
if rn.args.PlaintextNames {
// Unlinkat with AT_REMOVEDIR is equivalent to Rmdir
err := unix.Unlinkat(parentDirFd, cName, unix.AT_REMOVEDIR)
return fs.ToErrno(err)
}
if rn.args.DeterministicNames {
if err := unix.Unlinkat(parentDirFd, cName, unix.AT_REMOVEDIR); err != nil {
return fs.ToErrno(err)
}
if nametransform.IsLongContent(cName) {
nametransform.DeleteLongNameAt(parentDirFd, cName)
}
return 0
}
// Unless we are running as root, we need read, write and execute permissions
// to handle gocryptfs.diriv.
permWorkaround := false
var origMode uint32
if !rn.args.PreserveOwner {
var st unix.Stat_t
err := syscallcompat.Fstatat(parentDirFd, cName, &st, unix.AT_SYMLINK_NOFOLLOW)
if err != nil {
return fs.ToErrno(err)
}
if st.Mode&0700 != 0700 {
tlog.Debug.Printf("Rmdir: permWorkaround")
permWorkaround = true
// This cast is needed on Darwin, where st.Mode is uint1 | {
return errno
} | conditional_block |
node_dir_ops.go | err = nametransform.WriteDirIVAt(dirfd2)
syscall.Close(dirfd2)
}
if err != nil {
// Delete inconsistent directory (missing gocryptfs.diriv!)
err2 := syscallcompat.Unlinkat(dirfd, cName, unix.AT_REMOVEDIR)
if err2 != nil {
tlog.Warn.Printf("mkdirWithIv: rollback failed: %v", err2)
}
}
return err
}
// Mkdir - FUSE call. Create a directory at "newPath" with permissions "mode".
//
// Symlink-safe through use of Mkdirat().
func (n *Node) Mkdir(ctx context.Context, name string, mode uint32, out *fuse.EntryOut) (*fs.Inode, syscall.Errno) {
dirfd, cName, errno := n.prepareAtSyscall(name)
if errno != 0 {
return nil, errno
}
defer syscall.Close(dirfd)
rn := n.rootNode()
var context *fuse.Context
if rn.args.PreserveOwner {
context = toFuseCtx(ctx)
}
var st syscall.Stat_t
if rn.args.PlaintextNames {
err := syscallcompat.MkdiratUser(dirfd, cName, mode, context)
if err != nil {
return nil, fs.ToErrno(err)
}
var ust unix.Stat_t
err = syscallcompat.Fstatat(dirfd, cName, &ust, unix.AT_SYMLINK_NOFOLLOW)
if err != nil {
return nil, fs.ToErrno(err)
}
st = syscallcompat.Unix2syscall(ust)
// Create child node & return
ch := n.newChild(ctx, &st, out)
return ch, 0
}
// We need write and execute permissions to create gocryptfs.diriv.
// Also, we need read permissions to open the directory (to avoid
// race-conditions between getting and setting the mode).
origMode := mode
mode = mode | 0700
// Handle long file name
if nametransform.IsLongContent(cName) {
// Create ".name"
err := rn.nameTransform.WriteLongNameAt(dirfd, cName, name)
if err != nil {
return nil, fs.ToErrno(err)
}
// Create directory & rollback .name file on error
err = rn.mkdirWithIv(dirfd, cName, mode, context)
if err != nil {
nametransform.DeleteLongNameAt(dirfd, cName)
return nil, fs.ToErrno(err)
}
} else {
err := rn.mkdirWithIv(dirfd, cName, mode, context)
if err != nil {
return nil, fs.ToErrno(err)
}
}
// Fill `st`
fd, err := syscallcompat.Openat(dirfd, cName,
syscall.O_RDONLY|syscall.O_DIRECTORY|syscall.O_NOFOLLOW, 0)
if err != nil {
tlog.Warn.Printf("Mkdir %q: Openat failed: %v", cName, err)
return nil, fs.ToErrno(err)
}
defer syscall.Close(fd)
err = syscall.Fstat(fd, &st)
if err != nil {
tlog.Warn.Printf("Mkdir %q: Fstat failed: %v", cName, err)
return nil, fs.ToErrno(err)
}
// Fix permissions
if origMode != mode {
// Preserve SGID bit if it was set due to inheritance.
origMode = uint32(st.Mode&^0777) | origMode
err = syscall.Fchmod(fd, origMode)
if err != nil {
tlog.Warn.Printf("Mkdir %q: Fchmod %#o -> %#o failed: %v", cName, mode, origMode, err)
}
}
// Create child node & return
ch := n.newChild(ctx, &st, out)
return ch, 0
}
// Readdir - FUSE call.
//
// This function is symlink-safe through use of openBackingDir() and
// ReadDirIVAt().
func (n *Node) Readdir(ctx context.Context) (fs.DirStream, syscall.Errno) {
parentDirFd, cDirName, errno := n.prepareAtSyscallMyself()
if errno != 0 {
return nil, errno
}
defer syscall.Close(parentDirFd)
// Read ciphertext directory
fd, err := syscallcompat.Openat(parentDirFd, cDirName, syscall.O_RDONLY|syscall.O_DIRECTORY|syscall.O_NOFOLLOW, 0)
if err != nil {
return nil, fs.ToErrno(err)
}
defer syscall.Close(fd)
cipherEntries, specialEntries, err := syscallcompat.GetdentsSpecial(fd)
if err != nil {
return nil, fs.ToErrno(err)
}
// Get DirIV (stays nil if PlaintextNames is used)
var cachedIV []byte
rn := n.rootNode()
if !rn.args.PlaintextNames {
// Read the DirIV from disk
cachedIV, err = rn.nameTransform.ReadDirIVAt(fd)
if err != nil {
tlog.Warn.Printf("OpenDir %q: could not read %s: %v", cDirName, nametransform.DirIVFilename, err)
return nil, syscall.EIO
}
}
// Decrypted directory entries
var plain []fuse.DirEntry
// Add "." and ".."
plain = append(plain, specialEntries...)
// Filter and decrypt filenames
for i := range cipherEntries {
cName := cipherEntries[i].Name
if n.IsRoot() && cName == configfile.ConfDefaultName {
// silently ignore "gocryptfs.conf" in the top level dir
continue
}
if rn.args.PlaintextNames {
plain = append(plain, cipherEntries[i])
continue
}
if !rn.args.DeterministicNames && cName == nametransform.DirIVFilename {
// silently ignore "gocryptfs.diriv" everywhere if dirIV is enabled
continue
}
// Handle long file name
isLong := nametransform.LongNameNone
if rn.args.LongNames {
isLong = nametransform.NameType(cName)
}
if isLong == nametransform.LongNameContent {
cNameLong, err := nametransform.ReadLongNameAt(fd, cName)
if err != nil {
tlog.Warn.Printf("OpenDir %q: invalid entry %q: Could not read .name: %v",
cDirName, cName, err)
rn.reportMitigatedCorruption(cName)
continue
}
cName = cNameLong
} else if isLong == nametransform.LongNameFilename {
// ignore "gocryptfs.longname.*.name"
continue
}
name, err := rn.nameTransform.DecryptName(cName, cachedIV)
if err != nil {
tlog.Warn.Printf("OpenDir %q: invalid entry %q: %v",
cDirName, cName, err)
rn.reportMitigatedCorruption(cName)
continue
}
// Override the ciphertext name with the plaintext name but reuse the rest
// of the structure
cipherEntries[i].Name = name
plain = append(plain, cipherEntries[i])
}
return fs.NewListDirStream(plain), 0
}
// Rmdir - FUSE call.
//
// Symlink-safe through Unlinkat() + AT_REMOVEDIR.
func (n *Node) Rmdir(ctx context.Context, name string) (code syscall.Errno) | }
// Unless we are running as root, we need read, write and execute permissions
// to handle gocryptfs.diriv.
permWorkaround := false
var origMode uint32
if !rn.args.PreserveOwner {
var st unix.Stat_t
err := syscallcompat.Fstatat(parentDirFd, cName, &st, unix.AT_SYMLINK_NOFOLLOW)
if err != nil {
return fs.ToErrno(err)
}
if st.Mode&0700 != 0700 {
tlog.Debug.Printf("Rmdir: permWorkaround")
permWorkaround = true
// This cast is needed on Darwin, where st.Mode is uint16 | {
rn := n.rootNode()
parentDirFd, cName, errno := n.prepareAtSyscall(name)
if errno != 0 {
return errno
}
defer syscall.Close(parentDirFd)
if rn.args.PlaintextNames {
// Unlinkat with AT_REMOVEDIR is equivalent to Rmdir
err := unix.Unlinkat(parentDirFd, cName, unix.AT_REMOVEDIR)
return fs.ToErrno(err)
}
if rn.args.DeterministicNames {
if err := unix.Unlinkat(parentDirFd, cName, unix.AT_REMOVEDIR); err != nil {
return fs.ToErrno(err)
}
if nametransform.IsLongContent(cName) {
nametransform.DeleteLongNameAt(parentDirFd, cName)
}
return 0 | identifier_body |
node_dir_ops.go | (dirfd int, cName string, mode uint32, context *fuse.Context) error {
rn := n.rootNode()
if rn.args.DeterministicNames {
return syscallcompat.MkdiratUser(dirfd, cName, mode, context)
}
// Between the creation of the directory and the creation of gocryptfs.diriv
// the directory is inconsistent. Take the lock to prevent other readers
// from seeing it.
rn.dirIVLock.Lock()
defer rn.dirIVLock.Unlock()
err := syscallcompat.MkdiratUser(dirfd, cName, mode, context)
if err != nil {
return err
}
dirfd2, err := syscallcompat.Openat(dirfd, cName, syscall.O_DIRECTORY|syscall.O_NOFOLLOW|syscallcompat.O_PATH, 0)
if err == nil {
// Create gocryptfs.diriv
err = nametransform.WriteDirIVAt(dirfd2)
syscall.Close(dirfd2)
}
if err != nil {
// Delete inconsistent directory (missing gocryptfs.diriv!)
err2 := syscallcompat.Unlinkat(dirfd, cName, unix.AT_REMOVEDIR)
if err2 != nil {
tlog.Warn.Printf("mkdirWithIv: rollback failed: %v", err2)
}
}
return err
}
// Mkdir - FUSE call. Create a directory at "newPath" with permissions "mode".
//
// Symlink-safe through use of Mkdirat().
func (n *Node) Mkdir(ctx context.Context, name string, mode uint32, out *fuse.EntryOut) (*fs.Inode, syscall.Errno) {
dirfd, cName, errno := n.prepareAtSyscall(name)
if errno != 0 {
return nil, errno
}
defer syscall.Close(dirfd)
rn := n.rootNode()
var context *fuse.Context
if rn.args.PreserveOwner {
context = toFuseCtx(ctx)
}
var st syscall.Stat_t
if rn.args.PlaintextNames {
err := syscallcompat.MkdiratUser(dirfd, cName, mode, context)
if err != nil {
return nil, fs.ToErrno(err)
}
var ust unix.Stat_t
err = syscallcompat.Fstatat(dirfd, cName, &ust, unix.AT_SYMLINK_NOFOLLOW)
if err != nil {
return nil, fs.ToErrno(err)
}
st = syscallcompat.Unix2syscall(ust)
// Create child node & return
ch := n.newChild(ctx, &st, out)
return ch, 0
}
// We need write and execute permissions to create gocryptfs.diriv.
// Also, we need read permissions to open the directory (to avoid
// race-conditions between getting and setting the mode).
origMode := mode
mode = mode | 0700
// Handle long file name
if nametransform.IsLongContent(cName) {
// Create ".name"
err := rn.nameTransform.WriteLongNameAt(dirfd, cName, name)
if err != nil {
return nil, fs.ToErrno(err)
}
// Create directory & rollback .name file on error
err = rn.mkdirWithIv(dirfd, cName, mode, context)
if err != nil {
nametransform.DeleteLongNameAt(dirfd, cName)
return nil, fs.ToErrno(err)
}
} else {
err := rn.mkdirWithIv(dirfd, cName, mode, context)
if err != nil {
return nil, fs.ToErrno(err)
}
}
// Fill `st`
fd, err := syscallcompat.Openat(dirfd, cName,
syscall.O_RDONLY|syscall.O_DIRECTORY|syscall.O_NOFOLLOW, 0)
if err != nil {
tlog.Warn.Printf("Mkdir %q: Openat failed: %v", cName, err)
return nil, fs.ToErrno(err)
}
defer syscall.Close(fd)
err = syscall.Fstat(fd, &st)
if err != nil {
tlog.Warn.Printf("Mkdir %q: Fstat failed: %v", cName, err)
return nil, fs.ToErrno(err)
}
// Fix permissions
if origMode != mode {
// Preserve SGID bit if it was set due to inheritance.
origMode = uint32(st.Mode&^0777) | origMode
err = syscall.Fchmod(fd, origMode)
if err != nil {
tlog.Warn.Printf("Mkdir %q: Fchmod %#o -> %#o failed: %v", cName, mode, origMode, err)
}
}
// Create child node & return
ch := n.newChild(ctx, &st, out)
return ch, 0
}
// Readdir - FUSE call.
//
// This function is symlink-safe through use of openBackingDir() and
// ReadDirIVAt().
func (n *Node) Readdir(ctx context.Context) (fs.DirStream, syscall.Errno) {
parentDirFd, cDirName, errno := n.prepareAtSyscallMyself()
if errno != 0 {
return nil, errno
}
defer syscall.Close(parentDirFd)
// Read ciphertext directory
fd, err := syscallcompat.Openat(parentDirFd, cDirName, syscall.O_RDONLY|syscall.O_DIRECTORY|syscall.O_NOFOLLOW, 0)
if err != nil {
return nil, fs.ToErrno(err)
}
defer syscall.Close(fd)
cipherEntries, specialEntries, err := syscallcompat.GetdentsSpecial(fd)
if err != nil {
return nil, fs.ToErrno(err)
}
// Get DirIV (stays nil if PlaintextNames is used)
var cachedIV []byte
rn := n.rootNode()
if !rn.args.PlaintextNames {
// Read the DirIV from disk
cachedIV, err = rn.nameTransform.ReadDirIVAt(fd)
if err != nil {
tlog.Warn.Printf("OpenDir %q: could not read %s: %v", cDirName, nametransform.DirIVFilename, err)
return nil, syscall.EIO
}
}
// Decrypted directory entries
var plain []fuse.DirEntry
// Add "." and ".."
plain = append(plain, specialEntries...)
// Filter and decrypt filenames
for i := range cipherEntries {
cName := cipherEntries[i].Name
if n.IsRoot() && cName == configfile.ConfDefaultName {
// silently ignore "gocryptfs.conf" in the top level dir
continue
}
if rn.args.PlaintextNames {
plain = append(plain, cipherEntries[i])
continue
}
if !rn.args.DeterministicNames && cName == nametransform.DirIVFilename {
// silently ignore "gocryptfs.diriv" everywhere if dirIV is enabled
continue
}
// Handle long file name
isLong := nametransform.LongNameNone
if rn.args.LongNames {
isLong = nametransform.NameType(cName)
}
if isLong == nametransform.LongNameContent {
cNameLong, err := nametransform.ReadLongNameAt(fd, cName)
if err != nil {
tlog.Warn.Printf("OpenDir %q: invalid entry %q: Could not read .name: %v",
cDirName, cName, err)
rn.reportMitigatedCorruption(cName)
continue
}
cName = cNameLong
} else if isLong == nametransform.LongNameFilename {
// ignore "gocryptfs.longname.*.name"
continue
}
name, err := rn.nameTransform.DecryptName(cName, cachedIV)
if err != nil {
tlog.Warn.Printf("OpenDir %q: invalid entry %q: %v",
cDirName, cName, err)
rn.reportMitigatedCorruption(cName)
continue
}
// Override the ciphertext name with the plaintext name but reuse the rest
// of the structure
cipherEntries[i].Name = name
plain = append(plain, cipherEntries[i])
}
return fs.NewListDirStream(plain), 0
}
// Rmdir - FUSE call.
//
// Symlink-safe through Unlinkat() + AT_REMOVEDIR.
func (n *Node) Rmdir(ctx context.Context, name string) (code syscall.Errno) {
rn := n.rootNode()
parentDirFd, cName, errno := n.prepareAtSyscall(name)
if errno != 0 {
return errno
}
defer syscall.Close(parentDirFd)
if rn.args.PlaintextNames {
// Unlinkat with AT_REMOVEDIR is equivalent to Rmdir
err := unix.Unlinkat(parentDirFd, cName, unix.AT_REMOVEDIR)
return fs.ToErrno(err)
}
if rn.args.DeterministicNames {
if err := unix.Unlinkat(parentDirFd, cName, unix.AT_REMOVEDIR); err != nil {
return fs.ToErrno(err)
}
if nametrans | mkdirWithIv | identifier_name |
|
node_dir_ops.go | err = nametransform.WriteDirIVAt(dirfd2)
syscall.Close(dirfd2)
}
if err != nil { | // Delete inconsistent directory (missing gocryptfs.diriv!)
err2 := syscallcompat.Unlinkat(dirfd, cName, unix.AT_REMOVEDIR)
if err2 != nil {
tlog.Warn.Printf("mkdirWithIv: rollback failed: %v", err2)
}
}
return err
}
// Mkdir - FUSE call. Create a directory at "newPath" with permissions "mode".
//
// Symlink-safe through use of Mkdirat().
func (n *Node) Mkdir(ctx context.Context, name string, mode uint32, out *fuse.EntryOut) (*fs.Inode, syscall.Errno) {
dirfd, cName, errno := n.prepareAtSyscall(name)
if errno != 0 {
return nil, errno
}
defer syscall.Close(dirfd)
rn := n.rootNode()
var context *fuse.Context
if rn.args.PreserveOwner {
context = toFuseCtx(ctx)
}
var st syscall.Stat_t
if rn.args.PlaintextNames {
err := syscallcompat.MkdiratUser(dirfd, cName, mode, context)
if err != nil {
return nil, fs.ToErrno(err)
}
var ust unix.Stat_t
err = syscallcompat.Fstatat(dirfd, cName, &ust, unix.AT_SYMLINK_NOFOLLOW)
if err != nil {
return nil, fs.ToErrno(err)
}
st = syscallcompat.Unix2syscall(ust)
// Create child node & return
ch := n.newChild(ctx, &st, out)
return ch, 0
}
// We need write and execute permissions to create gocryptfs.diriv.
// Also, we need read permissions to open the directory (to avoid
// race-conditions between getting and setting the mode).
origMode := mode
mode = mode | 0700
// Handle long file name
if nametransform.IsLongContent(cName) {
// Create ".name"
err := rn.nameTransform.WriteLongNameAt(dirfd, cName, name)
if err != nil {
return nil, fs.ToErrno(err)
}
// Create directory & rollback .name file on error
err = rn.mkdirWithIv(dirfd, cName, mode, context)
if err != nil {
nametransform.DeleteLongNameAt(dirfd, cName)
return nil, fs.ToErrno(err)
}
} else {
err := rn.mkdirWithIv(dirfd, cName, mode, context)
if err != nil {
return nil, fs.ToErrno(err)
}
}
// Fill `st`
fd, err := syscallcompat.Openat(dirfd, cName,
syscall.O_RDONLY|syscall.O_DIRECTORY|syscall.O_NOFOLLOW, 0)
if err != nil {
tlog.Warn.Printf("Mkdir %q: Openat failed: %v", cName, err)
return nil, fs.ToErrno(err)
}
defer syscall.Close(fd)
err = syscall.Fstat(fd, &st)
if err != nil {
tlog.Warn.Printf("Mkdir %q: Fstat failed: %v", cName, err)
return nil, fs.ToErrno(err)
}
// Fix permissions
if origMode != mode {
// Preserve SGID bit if it was set due to inheritance.
origMode = uint32(st.Mode&^0777) | origMode
err = syscall.Fchmod(fd, origMode)
if err != nil {
tlog.Warn.Printf("Mkdir %q: Fchmod %#o -> %#o failed: %v", cName, mode, origMode, err)
}
}
// Create child node & return
ch := n.newChild(ctx, &st, out)
return ch, 0
}
// Readdir - FUSE call.
//
// This function is symlink-safe through use of openBackingDir() and
// ReadDirIVAt().
func (n *Node) Readdir(ctx context.Context) (fs.DirStream, syscall.Errno) {
parentDirFd, cDirName, errno := n.prepareAtSyscallMyself()
if errno != 0 {
return nil, errno
}
defer syscall.Close(parentDirFd)
// Read ciphertext directory
fd, err := syscallcompat.Openat(parentDirFd, cDirName, syscall.O_RDONLY|syscall.O_DIRECTORY|syscall.O_NOFOLLOW, 0)
if err != nil {
return nil, fs.ToErrno(err)
}
defer syscall.Close(fd)
cipherEntries, specialEntries, err := syscallcompat.GetdentsSpecial(fd)
if err != nil {
return nil, fs.ToErrno(err)
}
// Get DirIV (stays nil if PlaintextNames is used)
var cachedIV []byte
rn := n.rootNode()
if !rn.args.PlaintextNames {
// Read the DirIV from disk
cachedIV, err = rn.nameTransform.ReadDirIVAt(fd)
if err != nil {
tlog.Warn.Printf("OpenDir %q: could not read %s: %v", cDirName, nametransform.DirIVFilename, err)
return nil, syscall.EIO
}
}
// Decrypted directory entries
var plain []fuse.DirEntry
// Add "." and ".."
plain = append(plain, specialEntries...)
// Filter and decrypt filenames
for i := range cipherEntries {
cName := cipherEntries[i].Name
if n.IsRoot() && cName == configfile.ConfDefaultName {
// silently ignore "gocryptfs.conf" in the top level dir
continue
}
if rn.args.PlaintextNames {
plain = append(plain, cipherEntries[i])
continue
}
if !rn.args.DeterministicNames && cName == nametransform.DirIVFilename {
// silently ignore "gocryptfs.diriv" everywhere if dirIV is enabled
continue
}
// Handle long file name
isLong := nametransform.LongNameNone
if rn.args.LongNames {
isLong = nametransform.NameType(cName)
}
if isLong == nametransform.LongNameContent {
cNameLong, err := nametransform.ReadLongNameAt(fd, cName)
if err != nil {
tlog.Warn.Printf("OpenDir %q: invalid entry %q: Could not read .name: %v",
cDirName, cName, err)
rn.reportMitigatedCorruption(cName)
continue
}
cName = cNameLong
} else if isLong == nametransform.LongNameFilename {
// ignore "gocryptfs.longname.*.name"
continue
}
name, err := rn.nameTransform.DecryptName(cName, cachedIV)
if err != nil {
tlog.Warn.Printf("OpenDir %q: invalid entry %q: %v",
cDirName, cName, err)
rn.reportMitigatedCorruption(cName)
continue
}
// Override the ciphertext name with the plaintext name but reuse the rest
// of the structure
cipherEntries[i].Name = name
plain = append(plain, cipherEntries[i])
}
return fs.NewListDirStream(plain), 0
}
// Rmdir - FUSE call.
//
// Symlink-safe through Unlinkat() + AT_REMOVEDIR.
func (n *Node) Rmdir(ctx context.Context, name string) (code syscall.Errno) {
rn := n.rootNode()
parentDirFd, cName, errno := n.prepareAtSyscall(name)
if errno != 0 {
return errno
}
defer syscall.Close(parentDirFd)
if rn.args.PlaintextNames {
// Unlinkat with AT_REMOVEDIR is equivalent to Rmdir
err := unix.Unlinkat(parentDirFd, cName, unix.AT_REMOVEDIR)
return fs.ToErrno(err)
}
if rn.args.DeterministicNames {
if err := unix.Unlinkat(parentDirFd, cName, unix.AT_REMOVEDIR); err != nil {
return fs.ToErrno(err)
}
if nametransform.IsLongContent(cName) {
nametransform.DeleteLongNameAt(parentDirFd, cName)
}
return 0
}
// Unless we are running as root, we need read, write and execute permissions
// to handle gocryptfs.diriv.
permWorkaround := false
var origMode uint32
if !rn.args.PreserveOwner {
var st unix.Stat_t
err := syscallcompat.Fstatat(parentDirFd, cName, &st, unix.AT_SYMLINK_NOFOLLOW)
if err != nil {
return fs.ToErrno(err)
}
if st.Mode&0700 != 0700 {
tlog.Debug.Printf("Rmdir: permWorkaround")
permWorkaround = true
// This cast is needed on Darwin, where st.Mode is uint16 | random_line_split |
|
engine.py | * self.margin):
# we use the same coef as the noise coef
#cell.diffusive_mass += val * self.environment.sigma
def log_status(self):
if self.datalog == None:
return
row = []
#row.append(self.iteration)
dm = [cell.diffusive_mass for cell in self.cells if cell]
row.append(sum(dm))
cm = [cell.crystal_mass for cell in self.cells if cell]
row.append(sum(cm))
bm = [cell.boundary_mass for cell in self.cells if cell]
row.append(sum(bm))
acnt = len([cell for cell in self.cells if cell and cell.attached])
row.append(acnt)
bcnt = len([cell for cell in self.cells if cell and cell.boundary])
row.append(bcnt)
d = self.snowflake_radius()
row.append(d)
row.append(self.environment.beta)
row.append(self.environment.theta)
row.append(self.environment.alpha)
row.append(self.environment.kappa)
row.append(self.environment.mu)
row.append(self.environment.upsilon)
#row.append(self.environment.sigma)
#row.append(self.environment.gamma)
self.datalog.append(row)
# log the cells
self.celllog.append((self.iteration, dm, cm))
def write_log(self):
self.write_datalog()
self.write_celllog()
def write_datalog(self):
if self.datalog == None:
return
logfn = "datalog.csv"
msg = "Saving runtime data to %s" % logfn
log(msg)
f = open(logfn, 'w')
txt = ''
txt += str.join(',', self.LogHeader) + '\n'
for row in self.datalog:
txt += str.join(',', map(str, row)) + '\n'
f.write(txt)
def write_celllog(self):
if not self.celllog:
return
logfn = "cell_log_%d.pickle" % self.iteration
f = open(logfn, 'wb')
pickle.dump(self.celllog, f, protocol=-1)
self.celllog = []
def print_status(self):
dm = sum([cell.diffusive_mass for cell in self.cells if cell])
cm = sum([cell.crystal_mass for cell in self.cells if cell])
bm = sum([cell.boundary_mass for cell in self.cells if cell])
acnt = len([cell for cell in self.cells if cell and cell.attached])
bcnt = len([cell for cell in self.cells if cell and cell.boundary])
#msg = "Step #%d, %d attached, %d boundary, %.2f dM, %.2f bM, %.2f cM, tot %.2f M" % (self.iteration, acnt, bcnt, dm, bm, cm, dm + cm + bm)
d = self.snowflake_radius()
msg = "Step #%d/%dp (%.2f%% scl), %d/%d (%.2f%%), %.2f dM, %.2f bM, %.2f cM, tot %.2f M" % (self.iteration, d, (float(d * 2 * X_SCALE_FACTOR) / self.iteration) * 100, acnt, bcnt, (float(bcnt) / acnt) * 100, dm, bm, cm, dm + cm + bm)
log(msg)
def step(self):
self.log_status()
for cell in self.cells:
if cell == None or cell.attached:
continue
cell.step_one()
for cell in self.cells:
if cell == None or cell.attached:
continue
cell.step_two()
for cell in self.cells:
if cell == None or cell.attached:
continue
cell.step_three()
# run curves
self.iteration += 1
self.environment.step(self.iteration)
def translate_xy(self, xy):
(x, y) = xy
x = int(round(x * X_SCALE_FACTOR))
return (x, y)
def polar_to_xy(self, args):
(angle, distance) = args
half = self.size / 2.0
angle = math.radians(angle)
y = int(round(half - (math.sin(angle) * distance)))
x = int(round(half + (math.cos(angle) * distance)))
return (x, y)
def xy_to_polar(self, args):
(x, y) = args
half = self.size / 2.0
x -= half
y += half
angle = math.degrees(math.atan2(y, x))
distance = math.hypot(x, y)
return (angle, distance)
def snowflake_radius(self, angle=135):
# we cast a ray on the 135 degeree axis
radius = 0
half = self.size / 2.0
while radius < half:
radius += 1
xy = self.polar_to_xy((angle, radius))
cell = self.cells[self._cell_index(xy)]
if cell.attached or cell.boundary:
continue
return radius
# uhh
return int(round(half))
def crop_snowflake(self, margin=None):
def scale(val):
return int(round(X_SCALE_FACTOR * val))
if margin == None:
margin = 15
half = self.size / 2
radius = scale(self.snowflake_radius())
distance = min(radius + margin, half)
half_s = scale(half)
distance_s = scale(distance)
box = (half_s - distance, half - distance, half_s + distance, half + distance)
return box
def headroom(self, margin=None):
if self.max_steps and self.iteration >= self.max_steps:
return False
if margin == None:
margin = self.margin
assert margin > 0 and margin <= 1
cutoff = int(round(margin * (self.size / 2.0)))
radius = self.snowflake_radius()
if radius > cutoff:
return False
return True
def grow(self):
while True:
if self.debug:
self.print_status()
self.step()
if self.iteration % 50 == 0:
self.write_celllog()
if not self.debug:
self.print_status()
if not self.headroom():
break
if self.debug:
self.print_status()
def save_image(self, fn, **kw):
import sfgen
r = sfgen.RenderSnowflake(self)
r.save_image(fn, **kw)
class SnowflakeCell(object):
def __init__(self, xy, lattice):
self.xy = xy
self.lattice = lattice
self.env = lattice.environment
self.diffusive_mass = self.env.gamma
self.boundary_mass = 0.0
self.crystal_mass = 0.0
self.attached = False
self.age = 0
self.boundary = 0
self.attached_neighbors = []
self.__neighbors = None
def __getstate__(self):
return (self.xy, self.diffusive_mass, self.boundary_mass, self.crystal_mass, self.attached, self.age)
def __setstate__(self, state):
self.xy = state[0]
self.diffusive_mass = state[1]
self.boundary_mass = state[2]
self.crystal_mass = state[3]
self.attached = state[4]
# 0.2 -> 0.3
try:
self.age = state[5]
except IndexError:
self.age = 0
self.__neighbors = None
self.lattice = None
self.env = None
def reality_check(self):
assert len(self.neighbors)
for neighbor in self.neighbors:
assert self in neighbor.neighbors, "%s not in %s" % (str(self), str(neighbor.neighbors))
def __repr__(self):
return "(%d,%d)" % self.xy
@property
def neighbors(self):
if self.__neighbors == None:
self.__neighbors = self.lattice.get_neighbors(self.xy)
return self.__neighbors
#@property
#def attached_neighbors(self):
# return [cell for cell in self.neighbors if cell.attached]
#@property
#def boundary(self):
# return (not self.attached) and any([cell.attached for cell in self.neighbors])
def update_boundary(self):
self.boundary = (not self.attached) and any([cell.attached for cell in self.neighbors])
def step_one(self):
self.update_boundary()
if self.boundary:
self.attached_neighbors = [cell for cell in self.neighbors if cell.attached]
self._next_dm = self.diffusion_calc()
def step_two(self):
self.diffusive_mass = self._next_dm
self.attachment_flag = self.attached
self.freezing_step()
self.attachment_flag = self.attachment_step()
self.melting_step()
def step_three(self):
if self.boundary and self.attachment_flag:
self.attach()
self.noise_step()
def diffusion_calc(self):
next_dm = self.diffusive_mass | if self.attached:
return next_dm
self.age += 1
for cell in self.neighbors: | random_line_split |
|
engine.py | .cells[idx] = cell
self.reality_check()
center_pt = self._cell_index((self.size / 2, self.size / 2))
self.cells[center_pt].attach(1)
# fun experiments
#self.cells[center_pt+4].attach(1)
#self.cells[center_pt-4].attach(1)
def _xy_ok(self, xy):
(x, y) = xy
return (x >= 0 and x < self.size and y >= 0 and y < self.size)
def _cell_index(self, xy):
(x, y) = xy
return int(round(y * self.size + x))
def _cell_xy(self, idx):
y = idx / self.size
x = idx % self.size
return (x, y)
def adjust_humidity(self, val):
val = abs(val)
for cell in self.cells:
if cell.attached or cell.boundary:
continue
cell.diffusive_mass += val * self.environment.sigma
# only mutate the cells outside our margin
#if self.xy_to_polar(cell.xy)[1] > (self.size * self.margin):
# we use the same coef as the noise coef
#cell.diffusive_mass += val * self.environment.sigma
def log_status(self):
if self.datalog == None:
return
row = []
#row.append(self.iteration)
dm = [cell.diffusive_mass for cell in self.cells if cell]
row.append(sum(dm))
cm = [cell.crystal_mass for cell in self.cells if cell]
row.append(sum(cm))
bm = [cell.boundary_mass for cell in self.cells if cell]
row.append(sum(bm))
acnt = len([cell for cell in self.cells if cell and cell.attached])
row.append(acnt)
bcnt = len([cell for cell in self.cells if cell and cell.boundary])
row.append(bcnt)
d = self.snowflake_radius()
row.append(d)
row.append(self.environment.beta)
row.append(self.environment.theta)
row.append(self.environment.alpha)
row.append(self.environment.kappa)
row.append(self.environment.mu)
row.append(self.environment.upsilon)
#row.append(self.environment.sigma)
#row.append(self.environment.gamma)
self.datalog.append(row)
# log the cells
self.celllog.append((self.iteration, dm, cm))
def write_log(self):
self.write_datalog()
self.write_celllog()
def write_datalog(self):
if self.datalog == None:
return
logfn = "datalog.csv"
msg = "Saving runtime data to %s" % logfn
log(msg)
f = open(logfn, 'w')
txt = ''
txt += str.join(',', self.LogHeader) + '\n'
for row in self.datalog:
txt += str.join(',', map(str, row)) + '\n'
f.write(txt)
def write_celllog(self):
if not self.celllog:
return
logfn = "cell_log_%d.pickle" % self.iteration
f = open(logfn, 'wb')
pickle.dump(self.celllog, f, protocol=-1)
self.celllog = []
def print_status(self):
dm = sum([cell.diffusive_mass for cell in self.cells if cell])
cm = sum([cell.crystal_mass for cell in self.cells if cell])
bm = sum([cell.boundary_mass for cell in self.cells if cell])
acnt = len([cell for cell in self.cells if cell and cell.attached])
bcnt = len([cell for cell in self.cells if cell and cell.boundary])
#msg = "Step #%d, %d attached, %d boundary, %.2f dM, %.2f bM, %.2f cM, tot %.2f M" % (self.iteration, acnt, bcnt, dm, bm, cm, dm + cm + bm)
d = self.snowflake_radius()
msg = "Step #%d/%dp (%.2f%% scl), %d/%d (%.2f%%), %.2f dM, %.2f bM, %.2f cM, tot %.2f M" % (self.iteration, d, (float(d * 2 * X_SCALE_FACTOR) / self.iteration) * 100, acnt, bcnt, (float(bcnt) / acnt) * 100, dm, bm, cm, dm + cm + bm)
log(msg)
def step(self):
self.log_status()
for cell in self.cells:
if cell == None or cell.attached:
continue
cell.step_one()
for cell in self.cells:
if cell == None or cell.attached:
continue
cell.step_two()
for cell in self.cells:
if cell == None or cell.attached:
continue
cell.step_three()
# run curves
self.iteration += 1
self.environment.step(self.iteration)
def translate_xy(self, xy):
(x, y) = xy
x = int(round(x * X_SCALE_FACTOR))
return (x, y)
def polar_to_xy(self, args):
(angle, distance) = args
half = self.size / 2.0
angle = math.radians(angle)
y = int(round(half - (math.sin(angle) * distance)))
x = int(round(half + (math.cos(angle) * distance)))
return (x, y)
def xy_to_polar(self, args):
(x, y) = args
half = self.size / 2.0
x -= half
y += half
angle = math.degrees(math.atan2(y, x))
distance = math.hypot(x, y)
return (angle, distance)
def snowflake_radius(self, angle=135):
# we cast a ray on the 135 degeree axis
radius = 0
half = self.size / 2.0
while radius < half:
radius += 1
xy = self.polar_to_xy((angle, radius))
cell = self.cells[self._cell_index(xy)]
if cell.attached or cell.boundary:
continue
return radius
# uhh
return int(round(half))
def crop_snowflake(self, margin=None):
def scale(val):
return int(round(X_SCALE_FACTOR * val))
if margin == None:
margin = 15
half = self.size / 2
radius = scale(self.snowflake_radius())
distance = min(radius + margin, half)
half_s = scale(half)
distance_s = scale(distance)
box = (half_s - distance, half - distance, half_s + distance, half + distance)
return box
def headroom(self, margin=None):
if self.max_steps and self.iteration >= self.max_steps:
return False
if margin == None:
margin = self.margin
assert margin > 0 and margin <= 1
cutoff = int(round(margin * (self.size / 2.0)))
radius = self.snowflake_radius()
if radius > cutoff:
return False
return True
def grow(self):
while True:
if self.debug:
self.print_status()
self.step()
if self.iteration % 50 == 0:
self.write_celllog()
if not self.debug:
self.print_status()
if not self.headroom():
break
if self.debug:
self.print_status()
def save_image(self, fn, **kw):
import sfgen
r = sfgen.RenderSnowflake(self)
r.save_image(fn, **kw)
class SnowflakeCell(object):
def __init__(self, xy, lattice):
self.xy = xy
self.lattice = lattice
self.env = lattice.environment
self.diffusive_mass = self.env.gamma
self.boundary_mass = 0.0
self.crystal_mass = 0.0
self.attached = False
self.age = 0
self.boundary = 0
self.attached_neighbors = []
self.__neighbors = None
def __getstate__(self):
return (self.xy, self.diffusive_mass, self.boundary_mass, self.crystal_mass, self.attached, self.age)
def __setstate__(self, state):
self.xy = state[0]
self.diffusive_mass = state[1]
self.boundary_mass = state[2]
self.crystal_mass = state[3]
self.attached = state[4]
# 0.2 -> 0.3
try:
self.age = state[5]
except IndexError:
self.age = 0
self.__neighbors = None
self.lattice = None
self.env = None
def reality_check(self):
assert len(self.neighbors)
for neighbor in self.neighbors:
assert self in neighbor.neighbors, "%s not in %s" % (str(self), str(neighbor.neighbors))
def __repr__(self):
return "(%d,%d)" % self.xy
@property
def neighbors(self):
if self.__neighbors == None:
| self.__neighbors = self.lattice.get_neighbors(self.xy) | conditional_block |
|
engine.py |
def step(self, x):
if self.curves == None:
return
for key in self.curves:
self[key] = self.curves[key][x]
@classmethod
def build_env(self, name, steps, min_gamma=0.45, max_gamma=0.85):
curves = {
"beta": (1.3, 2),
"theta": (0.01, 0.04),
"alpha": (0.02, 0.1),
"kappa": (0.001, 0.01),
"mu": (0.01, 0.1),
"upilson": (0.00001, 0.0001),
"sigma": (0.00001, 0.000001),
}
cs = CurveSet(name, steps, curves)
cs.run_graph()
env = {key: cs[key][0] for key in curves}
env["gamma"] = random.random() * (max_gamma - min_gamma) + min_gamma
return CrystalEnvironment(curves=cs, **env)
def get_default(self, key):
return self.factory_settings[key]
def randomize(self):
for key in self:
if key == "sigma":
continue
if key == "gamma":
self[key] += 1.0 / random.randint(100, 1000)
else:
self[key] += random.choice([1.0, -1.0]) / random.randint(100, 1000)
self.set_factory_settings()
def _init_defaults(self):
# (3a)
# "A boundary site with 1 or 2 attached neighbors needs boundary mass at least beta to join the crystal
# This is the case when the local mesoscopic geometry near x corresponds to a tip or flat spot of the crystal.
# (Distinguishing the two cases turns out to be of minor significance.) In our simulations, beta is typically
# between about 1.05 and 3. We assume beta > 1 since 1 is the basic threshold of the case to follow next.
self["beta"] = 1.3
# (3b)
# "A boundary site with 3 attached neighbors joins the crystal if either it has boundary mass >= 1,
# or it has diffusive mass < theta in its neighborhood and it has boundary mass >= alpha"
self["theta"] = 0.025
self["alpha"] = 0.08
# (2)
# "Proportion kappa of the diffusive mass at each boundary site crystallizes.
# The remainder (proportion 1 - kappa) becomes boundary mass."
self["kappa"] = 0.003
# (4)
# "Proportion mu of the boundary mass and proportion upsilon of the crystal mass at each boundary site become diffusive mass.
# Melting represents mass flow at the boundary from ice and quasi-liquid back to vapor, reverse
# effects from the freezing of step ii. Typically mu is small and upsilon extremely small."
self["mu"] = 0.07
self["upsilon"] = 0.00005
# (5)
# "The diffusive mass at each site undergoes an independent random perturbation of proportion sigma"
self["sigma"] = 0.00001
# initial diffusion
self["gamma"] = 0.5
def _init_special(self):
pass
# class RenderMovie(object) moved to movie.py
# 2018-0212
# class LatticeReplay(object) moved to movie.py
# 2018-0212
class CrystalLattice(object):
LogHeader = ["dm", "cm", "bm", "acnt", "bcnt", "width", "beta", "theta", "alpha", "kappa", "mu", "upsilon"]
def __init__(self, size, environment=None, celltype=None, max_steps=0, margin=None, curves=None, datalog=False, debug=False):
self.size = size
if environment == None:
environment = CrystalEnvironment()
self.environment = environment
self.datalog = None
self.celllog = None
if datalog:
self.datalog = []
self.celllog = []
if celltype == None:
celltype = SnowflakeCell
self.debug = debug
self.celltype = celltype
self.iteration = 1
assert margin > 0 and margin <= 1.0
self.margin = margin
self.curves = curves
self.max_steps = max_steps
self._init_cells()
def __setstate__(self, state):
# 0.1->0.2 format changes
if "radius" in state:
state["size"] = state["radius"]
del state["radius"]
if "angle" in state:
del state["angle"]
self.__dict__.update(state)
def save_lattice(self, fn):
msg = "Saving %s..." % fn
log(msg)
f = open(fn, 'wb')
pickle.dump(self, f, protocol=-1)
@classmethod
def load_lattice(cls, fn):
msg = "Loading %s..." % fn
log(msg)
f = open(fn, 'rb')
obj = pickle.load(f)
for cell in obj.cells:
cell.lattice = obj
cell.env = obj.environment
cell.update_boundary()
return obj
def get_neighbors(self, xy):
(x, y) = xy
nlist = [(x, y + 1), (x, y - 1), (x - 1, y), (x + 1, y), (x - 1, y - 1), (x + 1, y + 1)]
nlist = map(self._cell_index, filter(self._xy_ok, nlist))
res = tuple([self.cells[nidx] for nidx in nlist if self.cells[nidx] != None])
return res
def reality_check(self):
for cell in self.cells:
cell.reality_check()
def _init_cells(self):
self.cells = [None] * (self.size * self.size)
for x in range(self.size):
for y in range(self.size):
xy = (x, y)
cell = self.celltype(xy, self)
idx = self._cell_index(xy)
self.cells[idx] = cell
self.reality_check()
center_pt = self._cell_index((self.size / 2, self.size / 2))
self.cells[center_pt].attach(1)
# fun experiments
#self.cells[center_pt+4].attach(1)
#self.cells[center_pt-4].attach(1)
def _xy_ok(self, xy):
(x, y) = xy
return (x >= 0 and x < self.size and y >= 0 and y < self.size)
def _cell_index(self, xy):
(x, y) = xy
return int(round(y * self.size + x))
def _cell_xy(self, idx):
y = idx / self.size
x = idx % self.size
return (x, y)
def adjust_humidity(self, val):
val = abs(val)
for cell in self.cells:
if cell.attached or cell.boundary:
continue
cell.diffusive_mass += val * self.environment.sigma
# only mutate the cells outside our margin
#if self.xy_to_polar(cell.xy)[1] > (self.size * self.margin):
# we use the same coef as the noise coef
#cell.diffusive_mass += val * self.environment.sigma
def log_status(self):
if self.datalog == None:
return
row = []
#row.append(self.iteration)
dm = [cell.diffusive_mass for cell in self.cells if cell]
row.append(sum(dm))
cm = [cell.crystal_mass for cell in self.cells if cell]
row.append(sum(cm))
bm = [cell.boundary_mass for cell in self.cells if cell]
row.append(sum(bm))
acnt = len([cell for cell in self.cells if cell and cell.attached])
row.append(acnt)
bcnt = len([cell for cell in self.cells if cell and cell.boundary])
row.append(bcnt)
d = self.snowflake_radius()
row.append(d)
row.append(self.environment.beta)
row.append(self.environment.theta)
row.append(self.environment.alpha)
row.append(self.environment.kappa)
row.append(self.environment.mu)
row.append(self.environment.upsilon)
#row.append(self.environment.sigma)
#row.append(self.environment.gamma)
self.datalog.append(row)
# log the cells
self.celllog.append((self.iter | if type(state) == dict:
self.update(state)
self.curves = None
self.set_factory_settings()
else:
self.curves = state[0]
self.factory_settings = state[1]
self.update(state[2]) | identifier_body |
|
engine.py | cell for cell in self.cells if cell and cell.boundary])
row.append(bcnt)
d = self.snowflake_radius()
row.append(d)
row.append(self.environment.beta)
row.append(self.environment.theta)
row.append(self.environment.alpha)
row.append(self.environment.kappa)
row.append(self.environment.mu)
row.append(self.environment.upsilon)
#row.append(self.environment.sigma)
#row.append(self.environment.gamma)
self.datalog.append(row)
# log the cells
self.celllog.append((self.iteration, dm, cm))
def write_log(self):
self.write_datalog()
self.write_celllog()
def write_datalog(self):
if self.datalog == None:
return
logfn = "datalog.csv"
msg = "Saving runtime data to %s" % logfn
log(msg)
f = open(logfn, 'w')
txt = ''
txt += str.join(',', self.LogHeader) + '\n'
for row in self.datalog:
txt += str.join(',', map(str, row)) + '\n'
f.write(txt)
def write_celllog(self):
if not self.celllog:
return
logfn = "cell_log_%d.pickle" % self.iteration
f = open(logfn, 'wb')
pickle.dump(self.celllog, f, protocol=-1)
self.celllog = []
def print_status(self):
dm = sum([cell.diffusive_mass for cell in self.cells if cell])
cm = sum([cell.crystal_mass for cell in self.cells if cell])
bm = sum([cell.boundary_mass for cell in self.cells if cell])
acnt = len([cell for cell in self.cells if cell and cell.attached])
bcnt = len([cell for cell in self.cells if cell and cell.boundary])
#msg = "Step #%d, %d attached, %d boundary, %.2f dM, %.2f bM, %.2f cM, tot %.2f M" % (self.iteration, acnt, bcnt, dm, bm, cm, dm + cm + bm)
d = self.snowflake_radius()
msg = "Step #%d/%dp (%.2f%% scl), %d/%d (%.2f%%), %.2f dM, %.2f bM, %.2f cM, tot %.2f M" % (self.iteration, d, (float(d * 2 * X_SCALE_FACTOR) / self.iteration) * 100, acnt, bcnt, (float(bcnt) / acnt) * 100, dm, bm, cm, dm + cm + bm)
log(msg)
def step(self):
self.log_status()
for cell in self.cells:
if cell == None or cell.attached:
continue
cell.step_one()
for cell in self.cells:
if cell == None or cell.attached:
continue
cell.step_two()
for cell in self.cells:
if cell == None or cell.attached:
continue
cell.step_three()
# run curves
self.iteration += 1
self.environment.step(self.iteration)
def translate_xy(self, xy):
(x, y) = xy
x = int(round(x * X_SCALE_FACTOR))
return (x, y)
def polar_to_xy(self, args):
(angle, distance) = args
half = self.size / 2.0
angle = math.radians(angle)
y = int(round(half - (math.sin(angle) * distance)))
x = int(round(half + (math.cos(angle) * distance)))
return (x, y)
def xy_to_polar(self, args):
(x, y) = args
half = self.size / 2.0
x -= half
y += half
angle = math.degrees(math.atan2(y, x))
distance = math.hypot(x, y)
return (angle, distance)
def snowflake_radius(self, angle=135):
# we cast a ray on the 135 degeree axis
radius = 0
half = self.size / 2.0
while radius < half:
radius += 1
xy = self.polar_to_xy((angle, radius))
cell = self.cells[self._cell_index(xy)]
if cell.attached or cell.boundary:
continue
return radius
# uhh
return int(round(half))
def crop_snowflake(self, margin=None):
def scale(val):
return int(round(X_SCALE_FACTOR * val))
if margin == None:
margin = 15
half = self.size / 2
radius = scale(self.snowflake_radius())
distance = min(radius + margin, half)
half_s = scale(half)
distance_s = scale(distance)
box = (half_s - distance, half - distance, half_s + distance, half + distance)
return box
def headroom(self, margin=None):
if self.max_steps and self.iteration >= self.max_steps:
return False
if margin == None:
margin = self.margin
assert margin > 0 and margin <= 1
cutoff = int(round(margin * (self.size / 2.0)))
radius = self.snowflake_radius()
if radius > cutoff:
return False
return True
def grow(self):
while True:
if self.debug:
self.print_status()
self.step()
if self.iteration % 50 == 0:
self.write_celllog()
if not self.debug:
self.print_status()
if not self.headroom():
break
if self.debug:
self.print_status()
def save_image(self, fn, **kw):
import sfgen
r = sfgen.RenderSnowflake(self)
r.save_image(fn, **kw)
class SnowflakeCell(object):
def __init__(self, xy, lattice):
self.xy = xy
self.lattice = lattice
self.env = lattice.environment
self.diffusive_mass = self.env.gamma
self.boundary_mass = 0.0
self.crystal_mass = 0.0
self.attached = False
self.age = 0
self.boundary = 0
self.attached_neighbors = []
self.__neighbors = None
def __getstate__(self):
return (self.xy, self.diffusive_mass, self.boundary_mass, self.crystal_mass, self.attached, self.age)
def __setstate__(self, state):
self.xy = state[0]
self.diffusive_mass = state[1]
self.boundary_mass = state[2]
self.crystal_mass = state[3]
self.attached = state[4]
# 0.2 -> 0.3
try:
self.age = state[5]
except IndexError:
self.age = 0
self.__neighbors = None
self.lattice = None
self.env = None
def reality_check(self):
assert len(self.neighbors)
for neighbor in self.neighbors:
assert self in neighbor.neighbors, "%s not in %s" % (str(self), str(neighbor.neighbors))
def __repr__(self):
return "(%d,%d)" % self.xy
@property
def neighbors(self):
if self.__neighbors == None:
self.__neighbors = self.lattice.get_neighbors(self.xy)
return self.__neighbors
#@property
#def attached_neighbors(self):
# return [cell for cell in self.neighbors if cell.attached]
#@property
#def boundary(self):
# return (not self.attached) and any([cell.attached for cell in self.neighbors])
def update_boundary(self):
self.boundary = (not self.attached) and any([cell.attached for cell in self.neighbors])
def step_one(self):
self.update_boundary()
if self.boundary:
self.attached_neighbors = [cell for cell in self.neighbors if cell.attached]
self._next_dm = self.diffusion_calc()
def step_two(self):
self.diffusive_mass = self._next_dm
self.attachment_flag = self.attached
self.freezing_step()
self.attachment_flag = self.attachment_step()
self.melting_step()
def step_three(self):
if self.boundary and self.attachment_flag:
self.attach()
self.noise_step()
def diffusion_calc(self):
next_dm = self.diffusive_mass
if self.attached:
return next_dm
self.age += 1
for cell in self.neighbors:
if cell.attached:
next_dm += self.diffusive_mass
else:
next_dm += cell.diffusive_mass
return float(next_dm) / (len(self.neighbors) + 1)
def attach(self, offset=0.0):
self.crystal_mass = self.boundary_mass + self.crystal_mass + offset
self.boundary_mass = 0
self.attached = True
def freezing_step(self):
if not self.boundary:
return
self.boundary_mass += (1 - self.env.kappa) * self.diffusive_mass
self.crystal_mass += (self.env.kappa * self.diffusive_mass)
self.diffusive_mass = 0
def | attachment_step | identifier_name |
|
rcparser.py |
class DialogDef:
name = ""
id = 0
style = 0
styleEx = None
caption = ""
font = "MS Sans Serif"
fontSize = 8
x = 0
y = 0
w = 0
h = 0
template = None
def __init__(self, n, i):
self.name = n
self.id = i
self.styles = []
self.stylesEx = []
self.controls = []
def createDialogTemplate(self):
t = None
self.template = [[self.caption, (self.x,self.y,self.w,self.h), self.style, self.styleEx, (self.fontSize, self.font)]]
for control in self.controls:
self.template.append(control.createDialogTemplate())
return self.template
class ControlDef:
id = ""
controlType = ""
subType = ""
idNum = 0
style = defaultControlStyle
label = ""
x = 0
y = 0
w = 0
h = 0
def __init__(self):
self.styles = []
def toString(self):
s = "<Control id:"+self.id+" controlType:"+self.controlType+" subType:"+self.subType\
+" idNum:"+str(self.idNum)+" style:"+str(self.style)+" styles:"+str(self.styles)+" label:"+self.label\
+" x:"+str(self.x)+" y:"+str(self.y)+" w:"+str(self.w)+" h:"+str(self.h)+">"
return s
def createDialogTemplate(self):
ct = self.controlType
if "CONTROL"==ct:
ct = self.subType
if ct in _addDefaults:
self.style |= _addDefaults[ct]
if ct in _controlMap:
ct = _controlMap[ct]
t = [ct, self.label, self.idNum, (self.x, self.y, self.w, self.h), self.style]
return t
class gt_str(str):
"""Change a string to a gettext version of itself."""
def __repr__(self):
if len(self) > 0:
return "_(" + super(gt_str, self).__repr__() + ")"
else:
return super(gt_str, self).__repr__()
class RCParser:
next_id = 1001
dialogs = {}
_dialogs = {}
debugEnabled = False
token = ""
def __init__(self):
self.ids = {"IDOK":1, "IDCANCEL":2, "IDC_STATIC": -1}
self.names = {1:"IDOK", 2:"IDCANCEL", -1:"IDC_STATIC"}
self.bitmaps = {}
self.gettexted = False
def debug(self, *args):
if self.debugEnabled:
print(args)
def | (self):
self.token = self.lex.get_token()
self.debug("getToken returns:", self.token)
if self.token=="":
self.token = None
return self.token
def getCommaToken(self):
tok = self.getToken()
assert tok == ",", "Token '%s' should be a comma!" % tok
def loadDialogs(self, rcFileName):
"""
RCParser.loadDialogs(rcFileName) -> None
Load the dialog information into the parser. Dialog Definations can then be accessed
using the "dialogs" dictionary member (name->DialogDef). The "ids" member contains the dictionary of id->name.
The "names" member contains the dictionary of name->id
"""
hFileName = rcFileName[:-2]+"h"
if not os.path.exists(hFileName):
hFileName = os.path.join(os.path.dirname(__file__),
os.path.basename(hFileName))
try:
h = open(hFileName, "rU")
self.parseH(h)
h.close()
except IOError:
print("No .h file. ignoring.")
f = open(rcFileName)
self.open(f)
self.getToken()
while self.token!=None:
self.parse()
self.getToken()
f.close()
def open(self, file):
self.lex = shlex.shlex(file)
self.lex.commenters = "//#"
def parseH(self, file):
lex = shlex.shlex(file)
lex.commenters = "//"
token = " "
while token is not None:
token = lex.get_token()
if token == "" or token is None:
token = None
else:
if token=='define':
n = lex.get_token()
i = int(lex.get_token())
self.ids[n] = i
if i in self.names:
if not n.startswith("_APS_"):
print("Duplicate id",i,"for",n,"is", self.names[i])
else:
self.names[i] = n
if self.next_id<=i:
self.next_id = i+1
def parse(self):
deep = 0
if self.token == None:
more == None
elif "BEGIN" == self.token:
deep = 1
while deep!=0:
self.getToken()
if "BEGIN" == self.token:
deep += 1
elif "END" == self.token:
deep -= 1
elif "IDD_" == self.token[:4]:
possibleDlgName = self.token
self.getToken()
if "DIALOG" == self.token or "DIALOGEX" == self.token:
self.dialog(possibleDlgName)
elif "IDB_" == self.token[:4]:
possibleBitmap = self.token
self.getToken()
if "BITMAP" == self.token:
self.getToken()
if self.token=="MOVEABLE":
self.getToken() # PURE
self.getToken() # bmpname
bmf = self.token[1:-1] # quotes
self.bitmaps[possibleBitmap] = bmf
print("BITMAP", possibleBitmap, bmf)
def addId(self, id_name):
if id_name in self.ids:
id = self.ids[id_name]
else:
id = self.next_id
self.next_id += 1
self.ids[id_name] = id
self.names[id] = id_name
return id
def lang(self):
while self.token[0:4]=="LANG" or self.token[0:7]=="SUBLANG" or self.token==',':
self.getToken();
def dialog(self, name):
dlg = DialogDef(name,self.addId(name))
assert len(dlg.controls)==0
self._dialogs[name] = dlg
extras = []
self.getToken()
while not self.token.isdigit():
self.debug("extra", self.token)
extras.append(self.token)
self.getToken()
dlg.x = int(self.token)
self.getCommaToken()
self.getToken() # number
dlg.y = int(self.token)
self.getCommaToken()
self.getToken() # number
dlg.w = int(self.token)
self.getCommaToken()
self.getToken() # number
dlg.h = int(self.token)
self.getToken()
while not (self.token==None or self.token=="" or self.token=="END"):
if self.token=="STYLE":
self.dialogStyle(dlg)
elif self.token=="EXSTYLE":
self.dialogExStyle(dlg)
elif self.token=="CAPTION":
self.dialogCaption(dlg)
elif self.token=="FONT":
self.dialogFont(dlg)
elif self.token=="BEGIN":
self.controls(dlg)
else:
break
self.dialogs[name] = dlg.createDialogTemplate()
def dialogStyle(self, dlg):
dlg.style, dlg.styles = self.styles( [], win32con.WS_VISIBLE | win32con.DS_SETFONT)
def dialogExStyle(self, dlg):
self.getToken()
dlg.styleEx, dlg.stylesEx = self.styles( [], 0)
def styles(self, defaults, defaultStyle):
list = defaults
style = defaultStyle
if "STYLE"==self.token:
self.getToken()
i = 0
Not = False
while ((i%2==1 and ("|"==self.token or "NOT"==self.token)) or (i%2==0)) and not self.token==None:
Not = False;
if "NOT"==self.token:
Not = True
self.getToken()
i += 1
if self.token!="|":
if self.token in win32con.__dict__:
value = getattr(win32con,self.token)
else:
if self.token in commctrl.__dict__:
value = getattr(commctrl,self.token)
else:
value = 0
if Not:
list.append("NOT "+self.token)
self.debug("styles add Not",self.token, value)
style &= ~value
else:
list.append(self.token)
self.debug("styles add", self.token, value)
style |= value
self.getToken()
self.debug("style is ",style)
return style, list
def dialogCaption(self, dlg):
if "CAPTION"==self.token:
self.getToken()
self.token = self.token[1:-1]
self.debug("Caption is:",self.token)
if self.gettexted:
dlg.caption = gt_str(self.token)
else:
dlg.caption = self.token
self.getToken()
def dialogFont | getToken | identifier_name |
rcparser.py |
class DialogDef:
name = ""
id = 0
style = 0
styleEx = None
caption = ""
font = "MS Sans Serif"
fontSize = 8
x = 0
y = 0
w = 0
h = 0
template = None
def __init__(self, n, i):
self.name = n
self.id = i
self.styles = []
self.stylesEx = []
self.controls = []
def createDialogTemplate(self):
t = None
self.template = [[self.caption, (self.x,self.y,self.w,self.h), self.style, self.styleEx, (self.fontSize, self.font)]]
for control in self.controls:
self.template.append(control.createDialogTemplate())
return self.template
class ControlDef:
id = ""
controlType = ""
subType = ""
idNum = 0
style = defaultControlStyle
label = ""
x = 0
y = 0
w = 0
h = 0
def __init__(self):
self.styles = []
def toString(self):
s = "<Control id:"+self.id+" controlType:"+self.controlType+" subType:"+self.subType\
+" idNum:"+str(self.idNum)+" style:"+str(self.style)+" styles:"+str(self.styles)+" label:"+self.label\
+" x:"+str(self.x)+" y:"+str(self.y)+" w:"+str(self.w)+" h:"+str(self.h)+">"
return s
def createDialogTemplate(self):
ct = self.controlType
if "CONTROL"==ct:
ct = self.subType
if ct in _addDefaults:
self.style |= _addDefaults[ct]
if ct in _controlMap:
ct = _controlMap[ct]
t = [ct, self.label, self.idNum, (self.x, self.y, self.w, self.h), self.style]
return t
class gt_str(str):
"""Change a string to a gettext version of itself."""
def __repr__(self):
if len(self) > 0:
return "_(" + super(gt_str, self).__repr__() + ")"
else:
return super(gt_str, self).__repr__()
class RCParser:
next_id = 1001
dialogs = {}
_dialogs = {}
debugEnabled = False
token = ""
def __init__(self):
self.ids = {"IDOK":1, "IDCANCEL":2, "IDC_STATIC": -1}
self.names = {1:"IDOK", 2:"IDCANCEL", -1:"IDC_STATIC"}
self.bitmaps = {}
self.gettexted = False
def debug(self, *args):
if self.debugEnabled:
print(args)
def getToken(self):
self.token = self.lex.get_token()
self.debug("getToken returns:", self.token)
if self.token=="":
self.token = None
return self.token
def getCommaToken(self):
tok = self.getToken()
assert tok == ",", "Token '%s' should be a comma!" % tok
def loadDialogs(self, rcFileName):
"""
RCParser.loadDialogs(rcFileName) -> None
Load the dialog information into the parser. Dialog Definations can then be accessed
using the "dialogs" dictionary member (name->DialogDef). The "ids" member contains the dictionary of id->name.
The "names" member contains the dictionary of name->id
"""
hFileName = rcFileName[:-2]+"h"
if not os.path.exists(hFileName):
hFileName = os.path.join(os.path.dirname(__file__),
os.path.basename(hFileName))
try:
h = open(hFileName, "rU")
self.parseH(h)
h.close()
except IOError:
print("No .h file. ignoring.")
f = open(rcFileName)
self.open(f)
self.getToken()
while self.token!=None:
self.parse()
self.getToken()
f.close()
def open(self, file):
self.lex = shlex.shlex(file)
self.lex.commenters = "//#"
def parseH(self, file):
lex = shlex.shlex(file)
lex.commenters = "//"
token = " "
while token is not None:
token = lex.get_token()
if token == "" or token is None:
token = None
else:
if token=='define':
n = lex.get_token()
i = int(lex.get_token())
self.ids[n] = i
if i in self.names:
if not n.startswith("_APS_"):
print("Duplicate id",i,"for",n,"is", self.names[i])
else:
self.names[i] = n
if self.next_id<=i:
self.next_id = i+1
def parse(self):
deep = 0
if self.token == None:
more == None
elif "BEGIN" == self.token:
deep = 1
while deep!=0:
self.getToken()
if "BEGIN" == self.token:
deep += 1
elif "END" == self.token:
deep -= 1
elif "IDD_" == self.token[:4]:
possibleDlgName = self.token
self.getToken()
if "DIALOG" == self.token or "DIALOGEX" == self.token:
self.dialog(possibleDlgName)
elif "IDB_" == self.token[:4]:
possibleBitmap = self.token
self.getToken()
if "BITMAP" == self.token:
self.getToken()
if self.token=="MOVEABLE":
self.getToken() # PURE
self.getToken() # bmpname
bmf = self.token[1:-1] # quotes
self.bitmaps[possibleBitmap] = bmf
print("BITMAP", possibleBitmap, bmf)
def addId(self, id_name):
if id_name in self.ids:
id = self.ids[id_name]
else:
id = self.next_id
self.next_id += 1
self.ids[id_name] = id
self.names[id] = id_name
return id
def lang(self):
while self.token[0:4]=="LANG" or self.token[0:7]=="SUBLANG" or self.token==',':
self.getToken();
def dialog(self, name):
dlg = DialogDef(name,self.addId(name))
assert len(dlg.controls)==0
self._dialogs[name] = dlg
extras = []
self.getToken()
while not self.token.isdigit():
self.debug("extra", self.token)
extras.append(self.token)
self.getToken()
dlg.x = int(self.token)
self.getCommaToken()
self.getToken() # number
dlg.y = int(self.token)
self.getCommaToken()
self.getToken() # number
dlg.w = int(self.token)
self.getCommaToken()
self.getToken() # number
dlg.h = int(self.token)
self.getToken()
while not (self.token==None or self.token=="" or self.token=="END"):
|
self.dialogs[name] = dlg.createDialogTemplate()
def dialogStyle(self, dlg):
dlg.style, dlg.styles = self.styles( [], win32con.WS_VISIBLE | win32con.DS_SETFONT)
def dialogExStyle(self, dlg):
self.getToken()
dlg.styleEx, dlg.stylesEx = self.styles( [], 0)
def styles(self, defaults, defaultStyle):
list = defaults
style = defaultStyle
if "STYLE"==self.token:
self.getToken()
i = 0
Not = False
while ((i%2==1 and ("|"==self.token or "NOT"==self.token)) or (i%2==0)) and not self.token==None:
Not = False;
if "NOT"==self.token:
Not = True
self.getToken()
i += 1
if self.token!="|":
if self.token in win32con.__dict__:
value = getattr(win32con,self.token)
else:
if self.token in commctrl.__dict__:
value = getattr(commctrl,self.token)
else:
value = 0
if Not:
list.append("NOT "+self.token)
self.debug("styles add Not",self.token, value)
style &= ~value
else:
list.append(self.token)
self.debug("styles add", self.token, value)
style |= value
self.getToken()
self.debug("style is ",style)
return style, list
def dialogCaption(self, dlg):
if "CAPTION"==self.token:
self.getToken()
self.token = self.token[1:-1]
self.debug("Caption is:",self.token)
if self.gettexted:
dlg.caption = gt_str(self.token)
else:
dlg.caption = self.token
self.getToken()
def dialogFont(self | if self.token=="STYLE":
self.dialogStyle(dlg)
elif self.token=="EXSTYLE":
self.dialogExStyle(dlg)
elif self.token=="CAPTION":
self.dialogCaption(dlg)
elif self.token=="FONT":
self.dialogFont(dlg)
elif self.token=="BEGIN":
self.controls(dlg)
else:
break | conditional_block |
rcparser.py |
class DialogDef:
name = ""
id = 0
style = 0
styleEx = None
caption = ""
font = "MS Sans Serif"
fontSize = 8
x = 0
y = 0
w = 0
h = 0
template = None
def __init__(self, n, i):
self.name = n
self.id = i
self.styles = []
self.stylesEx = []
self.controls = []
def createDialogTemplate(self):
t = None
self.template = [[self.caption, (self.x,self.y,self.w,self.h), self.style, self.styleEx, (self.fontSize, self.font)]]
for control in self.controls:
self.template.append(control.createDialogTemplate())
return self.template
class ControlDef:
id = ""
controlType = ""
subType = ""
idNum = 0
style = defaultControlStyle
label = ""
x = 0
y = 0
w = 0
h = 0
def __init__(self):
self.styles = []
def toString(self):
s = "<Control id:"+self.id+" controlType:"+self.controlType+" subType:"+self.subType\
+" idNum:"+str(self.idNum)+" style:"+str(self.style)+" styles:"+str(self.styles)+" label:"+self.label\
+" x:"+str(self.x)+" y:"+str(self.y)+" w:"+str(self.w)+" h:"+str(self.h)+">"
return s
def createDialogTemplate(self):
ct = self.controlType
if "CONTROL"==ct:
ct = self.subType
if ct in _addDefaults:
self.style |= _addDefaults[ct]
if ct in _controlMap:
ct = _controlMap[ct]
t = [ct, self.label, self.idNum, (self.x, self.y, self.w, self.h), self.style]
return t
class gt_str(str):
"""Change a string to a gettext version of itself."""
def __repr__(self):
if len(self) > 0:
return "_(" + super(gt_str, self).__repr__() + ")"
else:
return super(gt_str, self).__repr__()
class RCParser:
next_id = 1001
dialogs = {}
_dialogs = {}
debugEnabled = False
token = ""
def __init__(self):
self.ids = {"IDOK":1, "IDCANCEL":2, "IDC_STATIC": -1}
self.names = {1:"IDOK", 2:"IDCANCEL", -1:"IDC_STATIC"}
self.bitmaps = {}
self.gettexted = False
def debug(self, *args):
if self.debugEnabled:
print(args)
def getToken(self):
self.token = self.lex.get_token()
self.debug("getToken returns:", self.token)
if self.token=="":
self.token = None
return self.token
def getCommaToken(self):
tok = self.getToken()
assert tok == ",", "Token '%s' should be a comma!" % tok
def loadDialogs(self, rcFileName):
"""
RCParser.loadDialogs(rcFileName) -> None
Load the dialog information into the parser. Dialog Definations can then be accessed
using the "dialogs" dictionary member (name->DialogDef). The "ids" member contains the dictionary of id->name.
The "names" member contains the dictionary of name->id
"""
hFileName = rcFileName[:-2]+"h"
if not os.path.exists(hFileName):
hFileName = os.path.join(os.path.dirname(__file__),
os.path.basename(hFileName))
try:
h = open(hFileName, "rU")
self.parseH(h)
h.close()
except IOError:
print("No .h file. ignoring.")
f = open(rcFileName)
self.open(f)
self.getToken()
while self.token!=None:
self.parse()
self.getToken()
f.close()
def open(self, file):
self.lex = shlex.shlex(file)
self.lex.commenters = "//#"
def parseH(self, file):
lex = shlex.shlex(file)
lex.commenters = "//"
token = " "
while token is not None:
token = lex.get_token()
if token == "" or token is None:
token = None
else:
if token=='define':
n = lex.get_token()
i = int(lex.get_token())
self.ids[n] = i
if i in self.names:
if not n.startswith("_APS_"):
print("Duplicate id",i,"for",n,"is", self.names[i])
else:
self.names[i] = n
if self.next_id<=i:
self.next_id = i+1
def parse(self):
deep = 0
if self.token == None:
more == None
elif "BEGIN" == self.token:
deep = 1
while deep!=0:
self.getToken()
if "BEGIN" == self.token:
deep += 1
elif "END" == self.token:
deep -= 1
elif "IDD_" == self.token[:4]:
possibleDlgName = self.token
self.getToken()
if "DIALOG" == self.token or "DIALOGEX" == self.token:
self.dialog(possibleDlgName)
elif "IDB_" == self.token[:4]:
possibleBitmap = self.token
self.getToken()
if "BITMAP" == self.token:
self.getToken()
if self.token=="MOVEABLE":
self.getToken() # PURE
self.getToken() # bmpname
bmf = self.token[1:-1] # quotes
self.bitmaps[possibleBitmap] = bmf
print("BITMAP", possibleBitmap, bmf)
def addId(self, id_name):
if id_name in self.ids:
id = self.ids[id_name]
else:
id = self.next_id
self.next_id += 1
self.ids[id_name] = id
self.names[id] = id_name
return id
def lang(self):
while self.token[0:4]=="LANG" or self.token[0:7]=="SUBLANG" or self.token==',':
self.getToken();
def dialog(self, name):
dlg = DialogDef(name,self.addId(name))
assert len(dlg.controls)==0
self._dialogs[name] = dlg
extras = []
self.getToken()
while not self.token.isdigit():
self.debug("extra", self.token)
extras.append(self.token)
self.getToken()
dlg.x = int(self.token)
self.getCommaToken()
self.getToken() # number
dlg.y = int(self.token)
self.getCommaToken()
self.getToken() # number
dlg.w = int(self.token)
self.getCommaToken()
self.getToken() # number
dlg.h = int(self.token)
self.getToken()
while not (self.token==None or self.token=="" or self.token=="END"):
if self.token=="STYLE":
self.dialogStyle(dlg)
elif self.token=="EXSTYLE":
self.dialogExStyle(dlg)
elif self.token=="CAPTION":
self.dialogCaption(dlg)
elif self.token=="FONT":
self.dialogFont(dlg)
elif self.token=="BEGIN":
self.controls(dlg)
else:
break
self.dialogs[name] = dlg.createDialogTemplate()
def dialogStyle(self, dlg):
dlg.style, dlg.styles = self.styles( [], win32con.WS_VISIBLE | win32con.DS_SETFONT)
def dialogExStyle(self, dlg):
self.getToken()
dlg.styleEx, dlg.stylesEx = self.styles( [], 0)
def styles(self, defaults, defaultStyle):
list = defaults
style = defaultStyle
if "STYLE"==self.token:
self.getToken()
i = 0
Not = False
while ((i%2==1 and ("|"==self.token or "NOT"==self.token)) or (i%2==0)) and not self.token==None:
Not = False;
if "NOT"==self.token:
Not = True
self.getToken()
i += 1
if self.token!="|":
if self.token in win32con.__dict__:
value = getattr(win32con,self.token)
else:
if self.token in commctrl.__dict__:
value = getattr(commctrl,self.token)
else:
value = 0
if Not:
list.append("NOT "+self.token)
self.debug("styles add Not",self.token, value)
style &= ~value
else:
list.append(self.token)
self.debug("styles add", self.token, value)
style |= value
self.getToken()
self.debug("style is ",style)
return style, list
def dialogCaption(self, dlg):
|
def dialogFont | if "CAPTION"==self.token:
self.getToken()
self.token = self.token[1:-1]
self.debug("Caption is:",self.token)
if self.gettexted:
dlg.caption = gt_str(self.token)
else:
dlg.caption = self.token
self.getToken() | identifier_body |
rcparser.py | _VISIBLE
class DialogDef:
name = ""
id = 0
style = 0
styleEx = None
caption = ""
font = "MS Sans Serif"
fontSize = 8
x = 0
y = 0
w = 0
h = 0
template = None
def __init__(self, n, i):
self.name = n
self.id = i
self.styles = []
self.stylesEx = []
self.controls = []
def createDialogTemplate(self):
t = None
self.template = [[self.caption, (self.x,self.y,self.w,self.h), self.style, self.styleEx, (self.fontSize, self.font)]]
for control in self.controls:
self.template.append(control.createDialogTemplate())
return self.template
class ControlDef:
id = ""
controlType = ""
subType = "" | label = ""
x = 0
y = 0
w = 0
h = 0
def __init__(self):
self.styles = []
def toString(self):
s = "<Control id:"+self.id+" controlType:"+self.controlType+" subType:"+self.subType\
+" idNum:"+str(self.idNum)+" style:"+str(self.style)+" styles:"+str(self.styles)+" label:"+self.label\
+" x:"+str(self.x)+" y:"+str(self.y)+" w:"+str(self.w)+" h:"+str(self.h)+">"
return s
def createDialogTemplate(self):
ct = self.controlType
if "CONTROL"==ct:
ct = self.subType
if ct in _addDefaults:
self.style |= _addDefaults[ct]
if ct in _controlMap:
ct = _controlMap[ct]
t = [ct, self.label, self.idNum, (self.x, self.y, self.w, self.h), self.style]
return t
class gt_str(str):
"""Change a string to a gettext version of itself."""
def __repr__(self):
if len(self) > 0:
return "_(" + super(gt_str, self).__repr__() + ")"
else:
return super(gt_str, self).__repr__()
class RCParser:
next_id = 1001
dialogs = {}
_dialogs = {}
debugEnabled = False
token = ""
def __init__(self):
self.ids = {"IDOK":1, "IDCANCEL":2, "IDC_STATIC": -1}
self.names = {1:"IDOK", 2:"IDCANCEL", -1:"IDC_STATIC"}
self.bitmaps = {}
self.gettexted = False
def debug(self, *args):
if self.debugEnabled:
print(args)
def getToken(self):
self.token = self.lex.get_token()
self.debug("getToken returns:", self.token)
if self.token=="":
self.token = None
return self.token
def getCommaToken(self):
tok = self.getToken()
assert tok == ",", "Token '%s' should be a comma!" % tok
def loadDialogs(self, rcFileName):
"""
RCParser.loadDialogs(rcFileName) -> None
Load the dialog information into the parser. Dialog Definations can then be accessed
using the "dialogs" dictionary member (name->DialogDef). The "ids" member contains the dictionary of id->name.
The "names" member contains the dictionary of name->id
"""
hFileName = rcFileName[:-2]+"h"
if not os.path.exists(hFileName):
hFileName = os.path.join(os.path.dirname(__file__),
os.path.basename(hFileName))
try:
h = open(hFileName, "rU")
self.parseH(h)
h.close()
except IOError:
print("No .h file. ignoring.")
f = open(rcFileName)
self.open(f)
self.getToken()
while self.token!=None:
self.parse()
self.getToken()
f.close()
def open(self, file):
self.lex = shlex.shlex(file)
self.lex.commenters = "//#"
def parseH(self, file):
lex = shlex.shlex(file)
lex.commenters = "//"
token = " "
while token is not None:
token = lex.get_token()
if token == "" or token is None:
token = None
else:
if token=='define':
n = lex.get_token()
i = int(lex.get_token())
self.ids[n] = i
if i in self.names:
if not n.startswith("_APS_"):
print("Duplicate id",i,"for",n,"is", self.names[i])
else:
self.names[i] = n
if self.next_id<=i:
self.next_id = i+1
def parse(self):
deep = 0
if self.token == None:
more == None
elif "BEGIN" == self.token:
deep = 1
while deep!=0:
self.getToken()
if "BEGIN" == self.token:
deep += 1
elif "END" == self.token:
deep -= 1
elif "IDD_" == self.token[:4]:
possibleDlgName = self.token
self.getToken()
if "DIALOG" == self.token or "DIALOGEX" == self.token:
self.dialog(possibleDlgName)
elif "IDB_" == self.token[:4]:
possibleBitmap = self.token
self.getToken()
if "BITMAP" == self.token:
self.getToken()
if self.token=="MOVEABLE":
self.getToken() # PURE
self.getToken() # bmpname
bmf = self.token[1:-1] # quotes
self.bitmaps[possibleBitmap] = bmf
print("BITMAP", possibleBitmap, bmf)
def addId(self, id_name):
if id_name in self.ids:
id = self.ids[id_name]
else:
id = self.next_id
self.next_id += 1
self.ids[id_name] = id
self.names[id] = id_name
return id
def lang(self):
while self.token[0:4]=="LANG" or self.token[0:7]=="SUBLANG" or self.token==',':
self.getToken();
def dialog(self, name):
dlg = DialogDef(name,self.addId(name))
assert len(dlg.controls)==0
self._dialogs[name] = dlg
extras = []
self.getToken()
while not self.token.isdigit():
self.debug("extra", self.token)
extras.append(self.token)
self.getToken()
dlg.x = int(self.token)
self.getCommaToken()
self.getToken() # number
dlg.y = int(self.token)
self.getCommaToken()
self.getToken() # number
dlg.w = int(self.token)
self.getCommaToken()
self.getToken() # number
dlg.h = int(self.token)
self.getToken()
while not (self.token==None or self.token=="" or self.token=="END"):
if self.token=="STYLE":
self.dialogStyle(dlg)
elif self.token=="EXSTYLE":
self.dialogExStyle(dlg)
elif self.token=="CAPTION":
self.dialogCaption(dlg)
elif self.token=="FONT":
self.dialogFont(dlg)
elif self.token=="BEGIN":
self.controls(dlg)
else:
break
self.dialogs[name] = dlg.createDialogTemplate()
def dialogStyle(self, dlg):
dlg.style, dlg.styles = self.styles( [], win32con.WS_VISIBLE | win32con.DS_SETFONT)
def dialogExStyle(self, dlg):
self.getToken()
dlg.styleEx, dlg.stylesEx = self.styles( [], 0)
def styles(self, defaults, defaultStyle):
list = defaults
style = defaultStyle
if "STYLE"==self.token:
self.getToken()
i = 0
Not = False
while ((i%2==1 and ("|"==self.token or "NOT"==self.token)) or (i%2==0)) and not self.token==None:
Not = False;
if "NOT"==self.token:
Not = True
self.getToken()
i += 1
if self.token!="|":
if self.token in win32con.__dict__:
value = getattr(win32con,self.token)
else:
if self.token in commctrl.__dict__:
value = getattr(commctrl,self.token)
else:
value = 0
if Not:
list.append("NOT "+self.token)
self.debug("styles add Not",self.token, value)
style &= ~value
else:
list.append(self.token)
self.debug("styles add", self.token, value)
style |= value
self.getToken()
self.debug("style is ",style)
return style, list
def dialogCaption(self, dlg):
if "CAPTION"==self.token:
self.getToken()
self.token = self.token[1:-1]
self.debug("Caption is:",self.token)
if self.gettexted:
dlg.caption = gt_str(self.token)
else:
dlg.caption = self.token
self.getToken()
def dialogFont(self | idNum = 0
style = defaultControlStyle | random_line_split |
ann_interface.py | ]
if i < d_col: #data(speak)の可視化
# 音声Dataがrmsの場合
# set_data:範囲は 0-1
set_data = data[j][i] #self.speakDecibelNorm(data[j][i])
#ON/OFF むちゃくちゃすぎる
#set_data = 1 if set_data > 0.3 else 0
#self.edited_speaks[j][i] = set_data
#一時的なOffset
#set_data = data_proc2.speakNorm(set_data, upper=0.75, lower=0.25)
#self.edited_speaks[j][i] = set_data
#color_dataの範囲を0-255に変更
#color_data=int(set_data*255)
color_data = set_data
color_data = 1 if color_data > 0.5 else 0
color_data=int(color_data*255)
#print color_data
#print "at",color_data
color = [255-color_data]*3
elif i >= d_col and i < t_col-add_flag: # annotationの可視化
#print i
"""
set_data = anno[j][i-d_col] #anno(1000, 2)
if set_data == 0:
color = [255, 255, 255]
else:
color = [0, 0, 0]
"""
set_data = anno[j][i-d_col] #anno(1000, 2)
set_data = 0 if set_data < 0 else set_data
set_data = 1 if set_data > 1 else set_data
color_data=int(set_data*255)
color = [255-color_data]*3
else: #flag
set_data = self.edited_flags[j][0]
if set_data == 0:
color = [255, 255, 255]
else:
color = [0, 0, 0]
self.table.setItem(i, j, QtGui.QTableWidgetItem())
self.table.item(i, j).setBackground(QtGui.QColor(color[0],color[1],color[2]))
self.table.item(i, j).setToolTip(str(set_data))
hor = False
self.table.setVisible(False)
self.table.resizeRowsToContents()
self.table.resizeColumnsToContents()
self.table.setVisible(True)
# TableがClickされたとき
def clickUpdateTable(self, cItem):
self.tip = float(cItem.toolTip())
self.r = cItem.row()
self.c = cItem.column()
print "r:",self.r,", c:",self.c, ". tip:",self.tip
set_data = 0 if self.tip == 1 else 1
speak_dim = self.edited_speaks.shape[1]
anno_dim = self.edited_annos.shape[1]
if self.r < speak_dim:
self.edited_speaks[self.c][self.r] = set_data
elif self.r >= speak_dim and self.r < speak_dim+anno_dim:
self.edited_annos[self.c][self.r-speak_dim] = set_data
else:
self.edited_flags[self.c][0] = set_data
#indexes = self.table.selectedIndexes()
#print indexes
color = [[255, 255, 255], [0, 0, 0]]
self.table.setItem(self.r, self.c, QtGui.QTableWidgetItem())
self.table.item(self.r, self.c).setBackground(QtGui.QColor(color[set_data][0],color[set_data][1],color[set_data][2]))
self.table.item(self.r, self.c).setToolTip(str(set_data))
#jointsの可視化
# self.vizJoint(self.c)
def checkRangeData(self):
set_row = int(self.checkUser.text()) #0~4 User1=2, User2=3
start = int(self.checkSt.text())
end = int(self.checkEd.text())
color = [[255, 255, 255], [0, 0, 0]]
table_offset = self.edited_speaks.shape[1]
#変更するUserの指定
user = set_row - table_offset
#print user, table_offset
if user < 0:
print "Not Change!"
return
print "Change [User:", user, ", Start:",start,", end:",end,"]"
for i in range(start, end):
get_data = self.edited_annos[i][user]
set_data = 0 if get_data == 1 else 1
self.table.setItem(set_row, i, QtGui.QTableWidgetItem())
self.table.item(set_row, i).setBackground(QtGui.QColor(color[set_data][0],color[set_data][1],color[set_data][2]))
self.table.item(set_row, i).setToolTip(str(set_data))
self.edited_annos[i][user] = set_data
def sliderChange(self, timeline):
if len(self.input_joints)==0:
print "now no data:", timeline
return
#self.table.selectColum(timeline)
#self.table.verticalScrollBar().setValue(timeline)
self.table.setCurrentCell(self.r, timeline)
self.pubViz(timeline)
"""
def activatedUpdateTable(self, cItem):
row = cItem.row()
col = cItem.column()
print "row:", row,", col:", col#, ". tip:",self.tip
"""
# 生Dataの入力
def inputData(self):
filesize = int(self.frmSizeBox.text())
print "Input raw data:", filesize
self.fname = [str(self.txtSepFile.text())]
input_data = data_proc2.load_persons(self.fname, annobool=False, datalen=filesize)
#print input_data["times"][0]
# もし手をくわえるなら
#input_data[2] = data_proc.proc_anno(input_data[0], input_data[2], use_vote=True, use_speak=False)
self.loadInputData(input_data)
# 加工済のDataの入力
def inputProcedData(self):
filesize = int(self.frmSizeBox.text())
print "Input proced data", filesize
self.fname = [str(self.txtSepFile.text())]
input_data = data_proc2.load_proced_data_flag(self.fname, datalen=filesize)
#print input_data["times"]
self.loadInputData(input_data)
def loadInputData(self, input_data):
self.input_joints = input_data["joints"]
self.input_speaks = input_data["speaks"]
datalen=self.input_speaks.shape[0]
if input_data.has_key("annos"):
self.input_annos = input_data["annos"]
#とりあえずuser1だけ編集(0124_personsはまた別)
#self.input_annos[:,:1] = np.zeros((self.input_annos.shape[0],1))
"""
user_joint = self.input_joints[:,:36]
user_anno = self.input_annos[:,:1]
calc_user_anno = data_proc2.proc_anno(user_joint, user_anno, use_vote=True, use_anno=False, threshold=-0.2)
self.input_annos[:,:1] = calc_user_anno
print "now_persons mode"
user_joint = self.input_joints[:,36:]
user_anno = self.input_annos[:,1:]
calc_user_anno = data_proc2.proc_anno(user_joint, user_anno, use_vote=True, use_anno=False, threshold=-0.2)
self.input_annos[:,1:] = calc_user_anno
"""
#print "now model mode"
self.input_annos[:,:] = np.round(self.input_annos[:,:])
else:
self.input_annos = np.zeros((datalen, self.anno_dim))#(1000,2)
if input_data.has_key("flags"):
print "load flags"
self.input_flags = input_data["flags"]
else:
print "create flags"
self.input_flags = np.zeros((datalen, 1))
if input_data.has_key("times"):
print "load times"
self.input_times = input_data["times"]
else:
print "create times"
self.input_times = np.zeros((datalen, 1))
"""
diff_times = []
for t in range(len(self.input_times)):
if t == 0:
diff_times.append(0)
else:
diff_times.append(self.input_times[t]-self.input_times[t-1])
fps = np.round(1/float(np.mean(diff_times)))
plt.plot(self.input_annos[:,0], label="person", color="red")
plt.plot(self.input_annos[:,1], label="robot", color="green")
plt.xlabel("Frame (fps:" | +str(fps)+")")
plt.title("speaker/listener result")
plt.xl | conditional_block |
|
ann_interface.py | .txtSepFile.setText('')
# reset
def resetAnno(self):
print "reset Data before:",self.edited_joints.shape, self.edited_annos.shape
self.edited_joints = copy.deepcopy(self.input_joints)
self.edited_speaks = copy.deepcopy(self.input_speaks)
#datalen = self.edited_speaks.shape[1]
self.edited_annos = copy.deepcopy(self.input_annos) #np.zeros((datalen, self.anno_dim)) #(1000, 2)
self.edited_flags = copy.deepcopy(self.input_flags)
self.sld.setMaximum(self.edited_speaks.shape[0]-1) #
self.updateTable(self.edited_speaks, self.edited_annos, self.edited_times)
print "reset Data after:",self.edited_joints.shape, self.edited_speaks.shape, self.edited_annos.shape
def speakNorm(self, data, upper=300, lower=80):
# 上限
c_data=data if data <upper else upper
# 下限
c_data=c_data if c_data>lower else lower
# 最大値で割る
return (c_data-lower)/float(upper-lower)
def convDecibel(self, data):
return 20*math.log10(data)
#decibel
def speakDecibelNorm(self, data):
decibel = self.convDecibel(data)
return self.speakNorm(decibel, upper=70, lower=30)
def updateTable(self, data, anno, times):
#plt.plot(data)
#plt.show()
th = 0#float(self.ThesholdBox.text())
if(len(data)==0):
print "No Data! Push exec button..."
d_row, d_col = data.shape #(1000, 2)
add_ann = anno.shape[1] #annotationの個数
#print add_ann
add_flag = 1
#t_row:時間軸, t_col:次元数, add_ann:加えるAnnotation
t_col = d_col+add_ann+add_flag
t_row = d_row
#print "t, d, a",t_col,d_col,add_ann
self.table.clear()
font = QtGui.QFont()
font.setFamily(u"DejaVu Sans")
font.setPointSize(5)
self.table.horizontalHeader().setFont(font)
self.table.verticalHeader().setFont(font)
self.table.setColumnCount(t_row) # 次元数(tableはcolだけど値はrow!)+flag
self.table.setRowCount(t_col) # 時間軸(tableはrowだけど値はcol!)
# print "t:", t_row, t_col
#self.table.setRowCount(data.shape[0])
#self.table.setColumnCount(ann_num)
# 軸の値をSet
#for i in range(len(times)):
# jItem = QtGui.QTableWidgetItem(str(i))
# self.table.setHorizontalHeaderItem(i, jItem)
hor = True
for i in range(t_col):
iItem = QtGui.QTableWidgetItem(str(i))
self.table.setVerticalHeaderItem(i, iItem)
self.table.verticalHeaderItem(i).setToolTip(str(i))
#時間軸にデータを入れるなら↓
#self.table.verticalHeaderItem(i).setToolTip(str(times[i]))
for j in range(t_row):
if hor == True:
jItem = QtGui.QTableWidgetItem(str(j))
self.table.setHorizontalHeaderItem(j, jItem)
self.table.horizontalHeaderItem(j).setToolTip(str(times[j]))
#print "%.10f"%times[j]
if i < d_col: #data(speak)の可視化
# 音声Dataがrmsの場合
# set_data:範囲は 0-1
set_data = data[j][i] #self.speakDecibelNorm(data[j][i])
#ON/OFF むちゃくちゃすぎる
#set_data = 1 if set_data > 0.3 else 0
#self.edited_speaks[j][i] = set_data
#一時的なOffset
#set_data = data_proc2.speakNorm(set_data, upper=0.75, lower=0.25)
#self.edited_speaks[j][i] = set_data
#color_dataの範囲を0-255に変更
#color_data=int(set_data*255)
color_data = set_data
color_data = 1 if color_data > 0.5 else 0
color_data=int(color_data*255)
#print color_data
#print "at",color_data
color = [255-color_data]*3
elif i >= d_col and i < t_col-add_flag: # annotationの可視化
#print i
"""
set_data = anno[j][i-d_col] #anno(1000, 2)
if set_data == 0:
color = [255, 255, 255]
else:
color = [0, 0, 0]
"""
set_data = anno[j][i-d_col] #anno(1000, 2)
set_data = 0 if set_data < 0 else set_data
set_data = 1 if set_data > 1 else set_data
color_data=int(set_data*255)
color = [255-color_data]*3
else: #flag
set_data = self.edited_flags[j][0]
if set_data == 0:
color = [255, 255, 255]
else:
color = [0, 0, 0]
self.table.setItem(i, j, QtGui.QTableWidgetItem())
self.table.item(i, j).setBackground(QtGui.QColor(color[0],color[1],color[2]))
self.table.item(i, j).setToolTip(str(set_data))
hor = False
self.table.setVisible(False)
self.table.resizeRowsToContents()
self.table.resizeColumnsToContents()
self.table.setVisible(True)
# TableがClickされたとき
def clickUpdateTable(self, cItem):
self.tip = float(cItem.toolTip())
self.r = cItem.row()
self.c = cItem.column()
print "r:",self.r,", c:",self.c, ". tip:",self.tip
set_data = 0 if self.tip == 1 else 1
speak_dim = self.edited_speaks.shape[1]
anno_dim = self.edited_annos.shape[1]
if self.r < speak_dim:
self.edited_speaks[self.c][self.r] = set_data
elif self.r >= speak_dim and self.r < speak_dim+anno_dim:
self.edited_annos[self.c][self.r-speak_dim] = set_data
else:
self.edited_flags[self.c][0] = set_data
#indexes = self.table.selectedIndexes()
#print indexes
color = [[255, 255, 255], [0, 0, 0]]
self.table.setItem(self.r, self.c, QtGui.QTableWidgetItem())
self.table.item(self.r, self.c).setBackground(QtGui.QColor(color[set_data][0],color[set_data][1],color[set_data][2]))
self.table.item(self.r, self.c).setToolTip(str(set_data))
#jointsの可視化
# self.vizJoint(self.c)
def checkRangeData(self):
set_row = int(self.checkUser.text()) #0~4 User1=2, User2=3
start = int(self.checkSt.text())
end = int(self.checkEd.text())
color = [[255, 255, 255], [0, 0, 0]]
table_offset = self.edited_speaks.shape[1]
#変更するUserの指定
user = set_row - table_offset
#print user, table_offset
if user < 0:
print "Not Change!"
return
print "Change [User:", user, ", Start:",start,", end:",end,"]"
for i in range(start, end):
get_data = self.edited_annos[i][user]
set_data = 0 if get_data == 1 else 1
self.table.setItem(set_row, i, QtGui.QTableWidgetItem())
self.table.item(set_row, i).setBackground(QtGui.QColor(color[set_data][0],color[set_data][1],color[set_data][2]))
self.table.item(set_row, i).setToolTip(str(set_data))
self.edited_annos[i][user] = set_data
def sliderChange(self, timeline):
if len(self.input_joints)==0:
print "now no data:", timeline
return
#self.table.selectColum(timeline)
#self.table.verticalScrollBar().setValue(timeline)
self.table.setCurrentCell(self.r, timeline)
self.pubViz(timeline)
""" | random_line_split |
||
ann_interface.py | .2)
plt.legend()
plt.show()
"""
print "joints shape:", self.input_joints.shape
print "speaks shape:", self.input_speaks.shape
print "annos shape:", self.input_annos.shape
print "flags shape:", self.input_flags.shape
print "times shape:", self.input_times.shape
#for j in range(len(self.input_times)):
# print j,"%.10f"%self.input_times[j]
self.timediff(self.input_times)
self.edited_joints = copy.deepcopy(self.input_joints)#np.arrayでCopyされる
self.edited_speaks = copy.deepcopy(self.input_speaks)#speakは編集しない(今のところ)
self.edited_annos = copy.deepcopy(self.input_annos)
self.edited_flags = copy.deepcopy(self.input_flags)
self.edited_times = copy.deepcopy(self.input_times)
#self.updateTable(self.input_joints)
self.updateTable(self.edited_speaks, self.edited_annos, self.edited_times)
#Sliderの最大値をset
self.sld.setMaximum(datalen - 1)
print "end"
def outputData(self):
name_json = str(self.txtOutputFile.text())
keys = ["joints", "speaks", "annos", "flags", "times"]
data = [self.edited_joints, self.edited_speaks, self.edited_annos, self.edited_flags, self.edited_times]
"""
if len(keys) != len(data):
print "Save false! keys len:"
"""
data_proc2.save_data(name_json, keys, data)
def timediff(self, times):
fpss = []
for i in range(len(times)-1):
fps = 1/(times[i+1]-times[i])
fpss.append(fps)
#print i, fps
fpss = np.array(fpss)
#print "mean:",np.mean(fpss),",std:",np.std(fpss)
def cutRangeData(self):
start = int(self.cutStart.text())
end = int(self.cutEnd.text())
print "cut data:",start,"-",end
self.edited_joints = self.edited_joints[start:end]
self.edited_speaks = self.edited_speaks[start:end]
self.edited_annos = self.edited_annos[start:end]
self.edited_flags = self.edited_flags[start:end]
self.edited_times = self.edited_times[start:end]
print "joints shape:", self.edited_joints.shape
print "speaks shape:", self.edited_speaks.shape
print "annos shape:", self.edited_annos.shape
print "flags shape:", self.edited_flags.shape
print "times shape:", self.edited_times.shape
self.updateTable(self.edited_speaks, self.edited_annos, self.edited_times)
#Sliderの最大値をset
self.sld.setMaximum(self.edited_joints.shape[0]-1)
print "end"
def directJoints(self):
print "direct joints"
size, dim = self.edited_joints.shape
user1_nose_y_idx = 35+33+2
pair = 1
pair_stack = 0
offset_y = 0.03
for i in range(size):
anno = self.edited_annos[i]
if anno[1] == 0:#聞き手なら
if anno[0] == 1 and anno[2] == 0:
#print i,":user0",self.input_joints[i][user1_nose_y_idx]
self.edited_joints[i][user1_nose_y_idx] += offset_y
pair = 0
pair_stack = 0
#print self.input_joints[i][user1_nose_y_idx]
elif anno[0] == 0 and anno[2] == 1:
#print i,":user2",self.input_joints[i][user1_nose_y_idx]
self.edited_joints[i][user1_nose_y_idx] -= offset_y
pair = 2
pair_stack = 0
#print self.input_joints[i][user1_nose_y_idx]
else: #誰も話していないなら
pair_stack += 1
else:#話し手ならどこを向くか
if pair == 0:
self.edited_joints[i][user1_nose_y_idx] += offset_y
elif pair == 2:
self.edited_joints[i][user1_nose_y_idx] -= offset_y
# 引っ繰り返して追加
def reverseData(self):
#y方向だけ引っくり返す
def reverse_y(joints):
#(10000, 36)
rev = np.array([[1,-1,1]*12]*joints.shape[0])
return joints*rev
print "reverseData before:",self.edited_joints.shape, self.edited_annos.shape
datalen = self.edited_annos.shape[0]
if self.anno_dim == 2:
j_r = np.hstack((self.edited_joints[:,36:], self.edited_joints[:,:36]))
self.edited_joints = np.vstack((self.edited_joints, j_r))
s_r = np.hstack((self.edited_speaks[:,1].reshape(datalen,1), self.edited_speaks[:,0].reshape(datalen,1)))
self.edited_speaks = np.vstack((self.edited_speaks, s_r))
a_r = np.hstack((self.edited_annos[:,1].reshape(datalen,1), self.edited_annos[:,0].reshape(datalen,1)))
self.edited_annos = np.vstack((self.edited_annos, a_r))
self.edited_flags = np.vstack((self.edited_flags, self.edited_flags))
self.edited_times = np.append(self.edited_times, self.edited_times)
else:
print "bad shape"
#print self.edited_times.shape
self.updateTable(self.edited_speaks, self.edited_annos, self.edited_times)
#Sliderの最大値をset
self.sld.setMaximum(self.edited_joints.shape[0]-1)
print "reverseData after:",self.edited_joints.shape,self.edited_speaks.shape,self.edited_annos.shape
def rviz_obj(self, obj_id, obj_ns, obj_type, obj_size, obj_color=[0, 0, 0, 0], obj_life=0):
obj = Marker()
obj.header.frame_id, obj.header.stamp = "camera_link", rospy.Time.now()
obj.ns, obj.action, obj.type = str(obj_ns), 0, obj_type
obj.scale.x, obj.scale.y, obj.scale.z = obj_size[0], obj_size[1], obj_size[2]
obj.color = obj_color
obj.lifetime = rospy.Duration.from_sec(obj_life)
obj.pose.orientation.w = 1.0
return obj
def set_point(self, pos, addx=0, addy=0, addz=0, rotate=False):
pt = Point()
if rotate == True:
pt.x, pt.y, pt.z = -1*pos[0]+addx, -1*pos[1]+addy, pos[2]+addz
else:
pt.x, pt.y, pt.z = pos[0]+addx, pos[1]+addy, pos[2]+addz
return pt
def set_vizmsg_point(self, u, data, color, psize, ofs, addx=0, addy=0, rotate=False):
pmsg = self.rviz_obj(u, 'p'+str(u), 7, [psize, psize, psize], color, 0)
points = []
for p in range(data.shape[1]/ofs):
points.append(self.set_point([data[0, p*ofs],
data[0, p*ofs+1],
data[0, p*ofs+2]],
addx=addx, addy=addy, rotate=rotate))
pmsg.points = points
return pmsg
def set_vizmsg_line(self, u, data, color, lsize, llist, addx=0, addy=0, rotate=False):
lmsg = self.rviz_obj(u, 'l'+str(u), 5, [lsize, lsize, lsize], color, 0)
for ls in llist:
for l in range(len(ls)-1):
for add in r | ange(2):
#print person[0, ls[l+add]], ls[l+add], l, add
linepoint=self.set_point([data[0,ls[l+add]*3],
data[0,ls[l+add]*3+1],
data[0,ls[l+add]*3+2]],
addx=addx, addy=addy, rotate=rotate)
lmsg.points.append(linepoint)
return lmsg
| identifier_body |
|
ann_interface.py | else:
print "create flags"
self.input_flags = np.zeros((datalen, 1))
if input_data.has_key("times"):
print "load times"
self.input_times = input_data["times"]
else:
print "create times"
self.input_times = np.zeros((datalen, 1))
"""
diff_times = []
for t in range(len(self.input_times)):
if t == 0:
diff_times.append(0)
else:
diff_times.append(self.input_times[t]-self.input_times[t-1])
fps = np.round(1/float(np.mean(diff_times)))
plt.plot(self.input_annos[:,0], label="person", color="red")
plt.plot(self.input_annos[:,1], label="robot", color="green")
plt.xlabel("Frame (fps:"+str(fps)+")")
plt.title("speaker/listener result")
plt.xlim(0,len(self.input_annos))
plt.ylim(-0.2, 1.2)
plt.legend()
plt.show()
"""
print "joints shape:", self.input_joints.shape
print "speaks shape:", self.input_speaks.shape
print "annos shape:", self.input_annos.shape
print "flags shape:", self.input_flags.shape
print "times shape:", self.input_times.shape
#for j in range(len(self.input_times)):
# print j,"%.10f"%self.input_times[j]
self.timediff(self.input_times)
self.edited_joints = copy.deepcopy(self.input_joints)#np.arrayでCopyされる
self.edited_speaks = copy.deepcopy(self.input_speaks)#speakは編集しない(今のところ)
self.edited_annos = copy.deepcopy(self.input_annos)
self.edited_flags = copy.deepcopy(self.input_flags)
self.edited_times = copy.deepcopy(self.input_times)
#self.updateTable(self.input_joints)
self.updateTable(self.edited_speaks, self.edited_annos, self.edited_times)
#Sliderの最大値をset
self.sld.setMaximum(datalen - 1)
print "end"
def outputData(self):
name_json = str(self.txtOutputFile.text())
keys = ["joints", "speaks", "annos", "flags", "times"]
data = [self.edited_joints, self.edited_speaks, self.edited_annos, self.edited_flags, self.edited_times]
"""
if len(keys) != len(data):
print "Save false! keys len:"
"""
data_proc2.save_data(name_json, keys, data)
def timediff(self, times):
fpss = []
for i in range(len(times)-1):
fps = 1/(times[i+1]-times[i])
fpss.append(fps)
#print i, fps
fpss = np.array(fpss)
#print "mean:",np.mean(fpss),",std:",np.std(fpss)
def cutRangeData(self):
start = int(self.cutStart.text())
end = int(self.cutEnd.text())
print "cut data:",start,"-",end
self.edited_joints = self.edited_joints[start:end]
self.edited_speaks = self.edited_speaks[start:end]
self.edited_annos = self.edited_annos[start:end]
self.edited_flags = self.edited_flags[start:end]
self.edited_times = self.edited_times[start:end]
print "joints shape:", self.edited_joints.shape
print "speaks shape:", self.edited_speaks.shape
print "annos shape:", self.edited_annos.shape
print "flags shape:", self.edited_flags.shape
print "times shape:", self.edited_times.shape
self.updateTable(self.edited_speaks, self.edited_annos, self.edited_times)
#Sliderの最大値をset
self.sld.setMaximum(self.edited_joints.shape[0]-1)
print "end"
def directJoints(self):
print "direct joints"
size, dim = self.edited_joints.shape
user1_nose_y_idx = 35+33+2
pair = 1
pair_stack = 0
offset_y = 0.03
for i in range(size):
anno = self.edited_annos[i]
if anno[1] == 0:#聞き手なら
if anno[0] == 1 and anno[2] == 0:
#print i,":user0",self.input_joints[i][user1_nose_y_idx]
self.edited_joints[i][user1_nose_y_idx] += offset_y
pair = 0
pair_stack = 0
#print self.input_joints[i][user1_nose_y_idx]
elif anno[0] == 0 and anno[2] == 1:
#print i,":user2",self.input_joints[i][user1_nose_y_idx]
self.edited_joints[i][user1_nose_y_idx] -= offset_y
pair = 2
pair_stack = 0
#print self.input_joints[i][user1_nose_y_idx]
else: #誰も話していないなら
pair_stack += 1
else:#話し手ならどこを向くか
if pair == 0:
self.edited_joints[i][user1_nose_y_idx] += offset_y
elif pair == 2:
self.edited_joints[i][user1_nose_y_idx] -= offset_y
# 引っ繰り返して追加
def reverseData(self):
#y方向だけ引っくり返す
def reverse_y(joints):
#(10000, 36)
rev = np.array([[1,-1,1]*12]*joints.shape[0])
return joints*rev
print "reverseData before:",self.edited_joints.shape, self.edited_annos.shape
datalen = self.edited_annos.shape[0]
if self.anno_dim == 2:
j_r = np.hstack((self.edited_joints[:,36:], self.edited_joints[:,:36]))
self.edited_joints = np.vstack((self.edited_joints, j_r))
s_r = np.hstack((self.edited_speaks[:,1].reshape(datalen,1), self.edited_speaks[:,0].reshape(datalen,1)))
self.edited_speaks = np.vstack((self.edited_speaks, s_r))
a_r = np.hstack((self.edited_annos[:,1].reshape(datalen,1), self.edited_annos[:,0].reshape(datalen,1)))
self.edited_annos = np.vstack((self.edited_annos, a_r))
self.edited_flags = np.vstack((self.edited_flags, self.edited_flags))
self.edited_times = np.append(self.edited_times, self.edited_times)
else:
print "bad shape"
#print self.edited_times.shape
self.updateTable(self.edited_speaks, self.edited_annos, self.edited_times)
#Sliderの最大値をset
self.sld.setMaximum(self.edited_joints.shape[0]-1)
print "reverseData after:",self.edited_joints.shape,self.edited_speaks.shape,self.edited_annos.shape
def rviz_obj(self, obj_id, obj_ns, obj_type, obj_size, obj_color=[0, 0, 0, 0], obj_life=0):
obj = Marker()
obj.header.frame_id, obj.header.stamp = "camera_link", rospy.Time.now()
obj.ns, obj.action, obj.type = str(obj_ns), 0, obj_type
obj.scale.x, obj.scale.y, obj.scale.z = obj_size[0], obj_size[1], obj_size[2]
obj.color = obj_color
obj.lifetime = rospy.Duration.from_sec(obj_life)
obj.pose.orientation.w = 1.0
return obj
def set_point(self, pos, addx=0, addy=0, addz=0, rotate=False):
pt = Point()
if rotate == True:
pt.x, pt.y, pt.z = -1*pos[0]+addx, -1*pos[1]+addy, pos[2]+addz
else:
pt.x, pt.y, pt.z = pos[0]+addx, pos[1]+addy, pos[2]+addz
return pt
def set_vizmsg_point(self, u, data, color, psize, ofs, addx=0, addy=0, rotate=False):
pmsg = self.rviz_obj(u, 'p'+str(u), 7, [psize, psize, psize], color, 0)
points = []
for p in range(data.shape[1]/ofs):
points.append(self.set_point([data[0, p*ofs],
data[0, p*ofs+1],
data[0, p*ofs+2]], | identifier_name |
||
index.js | = [];
function next(index) {
if (index < array.length) | else {
resolve(results);
}
}
next(0);
});
}
function findPromise(array, callback, thisArg) {
return new Promise((resolve, reject) => {
function next(index) {
if (index < array.length) {
callback.call(thisArg, array[index], index, array).then(result => {
if (result) {
resolve(array[index]);
} else {
next(index + 1);
}
}).catch(error => {
reject(error);
});
} else {
resolve();
}
}
next(0);
});
}
async function readDirRecursively(dir, options) {
const rootDir = _.get(options, 'rootDir', dir);
const files = await fse.readdir(dir);
const result = await mapPromise(files, async (file) => {
const filePath = path.join(dir, file);
const relFilePath = path.relative(rootDir, filePath);
const stats = await fse.stat(filePath);
if (_.has(options, 'filter') && !options.filter(relFilePath, stats)) {
return Promise.resolve();
}
if (stats.isDirectory()) {
return readDirRecursively(filePath, {...options, rootDir});
} else if (stats.isFile()) {
return relFilePath;
} else {
return null;
}
});
return _.chain(result).compact().flatten().value();
}
/**
* Gets the value at the first path of object having non undefined value.
* If all paths resolve to undefined values, the defaultValue is returned.
*
* @param {Object} object The object to query.
* @param {Array<String | Array<String>>} paths The property paths to search for.
* @param {*} [defaultValue] The value returned if all paths resolve to undefined values
* @returns {*}
*/
function getFirst(object, paths, defaultValue) {
let result = _(object).at(paths).reject(_.isUndefined).first();
return _.isUndefined(result) ? defaultValue : result;
}
function append(object, path, value) {
if (!_.has(object, path)) {
_.set(object, path, []);
}
_.get(object, path).push(value);
}
function concat(object, path, value) {
if (!_.has(object, path)) {
_.set(object, path, []);
}
_.set(object, path, _.get(object, path).concat(value));
}
function copy(sourceObject, sourcePath, targetObject, targetPath, transform) {
if (_.has(sourceObject, sourcePath)) {
let value = _.get(sourceObject, sourcePath);
if (transform) {
value = transform(value);
}
_.set(targetObject, targetPath, value);
}
}
function copyIfNotSet(sourceObject, sourcePath, targetObject, targetPath, transform) {
if (!_.has(targetObject, targetPath)) {
copy(sourceObject, sourcePath, targetObject, targetPath, transform);
}
}
function rename(object, oldPath, newPath) {
if (_.has(object, oldPath)) {
_.set(object, newPath, _.get(object, oldPath));
oldPath = _.toPath(oldPath);
if (oldPath.length > 1) {
object = _.get(object, _.initial(oldPath));
}
delete object[_.last(oldPath)];
}
}
function failFunctionWithTag(tag) {
return function fail(message) {
throw new Error(`[${tag}] ${message}`);
};
}
function assertFunctionWithFail(fail) {
return function assert(value, message) {
if (!value) {
fail(message);
}
}
}
/**
* Deeply maps the passed `value` by recursively calling the `iteratee` on
* value's children. The value returned from the iteratee is used to map the
* children.
*
* The iteratee is invoked with three arguments - the `value` being iterated,
* the `fieldPath` of the current `value` relative to the original passed value,
* and the `stack` of ancestors of the current `value`.
*
* For the first time the `iterate` will be called with the original `value`
* and empty arrays for `fieldPath` and `stack`.
*
* In other words, any `value` passed to the iteratee (except the first call,
* and assuming the ancestors keys were not mapped)
* will be equal to: `_.get(originalValue, fieldPath)`
*
* The recursion is called in pre-order depth-first-search. Meaning, the
* iteratee is called first on parent nodes and then on its children. Therefore
* if iteratee maps/replaces the parent node, then the children of the replaced
* node will be traversed.
*
* @example
* mapDeep({ prop: 'foo', arr: [ 'bar' , 1, 2 ] }, (value) => {
* if (_.isString(value)) return '__' + value;
* if (_.isNumber(value)) return value * 10;
* return value;
* })
* => { prop: '__foo', arr: [ '__bar', 10, 20 ] }
*
* mapDeep({ prop: 'foo', arr: [ 'bar' ] }, (value, fieldPath) => {
* if ((_.isString(value)) return value + '__' + fieldPath.join('.');
* return value;
* })
* => { prop: 'foo__prop', arr: [ 'bar__arr.0' ] }
*
* @param {*} value A value to map
* @param {Function} iteratee Function (value: any, fieldPath: Array, stack: Array)
* @param {object} [options]
* @param {boolean} [options.iterateCollections] Default: true
* @param {boolean} [options.iterateScalars] Default: true
* @param {boolean} [options.postOrder] Change the invocation of iteratee to post-order depth-first-search. Default: false
* @param {Array} [_keyPath] For internal recursive use
* @param {Array} [_objectStack] For internal recursive use
* @returns {*}
*/
function mapDeep(value, iteratee, options = {}, _keyPath = [], _objectStack = []) {
const postOrder = _.get(options, 'postOrder', false);
let iterate;
if (_.isPlainObject(value) || _.isArray(value)) {
iterate = _.get(options, 'iterateCollections', true);
} else {
iterate = _.get(options, 'iterateScalars', true);
}
if (iterate && !postOrder) {
value = iteratee(value, _keyPath, _objectStack);
}
const childrenIterator = (val, key) => {
return mapDeep(val, iteratee, options, _.concat(_keyPath, key), _.concat(_objectStack, [value]));
};
if (_.isPlainObject(value)) {
value = _.mapValues(value, childrenIterator);
} else if (_.isArray(value)) {
value = _.map(value, childrenIterator);
}
if (iterate && postOrder) {
value = iteratee(value, _keyPath, _objectStack);
}
return value;
}
function fieldPathToString(fieldPath) {
return _.reduce(fieldPath, (accumulator, fieldName, index) => {
if (_.isString(fieldName) && /\W/.test(fieldName)) {
// field name is a string with non alphanumeric character
accumulator += `['${fieldName}']`;
} else if (_.isNumber(fieldName)) {
accumulator += `[${fieldName}]`;
} else {
if (index > 0) {
accumulator += '.';
}
accumulator += fieldName;
}
return accumulator;
}, '');
}
function getFirstExistingFile(fileNames, inputDir) {
const filePaths = _.map(fileNames, fileName => path.resolve(inputDir, fileName));
return findPromise(filePaths, (filePath) => fse.pathExists(filePath));
}
function parseFirstExistingFile(fileNames, inputDir) {
return getFirstExistingFile(fileNames, inputDir).then(filePath => {
if (filePath) {
return parseFile(filePath);
} else {
return null;
}
});
}
async function parseFile(filePath) {
const data = await fse.readFile(filePath, 'utf8');
return parseDataByFilePath(data, filePath);
}
function parseDataByFilePath(string, filePath) {
const extension = path.extname(filePath).substring(1);
let data;
switch (extension) {
case 'yml':
case 'yaml':
data = yaml.safeLoad(string, {schema: yaml.JSON_SCHEMA});
break;
case 'json':
data = JSON.parse(string);
break;
case 'toml':
data = toml.parse(string);
break;
case 'md':
data = parseMarkdownWithFrontMatter(string);
break;
default:
throw new Error(`parseDataByFilePath error, extension '${extension}' of file ${filePath} is not supported`);
}
return data;
}
function parseMarkdownWithFrontMatter(string) {
string = string.replace('\r\n', '\n');
let frontmatter = null;
let markdown = string;
let frontMatterTypes = [
{
type: 'yaml',
startDelimiter: '---\n | {
callback.call(thisArg, array[index], index, array).then(result => {
results[index] = result;
next(index + 1);
}).catch(error => {
reject(error);
});
} | conditional_block |
index.js | {Array<String | Array<String>>} paths The property paths to search for.
* @param {*} [defaultValue] The value returned if all paths resolve to undefined values
* @returns {*}
*/
function getFirst(object, paths, defaultValue) {
let result = _(object).at(paths).reject(_.isUndefined).first();
return _.isUndefined(result) ? defaultValue : result;
}
function append(object, path, value) {
if (!_.has(object, path)) {
_.set(object, path, []);
}
_.get(object, path).push(value);
}
function concat(object, path, value) {
if (!_.has(object, path)) {
_.set(object, path, []);
}
_.set(object, path, _.get(object, path).concat(value));
}
function copy(sourceObject, sourcePath, targetObject, targetPath, transform) {
if (_.has(sourceObject, sourcePath)) {
let value = _.get(sourceObject, sourcePath);
if (transform) {
value = transform(value);
}
_.set(targetObject, targetPath, value);
}
}
function copyIfNotSet(sourceObject, sourcePath, targetObject, targetPath, transform) {
if (!_.has(targetObject, targetPath)) {
copy(sourceObject, sourcePath, targetObject, targetPath, transform);
}
}
function rename(object, oldPath, newPath) {
if (_.has(object, oldPath)) {
_.set(object, newPath, _.get(object, oldPath));
oldPath = _.toPath(oldPath);
if (oldPath.length > 1) {
object = _.get(object, _.initial(oldPath));
}
delete object[_.last(oldPath)];
}
}
function failFunctionWithTag(tag) {
return function fail(message) {
throw new Error(`[${tag}] ${message}`);
};
}
function assertFunctionWithFail(fail) {
return function assert(value, message) {
if (!value) {
fail(message);
}
}
}
/**
* Deeply maps the passed `value` by recursively calling the `iteratee` on
* value's children. The value returned from the iteratee is used to map the
* children.
*
* The iteratee is invoked with three arguments - the `value` being iterated,
* the `fieldPath` of the current `value` relative to the original passed value,
* and the `stack` of ancestors of the current `value`.
*
* For the first time the `iterate` will be called with the original `value`
* and empty arrays for `fieldPath` and `stack`.
*
* In other words, any `value` passed to the iteratee (except the first call,
* and assuming the ancestors keys were not mapped)
* will be equal to: `_.get(originalValue, fieldPath)`
*
* The recursion is called in pre-order depth-first-search. Meaning, the
* iteratee is called first on parent nodes and then on its children. Therefore
* if iteratee maps/replaces the parent node, then the children of the replaced
* node will be traversed.
*
* @example
* mapDeep({ prop: 'foo', arr: [ 'bar' , 1, 2 ] }, (value) => {
* if (_.isString(value)) return '__' + value;
* if (_.isNumber(value)) return value * 10;
* return value;
* })
* => { prop: '__foo', arr: [ '__bar', 10, 20 ] }
*
* mapDeep({ prop: 'foo', arr: [ 'bar' ] }, (value, fieldPath) => {
* if ((_.isString(value)) return value + '__' + fieldPath.join('.');
* return value;
* })
* => { prop: 'foo__prop', arr: [ 'bar__arr.0' ] }
*
* @param {*} value A value to map
* @param {Function} iteratee Function (value: any, fieldPath: Array, stack: Array)
* @param {object} [options]
* @param {boolean} [options.iterateCollections] Default: true
* @param {boolean} [options.iterateScalars] Default: true
* @param {boolean} [options.postOrder] Change the invocation of iteratee to post-order depth-first-search. Default: false
* @param {Array} [_keyPath] For internal recursive use
* @param {Array} [_objectStack] For internal recursive use
* @returns {*}
*/
function mapDeep(value, iteratee, options = {}, _keyPath = [], _objectStack = []) {
const postOrder = _.get(options, 'postOrder', false);
let iterate;
if (_.isPlainObject(value) || _.isArray(value)) {
iterate = _.get(options, 'iterateCollections', true);
} else {
iterate = _.get(options, 'iterateScalars', true);
}
if (iterate && !postOrder) {
value = iteratee(value, _keyPath, _objectStack);
}
const childrenIterator = (val, key) => {
return mapDeep(val, iteratee, options, _.concat(_keyPath, key), _.concat(_objectStack, [value]));
};
if (_.isPlainObject(value)) {
value = _.mapValues(value, childrenIterator);
} else if (_.isArray(value)) {
value = _.map(value, childrenIterator);
}
if (iterate && postOrder) {
value = iteratee(value, _keyPath, _objectStack);
}
return value;
}
function fieldPathToString(fieldPath) {
return _.reduce(fieldPath, (accumulator, fieldName, index) => {
if (_.isString(fieldName) && /\W/.test(fieldName)) {
// field name is a string with non alphanumeric character
accumulator += `['${fieldName}']`;
} else if (_.isNumber(fieldName)) {
accumulator += `[${fieldName}]`;
} else {
if (index > 0) {
accumulator += '.';
}
accumulator += fieldName;
}
return accumulator;
}, '');
}
function getFirstExistingFile(fileNames, inputDir) {
const filePaths = _.map(fileNames, fileName => path.resolve(inputDir, fileName));
return findPromise(filePaths, (filePath) => fse.pathExists(filePath));
}
function parseFirstExistingFile(fileNames, inputDir) {
return getFirstExistingFile(fileNames, inputDir).then(filePath => {
if (filePath) {
return parseFile(filePath);
} else {
return null;
}
});
}
async function parseFile(filePath) {
const data = await fse.readFile(filePath, 'utf8');
return parseDataByFilePath(data, filePath);
}
function parseDataByFilePath(string, filePath) {
const extension = path.extname(filePath).substring(1);
let data;
switch (extension) {
case 'yml':
case 'yaml':
data = yaml.safeLoad(string, {schema: yaml.JSON_SCHEMA});
break;
case 'json':
data = JSON.parse(string);
break;
case 'toml':
data = toml.parse(string);
break;
case 'md':
data = parseMarkdownWithFrontMatter(string);
break;
default:
throw new Error(`parseDataByFilePath error, extension '${extension}' of file ${filePath} is not supported`);
}
return data;
}
function parseMarkdownWithFrontMatter(string) {
string = string.replace('\r\n', '\n');
let frontmatter = null;
let markdown = string;
let frontMatterTypes = [
{
type: 'yaml',
startDelimiter: '---\n',
endDelimiter: '\n---',
parse: (string) => yaml.safeLoad(string, {schema: yaml.JSON_SCHEMA})
},
{
type: 'toml',
startDelimiter: '+++\n',
endDelimiter: '\n+++',
parse: (string) => toml.parse(string)
},
{
type: 'json',
startDelimiter: '{\n',
endDelimiter: '\n}',
parse: (string) => JSON.parse(string)
}
];
_.forEach(frontMatterTypes, fmType => {
if (string.startsWith(fmType.startDelimiter)) {
let index = string.indexOf(fmType.endDelimiter);
if (index !== -1) {
// The end delimiter must be followed by EOF or by a new line (possibly preceded with spaces)
// For example ("." used for spaces):
// |---
// |title: Title
// |---...
// |
// |Markdown Content
// |
// "index" points to the beginning of the second "---"
// "endDelimEndIndex" points to the end of the second "---"
// "afterEndDelim" is everything after the second "---"
// "afterEndDelimMatch" is the matched "...\n" after the second "---"
// frontmatter will be: {title: "Title"}
// markdown will be "\nMarkdown Content\n" (the first \n after end delimiter is discarded)
const endDelimEndIndex = index + fmType.endDelimiter.length;
const afterEndDelim = string.substring(endDelimEndIndex);
const afterEndDelimMatch = afterEndDelim.match(/^\s*?(\n|$)/);
if (afterEndDelimMatch) {
const data = string.substring(fmType.startDelimiter.length, index);
frontmatter = fmType.parse(data); | random_line_split |
||
index.js | , thisArg) {
return new Promise((resolve, reject) => {
function next(index) {
if (index < array.length) {
callback.call(thisArg, array[index], index, array).then(result => {
if (result) {
resolve(array[index]);
} else {
next(index + 1);
}
}).catch(error => {
reject(error);
});
} else {
resolve();
}
}
next(0);
});
}
async function readDirRecursively(dir, options) {
const rootDir = _.get(options, 'rootDir', dir);
const files = await fse.readdir(dir);
const result = await mapPromise(files, async (file) => {
const filePath = path.join(dir, file);
const relFilePath = path.relative(rootDir, filePath);
const stats = await fse.stat(filePath);
if (_.has(options, 'filter') && !options.filter(relFilePath, stats)) {
return Promise.resolve();
}
if (stats.isDirectory()) {
return readDirRecursively(filePath, {...options, rootDir});
} else if (stats.isFile()) {
return relFilePath;
} else {
return null;
}
});
return _.chain(result).compact().flatten().value();
}
/**
* Gets the value at the first path of object having non undefined value.
* If all paths resolve to undefined values, the defaultValue is returned.
*
* @param {Object} object The object to query.
* @param {Array<String | Array<String>>} paths The property paths to search for.
* @param {*} [defaultValue] The value returned if all paths resolve to undefined values
* @returns {*}
*/
function getFirst(object, paths, defaultValue) {
let result = _(object).at(paths).reject(_.isUndefined).first();
return _.isUndefined(result) ? defaultValue : result;
}
function append(object, path, value) {
if (!_.has(object, path)) {
_.set(object, path, []);
}
_.get(object, path).push(value);
}
function concat(object, path, value) {
if (!_.has(object, path)) {
_.set(object, path, []);
}
_.set(object, path, _.get(object, path).concat(value));
}
function copy(sourceObject, sourcePath, targetObject, targetPath, transform) {
if (_.has(sourceObject, sourcePath)) {
let value = _.get(sourceObject, sourcePath);
if (transform) {
value = transform(value);
}
_.set(targetObject, targetPath, value);
}
}
function copyIfNotSet(sourceObject, sourcePath, targetObject, targetPath, transform) {
if (!_.has(targetObject, targetPath)) {
copy(sourceObject, sourcePath, targetObject, targetPath, transform);
}
}
function rename(object, oldPath, newPath) {
if (_.has(object, oldPath)) {
_.set(object, newPath, _.get(object, oldPath));
oldPath = _.toPath(oldPath);
if (oldPath.length > 1) {
object = _.get(object, _.initial(oldPath));
}
delete object[_.last(oldPath)];
}
}
function failFunctionWithTag(tag) {
return function fail(message) {
throw new Error(`[${tag}] ${message}`);
};
}
function assertFunctionWithFail(fail) {
return function assert(value, message) {
if (!value) {
fail(message);
}
}
}
/**
* Deeply maps the passed `value` by recursively calling the `iteratee` on
* value's children. The value returned from the iteratee is used to map the
* children.
*
* The iteratee is invoked with three arguments - the `value` being iterated,
* the `fieldPath` of the current `value` relative to the original passed value,
* and the `stack` of ancestors of the current `value`.
*
* For the first time the `iterate` will be called with the original `value`
* and empty arrays for `fieldPath` and `stack`.
*
* In other words, any `value` passed to the iteratee (except the first call,
* and assuming the ancestors keys were not mapped)
* will be equal to: `_.get(originalValue, fieldPath)`
*
* The recursion is called in pre-order depth-first-search. Meaning, the
* iteratee is called first on parent nodes and then on its children. Therefore
* if iteratee maps/replaces the parent node, then the children of the replaced
* node will be traversed.
*
* @example
* mapDeep({ prop: 'foo', arr: [ 'bar' , 1, 2 ] }, (value) => {
* if (_.isString(value)) return '__' + value;
* if (_.isNumber(value)) return value * 10;
* return value;
* })
* => { prop: '__foo', arr: [ '__bar', 10, 20 ] }
*
* mapDeep({ prop: 'foo', arr: [ 'bar' ] }, (value, fieldPath) => {
* if ((_.isString(value)) return value + '__' + fieldPath.join('.');
* return value;
* })
* => { prop: 'foo__prop', arr: [ 'bar__arr.0' ] }
*
* @param {*} value A value to map
* @param {Function} iteratee Function (value: any, fieldPath: Array, stack: Array)
* @param {object} [options]
* @param {boolean} [options.iterateCollections] Default: true
* @param {boolean} [options.iterateScalars] Default: true
* @param {boolean} [options.postOrder] Change the invocation of iteratee to post-order depth-first-search. Default: false
* @param {Array} [_keyPath] For internal recursive use
* @param {Array} [_objectStack] For internal recursive use
* @returns {*}
*/
function mapDeep(value, iteratee, options = {}, _keyPath = [], _objectStack = []) {
const postOrder = _.get(options, 'postOrder', false);
let iterate;
if (_.isPlainObject(value) || _.isArray(value)) {
iterate = _.get(options, 'iterateCollections', true);
} else {
iterate = _.get(options, 'iterateScalars', true);
}
if (iterate && !postOrder) {
value = iteratee(value, _keyPath, _objectStack);
}
const childrenIterator = (val, key) => {
return mapDeep(val, iteratee, options, _.concat(_keyPath, key), _.concat(_objectStack, [value]));
};
if (_.isPlainObject(value)) {
value = _.mapValues(value, childrenIterator);
} else if (_.isArray(value)) {
value = _.map(value, childrenIterator);
}
if (iterate && postOrder) {
value = iteratee(value, _keyPath, _objectStack);
}
return value;
}
function fieldPathToString(fieldPath) {
return _.reduce(fieldPath, (accumulator, fieldName, index) => {
if (_.isString(fieldName) && /\W/.test(fieldName)) {
// field name is a string with non alphanumeric character
accumulator += `['${fieldName}']`;
} else if (_.isNumber(fieldName)) {
accumulator += `[${fieldName}]`;
} else {
if (index > 0) {
accumulator += '.';
}
accumulator += fieldName;
}
return accumulator;
}, '');
}
function getFirstExistingFile(fileNames, inputDir) {
const filePaths = _.map(fileNames, fileName => path.resolve(inputDir, fileName));
return findPromise(filePaths, (filePath) => fse.pathExists(filePath));
}
function parseFirstExistingFile(fileNames, inputDir) {
return getFirstExistingFile(fileNames, inputDir).then(filePath => {
if (filePath) {
return parseFile(filePath);
} else {
return null;
}
});
}
async function parseFile(filePath) {
const data = await fse.readFile(filePath, 'utf8');
return parseDataByFilePath(data, filePath);
}
function parseDataByFilePath(string, filePath) {
const extension = path.extname(filePath).substring(1);
let data;
switch (extension) {
case 'yml':
case 'yaml':
data = yaml.safeLoad(string, {schema: yaml.JSON_SCHEMA});
break;
case 'json':
data = JSON.parse(string);
break;
case 'toml':
data = toml.parse(string);
break;
case 'md':
data = parseMarkdownWithFrontMatter(string);
break;
default:
throw new Error(`parseDataByFilePath error, extension '${extension}' of file ${filePath} is not supported`);
}
return data;
}
function parseMarkdownWithFrontMatter(string) | {
string = string.replace('\r\n', '\n');
let frontmatter = null;
let markdown = string;
let frontMatterTypes = [
{
type: 'yaml',
startDelimiter: '---\n',
endDelimiter: '\n---',
parse: (string) => yaml.safeLoad(string, {schema: yaml.JSON_SCHEMA})
},
{
type: 'toml',
startDelimiter: '+++\n',
endDelimiter: '\n+++',
parse: (string) => toml.parse(string)
},
{
type: 'json',
startDelimiter: '{\n', | identifier_body |
|
index.js | (array, callback, thisArg) {
return new Promise((resolve, reject) => {
function next(index) {
if (index < array.length) {
callback.call(thisArg, array[index], index, array).then(() => {
next(index + 1);
}).catch(error => {
reject(error);
});
} else {
resolve();
}
}
next(0);
});
}
function mapPromise(array, callback, thisArg) {
return new Promise((resolve, reject) => {
let results = [];
function next(index) {
if (index < array.length) {
callback.call(thisArg, array[index], index, array).then(result => {
results[index] = result;
next(index + 1);
}).catch(error => {
reject(error);
});
} else {
resolve(results);
}
}
next(0);
});
}
function findPromise(array, callback, thisArg) {
return new Promise((resolve, reject) => {
function next(index) {
if (index < array.length) {
callback.call(thisArg, array[index], index, array).then(result => {
if (result) {
resolve(array[index]);
} else {
next(index + 1);
}
}).catch(error => {
reject(error);
});
} else {
resolve();
}
}
next(0);
});
}
async function readDirRecursively(dir, options) {
const rootDir = _.get(options, 'rootDir', dir);
const files = await fse.readdir(dir);
const result = await mapPromise(files, async (file) => {
const filePath = path.join(dir, file);
const relFilePath = path.relative(rootDir, filePath);
const stats = await fse.stat(filePath);
if (_.has(options, 'filter') && !options.filter(relFilePath, stats)) {
return Promise.resolve();
}
if (stats.isDirectory()) {
return readDirRecursively(filePath, {...options, rootDir});
} else if (stats.isFile()) {
return relFilePath;
} else {
return null;
}
});
return _.chain(result).compact().flatten().value();
}
/**
* Gets the value at the first path of object having non undefined value.
* If all paths resolve to undefined values, the defaultValue is returned.
*
* @param {Object} object The object to query.
* @param {Array<String | Array<String>>} paths The property paths to search for.
* @param {*} [defaultValue] The value returned if all paths resolve to undefined values
* @returns {*}
*/
function getFirst(object, paths, defaultValue) {
let result = _(object).at(paths).reject(_.isUndefined).first();
return _.isUndefined(result) ? defaultValue : result;
}
function append(object, path, value) {
if (!_.has(object, path)) {
_.set(object, path, []);
}
_.get(object, path).push(value);
}
function concat(object, path, value) {
if (!_.has(object, path)) {
_.set(object, path, []);
}
_.set(object, path, _.get(object, path).concat(value));
}
function copy(sourceObject, sourcePath, targetObject, targetPath, transform) {
if (_.has(sourceObject, sourcePath)) {
let value = _.get(sourceObject, sourcePath);
if (transform) {
value = transform(value);
}
_.set(targetObject, targetPath, value);
}
}
function copyIfNotSet(sourceObject, sourcePath, targetObject, targetPath, transform) {
if (!_.has(targetObject, targetPath)) {
copy(sourceObject, sourcePath, targetObject, targetPath, transform);
}
}
function rename(object, oldPath, newPath) {
if (_.has(object, oldPath)) {
_.set(object, newPath, _.get(object, oldPath));
oldPath = _.toPath(oldPath);
if (oldPath.length > 1) {
object = _.get(object, _.initial(oldPath));
}
delete object[_.last(oldPath)];
}
}
function failFunctionWithTag(tag) {
return function fail(message) {
throw new Error(`[${tag}] ${message}`);
};
}
function assertFunctionWithFail(fail) {
return function assert(value, message) {
if (!value) {
fail(message);
}
}
}
/**
* Deeply maps the passed `value` by recursively calling the `iteratee` on
* value's children. The value returned from the iteratee is used to map the
* children.
*
* The iteratee is invoked with three arguments - the `value` being iterated,
* the `fieldPath` of the current `value` relative to the original passed value,
* and the `stack` of ancestors of the current `value`.
*
* For the first time the `iterate` will be called with the original `value`
* and empty arrays for `fieldPath` and `stack`.
*
* In other words, any `value` passed to the iteratee (except the first call,
* and assuming the ancestors keys were not mapped)
* will be equal to: `_.get(originalValue, fieldPath)`
*
* The recursion is called in pre-order depth-first-search. Meaning, the
* iteratee is called first on parent nodes and then on its children. Therefore
* if iteratee maps/replaces the parent node, then the children of the replaced
* node will be traversed.
*
* @example
* mapDeep({ prop: 'foo', arr: [ 'bar' , 1, 2 ] }, (value) => {
* if (_.isString(value)) return '__' + value;
* if (_.isNumber(value)) return value * 10;
* return value;
* })
* => { prop: '__foo', arr: [ '__bar', 10, 20 ] }
*
* mapDeep({ prop: 'foo', arr: [ 'bar' ] }, (value, fieldPath) => {
* if ((_.isString(value)) return value + '__' + fieldPath.join('.');
* return value;
* })
* => { prop: 'foo__prop', arr: [ 'bar__arr.0' ] }
*
* @param {*} value A value to map
* @param {Function} iteratee Function (value: any, fieldPath: Array, stack: Array)
* @param {object} [options]
* @param {boolean} [options.iterateCollections] Default: true
* @param {boolean} [options.iterateScalars] Default: true
* @param {boolean} [options.postOrder] Change the invocation of iteratee to post-order depth-first-search. Default: false
* @param {Array} [_keyPath] For internal recursive use
* @param {Array} [_objectStack] For internal recursive use
* @returns {*}
*/
function mapDeep(value, iteratee, options = {}, _keyPath = [], _objectStack = []) {
const postOrder = _.get(options, 'postOrder', false);
let iterate;
if (_.isPlainObject(value) || _.isArray(value)) {
iterate = _.get(options, 'iterateCollections', true);
} else {
iterate = _.get(options, 'iterateScalars', true);
}
if (iterate && !postOrder) {
value = iteratee(value, _keyPath, _objectStack);
}
const childrenIterator = (val, key) => {
return mapDeep(val, iteratee, options, _.concat(_keyPath, key), _.concat(_objectStack, [value]));
};
if (_.isPlainObject(value)) {
value = _.mapValues(value, childrenIterator);
} else if (_.isArray(value)) {
value = _.map(value, childrenIterator);
}
if (iterate && postOrder) {
value = iteratee(value, _keyPath, _objectStack);
}
return value;
}
function fieldPathToString(fieldPath) {
return _.reduce(fieldPath, (accumulator, fieldName, index) => {
if (_.isString(fieldName) && /\W/.test(fieldName)) {
// field name is a string with non alphanumeric character
accumulator += `['${fieldName}']`;
} else if (_.isNumber(fieldName)) {
accumulator += `[${fieldName}]`;
} else {
if (index > 0) {
accumulator += '.';
}
accumulator += fieldName;
}
return accumulator;
}, '');
}
function getFirstExistingFile(fileNames, inputDir) {
const filePaths = _.map(fileNames, fileName => path.resolve(inputDir, fileName));
return findPromise(filePaths, (filePath) => fse.pathExists(filePath));
}
function parseFirstExistingFile(fileNames, inputDir) {
return getFirstExistingFile(fileNames, inputDir).then(filePath => {
if (filePath) {
return parseFile(filePath);
} else {
return null;
}
});
}
async function parseFile(filePath) {
const data = await fse.readFile(filePath, 'utf8');
return parseDataByFilePath(data, filePath);
}
function parseDataByFilePath(string, filePath) {
const extension = path.extname(filePath).substring(1);
let data;
switch (extension) {
case 'yml':
case 'yaml':
data = yaml.safeLoad(string, {schema: yaml.JSON_SCHEMA});
break;
case 'json':
data = JSON.parse(string);
break;
case 'toml':
data = toml.parse(string);
| forEachPromise | identifier_name |
|
server.go | server.log[i] = NewConsensusState(len(server.BGsInfo), -1, server)
}
i1, err := strconv.Atoi(server.RPCPort)
common.AssertNil(err, "RPC port cannot be recognized.", logger)
tcpPort := common.CalcTCPPortFromRPCPort(i1)
server.tcpManager = NewTCPManager(strconv.Itoa(tcpPort))
server.tcpManager.SetServer(server)
logger.Infof("Server %s start ", server.Address+":"+server.RPCPort)
for _, s := range c["raft_peers"].([]interface{}) {
server.raftPeers = append(server.raftPeers, s.(string))
}
for _, s := range c["rpc_peers"].([]interface{}) {
server.rpcPeers = append(server.rpcPeers, s.(string))
}
server.resultReceiver = NewBftResultReceiver(server.SBFTReplicaAddressForBFTResult)
server.resultVerifier = NewBftResultVerifier(server.SBFTResultVerifierAddress)
server.disseminationLayer.SetServer(server)
server.config = configuration.NewConfiguration(server.ID, configHome)
server.addr = server.Address + ":" + server.RPCPort
// initialize byzantine groups (list of superleafs)
for _, bg := range c["byzantine_groups"].([]interface{}) {
bgInfo := byzantineGroup.NewByzantineGroup(bg)
server.BGsInfo = append(server.BGsInfo, bgInfo)
}
var succ bool
server.SuperLeafID, succ = server.BGsInfo[server.BGID].GetSuperLeafIdByServer(
server.Address + ":" + server.RPCPort)
common.AssertTrue(succ, "Cannot fetch SuperleafID", logger)
logger.Infof("Server %s SuperLeafId %d", server.Address+":"+server.RPCPort, server.SuperLeafID)
//create grpc server
sopts := []grpc.ServerOption{grpc.InitialWindowSize(2000000)}
sopts = append(sopts, grpc.InitialConnWindowSize(2000000))
server.grpcServer = grpc.NewServer(sopts...)
pb.RegisterPolarisAllServer(server.grpcServer, server)
reflection.Register(server.grpcServer)
logger.Info("GRPC server created")
return server
}
func (s *Server) Start() {
logger.Debugf("Server %v start...", s.ID)
// initialize Raft layer
proposeC := make(chan string)
defer close(proposeC)
confChangeC := make(chan raftpb.ConfChange)
defer close(confChangeC)
var raft *Raft
getSnapshot := func() ([]byte, error) { return raft.GetSnapshot() }
commitC, errorC, snapshotterReady, getLeaderID, raftNode, leaderC := raftnode.NewRaftNode(
s.RaftID, s.RaftPort, s.raftPeers, s.join, getSnapshot, proposeC, confChangeC, RaftOutputChannelSize)
s.getLeaderID = getLeaderID
s.raftNode = raftNode
s.raftLayer = NewRaft(s, <-snapshotterReady, proposeC, commitC, errorC)
s.requestManager = NewRequestManager(s,
int(s.config.GetClientRequestCycleTimeout()),
s.config.GetMaxBatchSize(),
s.raftLayer.slRequestChan,
s.raftLayer.txnBodyChan)
// wait for raft leader election
<-leaderC
go s.requestManager.Start()
go s.raftLayer.Start()
go s.disseminationLayer.Start()
if s.IsLeader() {
// create a thread receiving the result from SBFT replica
go s.resultReceiver.Start()
logger.Debugf("leader %v start process sbft result thread", s.ID)
// create a thread handling the result from SBFT replica
go s.processSBFTResult()
}
s.saveLeaderAddress()
s.tcpManager.Start()
tcpAddrList := make([]string, 0)
for _, bg := range s.BGsInfo {
for _, sAddr := range bg.GetServerList() {
items := strings.Split(sAddr, ":")
portNum, _ := strconv.Atoi(items[1])
tPort := common.CalcTCPPortFromRPCPort(portNum)
tAddr := items[0] + ":" + strconv.Itoa(tPort)
s.serverTCPAddrMap[sAddr] = tAddr
tcpAddrList = append(tcpAddrList, tAddr)
}
}
logger.Debugf("tpc server addrs %v", tcpAddrList)
s.tcpManager.DialAll(tcpAddrList)
// Start a RPC Listener
RPCListener, err := net.Listen("tcp", ":"+s.RPCPort)
if err != nil {
logger.Errorf("Failed to listen port %s, %v", s.RPCPort, err)
}
err = s.grpcServer.Serve(RPCListener)
if err != nil {
logger.Errorf("Cannot start to serve RPC calls %v", err)
}
}
// get monitor id (the leader of superleaf(Raft group))
func (s *Server) GetLeader() string {
id := s.getLeaderID()
if id == 0 {
return ""
}
leader := s.rpcPeers[id-1]
go s.BGsInfo[s.BGID].UpdateLeader(leader)
logger.Debugf("server %v get leader %v, raft id %v", s.addr, leader, id)
return leader
}
func (s *Server) IsLeader() bool {
// when the raft leader call getLeaderID, it will return 0
// it only shows in the first time
return int(s.getLeaderID()) == s.RaftID || int(s.getLeaderID()) == 0
}
func (s *Server) IsBGLeader() bool {
if s.currentBGLeaderSuperLeadId != s.SuperLeafID {
return false
}
return s.IsLeader()
}
func (s *Server) saveLeaderAddress() {
fName := fmt.Sprintf("./leader_%v.log", s.ID)
f, err := os.Create(fName)
if err != nil {
logger.Fatalf("create leader log file error %v", err)
}
defer f.Close()
leader := s.GetLeader()
if leader == "" {
leader = s.addr
}
logger.Debugf("save leader id %v", leader)
f.WriteString(leader)
f.Sync()
}
func (s *Server) GetRaftTerm() uint64 {
if s.raftNode == nil {
logger.Error("Raft Node is nil!")
return raft.InvalidRaftTerm
}
return s.raftNode.GetRaftTerm()
}
func (s *Server) GetSBFTClientConnection() *byzantineGroup.Connection {
if s.SBFTClientConnection == nil {
s.SBFTClientConnection = byzantineGroup.NewConnection(s.SBFTClientAddressForBFTLayer, queueLen, 1)
}
return s.SBFTClientConnection
}
func (s *Server) handleRPCError(err error) (*grpcpool.Pool, bool) {
errCode := grpc.Code(err)
if errCode == common.NotLeaderErrorCode {
leaderId := grpc.ErrorDesc(err)
s.BGsInfo[s.BGID].UpdateLeader(leaderId)
conn := s.BGsInfo[s.BGID].GetConnection(leaderId).GetConnectionPool()
return conn, true
}
logger.Warningf("Unhandled ERROR: %v", err)
return nil, false
}
// list of monitors (leaders of superleafs)
func (s *Server) GetBGMembers() []string {
leaders := make([]string, 0)
for i := 0; i < s.BGsInfo[s.BGID].GetSuperLeafNum(); i++ {
l := s.BGsInfo[s.BGID].GetLeader(i)
if l == s.GetServerAddr() {
continue
}
leaders = append(leaders, l)
}
return leaders
}
func (s *Server) GetServerAddr() string {
return s.addr
}
// return consensusState for cycleId
func (s *Server) GetConsensusStateByCycleId(cycleId int) *ConsensusState {
s.logLock.Lock()
defer s.logLock.Unlock()
common.AssertTrue(cycleId >= 0, "cycleID cannot be negative!", logger)
common.AssertTrue(cycleId > s.round2LastFinishedCycle-common.MaxCycleNumberInStorage, "Cannot access stale cycles!", logger)
// cycle may not complete in order
index := cycleId % common.MaxCycleNumberInStorage
if s.log[index].cycleId != cycleId {
// name-removed: Here we overwrite stale records
s.log[index] = NewConsensusState(len(s.BGsInfo), cycleId, s)
}
return s.log[index]
}
// check if the server recieves the state response of bgId for cycleId
// return true: already received
// false: not received
func (s *Server) CheckBGState(cycleId int, bgId int) bool {
logger.Debugf("check bg state response for cycle %v bgId %v", cycleId, bgId)
cs := s.GetConsensusStateByCycleId(cycleId)
op := NewCheckBGStateOp(bgId, cs)
if !cs.AddOPandWait(op) {
logger.Fatal("Error: should return true")
}
logger.Debugf("state response for cycle %v bgId %v is %v",
cycleId, | }
for i := 0; i < common.MaxCycleNumberInStorage; i++ {
// Fill it with -1 | random_line_split |
|
server.go | raft_peers"].([]interface{}) {
server.raftPeers = append(server.raftPeers, s.(string))
}
for _, s := range c["rpc_peers"].([]interface{}) {
server.rpcPeers = append(server.rpcPeers, s.(string))
}
server.resultReceiver = NewBftResultReceiver(server.SBFTReplicaAddressForBFTResult)
server.resultVerifier = NewBftResultVerifier(server.SBFTResultVerifierAddress)
server.disseminationLayer.SetServer(server)
server.config = configuration.NewConfiguration(server.ID, configHome)
server.addr = server.Address + ":" + server.RPCPort
// initialize byzantine groups (list of superleafs)
for _, bg := range c["byzantine_groups"].([]interface{}) {
bgInfo := byzantineGroup.NewByzantineGroup(bg)
server.BGsInfo = append(server.BGsInfo, bgInfo)
}
var succ bool
server.SuperLeafID, succ = server.BGsInfo[server.BGID].GetSuperLeafIdByServer(
server.Address + ":" + server.RPCPort)
common.AssertTrue(succ, "Cannot fetch SuperleafID", logger)
logger.Infof("Server %s SuperLeafId %d", server.Address+":"+server.RPCPort, server.SuperLeafID)
//create grpc server
sopts := []grpc.ServerOption{grpc.InitialWindowSize(2000000)}
sopts = append(sopts, grpc.InitialConnWindowSize(2000000))
server.grpcServer = grpc.NewServer(sopts...)
pb.RegisterPolarisAllServer(server.grpcServer, server)
reflection.Register(server.grpcServer)
logger.Info("GRPC server created")
return server
}
func (s *Server) Start() {
logger.Debugf("Server %v start...", s.ID)
// initialize Raft layer
proposeC := make(chan string)
defer close(proposeC)
confChangeC := make(chan raftpb.ConfChange)
defer close(confChangeC)
var raft *Raft
getSnapshot := func() ([]byte, error) { return raft.GetSnapshot() }
commitC, errorC, snapshotterReady, getLeaderID, raftNode, leaderC := raftnode.NewRaftNode(
s.RaftID, s.RaftPort, s.raftPeers, s.join, getSnapshot, proposeC, confChangeC, RaftOutputChannelSize)
s.getLeaderID = getLeaderID
s.raftNode = raftNode
s.raftLayer = NewRaft(s, <-snapshotterReady, proposeC, commitC, errorC)
s.requestManager = NewRequestManager(s,
int(s.config.GetClientRequestCycleTimeout()),
s.config.GetMaxBatchSize(),
s.raftLayer.slRequestChan,
s.raftLayer.txnBodyChan)
// wait for raft leader election
<-leaderC
go s.requestManager.Start()
go s.raftLayer.Start()
go s.disseminationLayer.Start()
if s.IsLeader() {
// create a thread receiving the result from SBFT replica
go s.resultReceiver.Start()
logger.Debugf("leader %v start process sbft result thread", s.ID)
// create a thread handling the result from SBFT replica
go s.processSBFTResult()
}
s.saveLeaderAddress()
s.tcpManager.Start()
tcpAddrList := make([]string, 0)
for _, bg := range s.BGsInfo {
for _, sAddr := range bg.GetServerList() {
items := strings.Split(sAddr, ":")
portNum, _ := strconv.Atoi(items[1])
tPort := common.CalcTCPPortFromRPCPort(portNum)
tAddr := items[0] + ":" + strconv.Itoa(tPort)
s.serverTCPAddrMap[sAddr] = tAddr
tcpAddrList = append(tcpAddrList, tAddr)
}
}
logger.Debugf("tpc server addrs %v", tcpAddrList)
s.tcpManager.DialAll(tcpAddrList)
// Start a RPC Listener
RPCListener, err := net.Listen("tcp", ":"+s.RPCPort)
if err != nil {
logger.Errorf("Failed to listen port %s, %v", s.RPCPort, err)
}
err = s.grpcServer.Serve(RPCListener)
if err != nil {
logger.Errorf("Cannot start to serve RPC calls %v", err)
}
}
// get monitor id (the leader of superleaf(Raft group))
func (s *Server) GetLeader() string {
id := s.getLeaderID()
if id == 0 {
return ""
}
leader := s.rpcPeers[id-1]
go s.BGsInfo[s.BGID].UpdateLeader(leader)
logger.Debugf("server %v get leader %v, raft id %v", s.addr, leader, id)
return leader
}
func (s *Server) IsLeader() bool {
// when the raft leader call getLeaderID, it will return 0
// it only shows in the first time
return int(s.getLeaderID()) == s.RaftID || int(s.getLeaderID()) == 0
}
func (s *Server) IsBGLeader() bool {
if s.currentBGLeaderSuperLeadId != s.SuperLeafID {
return false
}
return s.IsLeader()
}
func (s *Server) saveLeaderAddress() {
fName := fmt.Sprintf("./leader_%v.log", s.ID)
f, err := os.Create(fName)
if err != nil {
logger.Fatalf("create leader log file error %v", err)
}
defer f.Close()
leader := s.GetLeader()
if leader == "" {
leader = s.addr
}
logger.Debugf("save leader id %v", leader)
f.WriteString(leader)
f.Sync()
}
func (s *Server) GetRaftTerm() uint64 {
if s.raftNode == nil {
logger.Error("Raft Node is nil!")
return raft.InvalidRaftTerm
}
return s.raftNode.GetRaftTerm()
}
func (s *Server) GetSBFTClientConnection() *byzantineGroup.Connection {
if s.SBFTClientConnection == nil {
s.SBFTClientConnection = byzantineGroup.NewConnection(s.SBFTClientAddressForBFTLayer, queueLen, 1)
}
return s.SBFTClientConnection
}
func (s *Server) | (err error) (*grpcpool.Pool, bool) {
errCode := grpc.Code(err)
if errCode == common.NotLeaderErrorCode {
leaderId := grpc.ErrorDesc(err)
s.BGsInfo[s.BGID].UpdateLeader(leaderId)
conn := s.BGsInfo[s.BGID].GetConnection(leaderId).GetConnectionPool()
return conn, true
}
logger.Warningf("Unhandled ERROR: %v", err)
return nil, false
}
// list of monitors (leaders of superleafs)
func (s *Server) GetBGMembers() []string {
leaders := make([]string, 0)
for i := 0; i < s.BGsInfo[s.BGID].GetSuperLeafNum(); i++ {
l := s.BGsInfo[s.BGID].GetLeader(i)
if l == s.GetServerAddr() {
continue
}
leaders = append(leaders, l)
}
return leaders
}
func (s *Server) GetServerAddr() string {
return s.addr
}
// return consensusState for cycleId
func (s *Server) GetConsensusStateByCycleId(cycleId int) *ConsensusState {
s.logLock.Lock()
defer s.logLock.Unlock()
common.AssertTrue(cycleId >= 0, "cycleID cannot be negative!", logger)
common.AssertTrue(cycleId > s.round2LastFinishedCycle-common.MaxCycleNumberInStorage, "Cannot access stale cycles!", logger)
// cycle may not complete in order
index := cycleId % common.MaxCycleNumberInStorage
if s.log[index].cycleId != cycleId {
// name-removed: Here we overwrite stale records
s.log[index] = NewConsensusState(len(s.BGsInfo), cycleId, s)
}
return s.log[index]
}
// check if the server recieves the state response of bgId for cycleId
// return true: already received
// false: not received
func (s *Server) CheckBGState(cycleId int, bgId int) bool {
logger.Debugf("check bg state response for cycle %v bgId %v", cycleId, bgId)
cs := s.GetConsensusStateByCycleId(cycleId)
op := NewCheckBGStateOp(bgId, cs)
if !cs.AddOPandWait(op) {
logger.Fatal("Error: should return true")
}
logger.Debugf("state response for cycle %v bgId %v is %v",
cycleId, bgId, op.GetResult())
return op.GetResult()
}
func (s *Server) CheckStateRequest(sp *pb.StateRequestMessage) bool {
return s.checkRequestQC(sp)
}
func (s *Server) CheckStateResponse(sp *pb.StateResponseMessage) bool {
r := s.checkQC(sp)
if r {
// if QC check ok, then replicate to the Raft group
s.disseminationLayer.AddStateResponse(sp)
}
return r
}
// if the state response is ready, then get the state response
// otherwise return nil and put request to pending list
func (s *Server) GetOrWaitBGState | handleRPCError | identifier_name |
server.go | raft_peers"].([]interface{}) {
server.raftPeers = append(server.raftPeers, s.(string))
}
for _, s := range c["rpc_peers"].([]interface{}) {
server.rpcPeers = append(server.rpcPeers, s.(string))
}
server.resultReceiver = NewBftResultReceiver(server.SBFTReplicaAddressForBFTResult)
server.resultVerifier = NewBftResultVerifier(server.SBFTResultVerifierAddress)
server.disseminationLayer.SetServer(server)
server.config = configuration.NewConfiguration(server.ID, configHome)
server.addr = server.Address + ":" + server.RPCPort
// initialize byzantine groups (list of superleafs)
for _, bg := range c["byzantine_groups"].([]interface{}) {
bgInfo := byzantineGroup.NewByzantineGroup(bg)
server.BGsInfo = append(server.BGsInfo, bgInfo)
}
var succ bool
server.SuperLeafID, succ = server.BGsInfo[server.BGID].GetSuperLeafIdByServer(
server.Address + ":" + server.RPCPort)
common.AssertTrue(succ, "Cannot fetch SuperleafID", logger)
logger.Infof("Server %s SuperLeafId %d", server.Address+":"+server.RPCPort, server.SuperLeafID)
//create grpc server
sopts := []grpc.ServerOption{grpc.InitialWindowSize(2000000)}
sopts = append(sopts, grpc.InitialConnWindowSize(2000000))
server.grpcServer = grpc.NewServer(sopts...)
pb.RegisterPolarisAllServer(server.grpcServer, server)
reflection.Register(server.grpcServer)
logger.Info("GRPC server created")
return server
}
func (s *Server) Start() {
logger.Debugf("Server %v start...", s.ID)
// initialize Raft layer
proposeC := make(chan string)
defer close(proposeC)
confChangeC := make(chan raftpb.ConfChange)
defer close(confChangeC)
var raft *Raft
getSnapshot := func() ([]byte, error) { return raft.GetSnapshot() }
commitC, errorC, snapshotterReady, getLeaderID, raftNode, leaderC := raftnode.NewRaftNode(
s.RaftID, s.RaftPort, s.raftPeers, s.join, getSnapshot, proposeC, confChangeC, RaftOutputChannelSize)
s.getLeaderID = getLeaderID
s.raftNode = raftNode
s.raftLayer = NewRaft(s, <-snapshotterReady, proposeC, commitC, errorC)
s.requestManager = NewRequestManager(s,
int(s.config.GetClientRequestCycleTimeout()),
s.config.GetMaxBatchSize(),
s.raftLayer.slRequestChan,
s.raftLayer.txnBodyChan)
// wait for raft leader election
<-leaderC
go s.requestManager.Start()
go s.raftLayer.Start()
go s.disseminationLayer.Start()
if s.IsLeader() {
// create a thread receiving the result from SBFT replica
go s.resultReceiver.Start()
logger.Debugf("leader %v start process sbft result thread", s.ID)
// create a thread handling the result from SBFT replica
go s.processSBFTResult()
}
s.saveLeaderAddress()
s.tcpManager.Start()
tcpAddrList := make([]string, 0)
for _, bg := range s.BGsInfo {
for _, sAddr := range bg.GetServerList() {
items := strings.Split(sAddr, ":")
portNum, _ := strconv.Atoi(items[1])
tPort := common.CalcTCPPortFromRPCPort(portNum)
tAddr := items[0] + ":" + strconv.Itoa(tPort)
s.serverTCPAddrMap[sAddr] = tAddr
tcpAddrList = append(tcpAddrList, tAddr)
}
}
logger.Debugf("tpc server addrs %v", tcpAddrList)
s.tcpManager.DialAll(tcpAddrList)
// Start a RPC Listener
RPCListener, err := net.Listen("tcp", ":"+s.RPCPort)
if err != nil {
logger.Errorf("Failed to listen port %s, %v", s.RPCPort, err)
}
err = s.grpcServer.Serve(RPCListener)
if err != nil {
logger.Errorf("Cannot start to serve RPC calls %v", err)
}
}
// get monitor id (the leader of superleaf(Raft group))
func (s *Server) GetLeader() string {
id := s.getLeaderID()
if id == 0 {
return ""
}
leader := s.rpcPeers[id-1]
go s.BGsInfo[s.BGID].UpdateLeader(leader)
logger.Debugf("server %v get leader %v, raft id %v", s.addr, leader, id)
return leader
}
func (s *Server) IsLeader() bool |
func (s *Server) IsBGLeader() bool {
if s.currentBGLeaderSuperLeadId != s.SuperLeafID {
return false
}
return s.IsLeader()
}
func (s *Server) saveLeaderAddress() {
fName := fmt.Sprintf("./leader_%v.log", s.ID)
f, err := os.Create(fName)
if err != nil {
logger.Fatalf("create leader log file error %v", err)
}
defer f.Close()
leader := s.GetLeader()
if leader == "" {
leader = s.addr
}
logger.Debugf("save leader id %v", leader)
f.WriteString(leader)
f.Sync()
}
func (s *Server) GetRaftTerm() uint64 {
if s.raftNode == nil {
logger.Error("Raft Node is nil!")
return raft.InvalidRaftTerm
}
return s.raftNode.GetRaftTerm()
}
func (s *Server) GetSBFTClientConnection() *byzantineGroup.Connection {
if s.SBFTClientConnection == nil {
s.SBFTClientConnection = byzantineGroup.NewConnection(s.SBFTClientAddressForBFTLayer, queueLen, 1)
}
return s.SBFTClientConnection
}
func (s *Server) handleRPCError(err error) (*grpcpool.Pool, bool) {
errCode := grpc.Code(err)
if errCode == common.NotLeaderErrorCode {
leaderId := grpc.ErrorDesc(err)
s.BGsInfo[s.BGID].UpdateLeader(leaderId)
conn := s.BGsInfo[s.BGID].GetConnection(leaderId).GetConnectionPool()
return conn, true
}
logger.Warningf("Unhandled ERROR: %v", err)
return nil, false
}
// list of monitors (leaders of superleafs)
func (s *Server) GetBGMembers() []string {
leaders := make([]string, 0)
for i := 0; i < s.BGsInfo[s.BGID].GetSuperLeafNum(); i++ {
l := s.BGsInfo[s.BGID].GetLeader(i)
if l == s.GetServerAddr() {
continue
}
leaders = append(leaders, l)
}
return leaders
}
func (s *Server) GetServerAddr() string {
return s.addr
}
// return consensusState for cycleId
func (s *Server) GetConsensusStateByCycleId(cycleId int) *ConsensusState {
s.logLock.Lock()
defer s.logLock.Unlock()
common.AssertTrue(cycleId >= 0, "cycleID cannot be negative!", logger)
common.AssertTrue(cycleId > s.round2LastFinishedCycle-common.MaxCycleNumberInStorage, "Cannot access stale cycles!", logger)
// cycle may not complete in order
index := cycleId % common.MaxCycleNumberInStorage
if s.log[index].cycleId != cycleId {
// name-removed: Here we overwrite stale records
s.log[index] = NewConsensusState(len(s.BGsInfo), cycleId, s)
}
return s.log[index]
}
// check if the server recieves the state response of bgId for cycleId
// return true: already received
// false: not received
func (s *Server) CheckBGState(cycleId int, bgId int) bool {
logger.Debugf("check bg state response for cycle %v bgId %v", cycleId, bgId)
cs := s.GetConsensusStateByCycleId(cycleId)
op := NewCheckBGStateOp(bgId, cs)
if !cs.AddOPandWait(op) {
logger.Fatal("Error: should return true")
}
logger.Debugf("state response for cycle %v bgId %v is %v",
cycleId, bgId, op.GetResult())
return op.GetResult()
}
func (s *Server) CheckStateRequest(sp *pb.StateRequestMessage) bool {
return s.checkRequestQC(sp)
}
func (s *Server) CheckStateResponse(sp *pb.StateResponseMessage) bool {
r := s.checkQC(sp)
if r {
// if QC check ok, then replicate to the Raft group
s.disseminationLayer.AddStateResponse(sp)
}
return r
}
// if the state response is ready, then get the state response
// otherwise return nil and put request to pending list
func (s *Server) GetOrWaitBG | {
// when the raft leader call getLeaderID, it will return 0
// it only shows in the first time
return int(s.getLeaderID()) == s.RaftID || int(s.getLeaderID()) == 0
} | identifier_body |
server.go | raft_peers"].([]interface{}) {
server.raftPeers = append(server.raftPeers, s.(string))
}
for _, s := range c["rpc_peers"].([]interface{}) {
server.rpcPeers = append(server.rpcPeers, s.(string))
}
server.resultReceiver = NewBftResultReceiver(server.SBFTReplicaAddressForBFTResult)
server.resultVerifier = NewBftResultVerifier(server.SBFTResultVerifierAddress)
server.disseminationLayer.SetServer(server)
server.config = configuration.NewConfiguration(server.ID, configHome)
server.addr = server.Address + ":" + server.RPCPort
// initialize byzantine groups (list of superleafs)
for _, bg := range c["byzantine_groups"].([]interface{}) {
bgInfo := byzantineGroup.NewByzantineGroup(bg)
server.BGsInfo = append(server.BGsInfo, bgInfo)
}
var succ bool
server.SuperLeafID, succ = server.BGsInfo[server.BGID].GetSuperLeafIdByServer(
server.Address + ":" + server.RPCPort)
common.AssertTrue(succ, "Cannot fetch SuperleafID", logger)
logger.Infof("Server %s SuperLeafId %d", server.Address+":"+server.RPCPort, server.SuperLeafID)
//create grpc server
sopts := []grpc.ServerOption{grpc.InitialWindowSize(2000000)}
sopts = append(sopts, grpc.InitialConnWindowSize(2000000))
server.grpcServer = grpc.NewServer(sopts...)
pb.RegisterPolarisAllServer(server.grpcServer, server)
reflection.Register(server.grpcServer)
logger.Info("GRPC server created")
return server
}
func (s *Server) Start() {
logger.Debugf("Server %v start...", s.ID)
// initialize Raft layer
proposeC := make(chan string)
defer close(proposeC)
confChangeC := make(chan raftpb.ConfChange)
defer close(confChangeC)
var raft *Raft
getSnapshot := func() ([]byte, error) { return raft.GetSnapshot() }
commitC, errorC, snapshotterReady, getLeaderID, raftNode, leaderC := raftnode.NewRaftNode(
s.RaftID, s.RaftPort, s.raftPeers, s.join, getSnapshot, proposeC, confChangeC, RaftOutputChannelSize)
s.getLeaderID = getLeaderID
s.raftNode = raftNode
s.raftLayer = NewRaft(s, <-snapshotterReady, proposeC, commitC, errorC)
s.requestManager = NewRequestManager(s,
int(s.config.GetClientRequestCycleTimeout()),
s.config.GetMaxBatchSize(),
s.raftLayer.slRequestChan,
s.raftLayer.txnBodyChan)
// wait for raft leader election
<-leaderC
go s.requestManager.Start()
go s.raftLayer.Start()
go s.disseminationLayer.Start()
if s.IsLeader() {
// create a thread receiving the result from SBFT replica
go s.resultReceiver.Start()
logger.Debugf("leader %v start process sbft result thread", s.ID)
// create a thread handling the result from SBFT replica
go s.processSBFTResult()
}
s.saveLeaderAddress()
s.tcpManager.Start()
tcpAddrList := make([]string, 0)
for _, bg := range s.BGsInfo {
for _, sAddr := range bg.GetServerList() {
items := strings.Split(sAddr, ":")
portNum, _ := strconv.Atoi(items[1])
tPort := common.CalcTCPPortFromRPCPort(portNum)
tAddr := items[0] + ":" + strconv.Itoa(tPort)
s.serverTCPAddrMap[sAddr] = tAddr
tcpAddrList = append(tcpAddrList, tAddr)
}
}
logger.Debugf("tpc server addrs %v", tcpAddrList)
s.tcpManager.DialAll(tcpAddrList)
// Start a RPC Listener
RPCListener, err := net.Listen("tcp", ":"+s.RPCPort)
if err != nil {
logger.Errorf("Failed to listen port %s, %v", s.RPCPort, err)
}
err = s.grpcServer.Serve(RPCListener)
if err != nil {
logger.Errorf("Cannot start to serve RPC calls %v", err)
}
}
// get monitor id (the leader of superleaf(Raft group))
func (s *Server) GetLeader() string {
id := s.getLeaderID()
if id == 0 |
leader := s.rpcPeers[id-1]
go s.BGsInfo[s.BGID].UpdateLeader(leader)
logger.Debugf("server %v get leader %v, raft id %v", s.addr, leader, id)
return leader
}
func (s *Server) IsLeader() bool {
// when the raft leader call getLeaderID, it will return 0
// it only shows in the first time
return int(s.getLeaderID()) == s.RaftID || int(s.getLeaderID()) == 0
}
func (s *Server) IsBGLeader() bool {
if s.currentBGLeaderSuperLeadId != s.SuperLeafID {
return false
}
return s.IsLeader()
}
func (s *Server) saveLeaderAddress() {
fName := fmt.Sprintf("./leader_%v.log", s.ID)
f, err := os.Create(fName)
if err != nil {
logger.Fatalf("create leader log file error %v", err)
}
defer f.Close()
leader := s.GetLeader()
if leader == "" {
leader = s.addr
}
logger.Debugf("save leader id %v", leader)
f.WriteString(leader)
f.Sync()
}
func (s *Server) GetRaftTerm() uint64 {
if s.raftNode == nil {
logger.Error("Raft Node is nil!")
return raft.InvalidRaftTerm
}
return s.raftNode.GetRaftTerm()
}
func (s *Server) GetSBFTClientConnection() *byzantineGroup.Connection {
if s.SBFTClientConnection == nil {
s.SBFTClientConnection = byzantineGroup.NewConnection(s.SBFTClientAddressForBFTLayer, queueLen, 1)
}
return s.SBFTClientConnection
}
func (s *Server) handleRPCError(err error) (*grpcpool.Pool, bool) {
errCode := grpc.Code(err)
if errCode == common.NotLeaderErrorCode {
leaderId := grpc.ErrorDesc(err)
s.BGsInfo[s.BGID].UpdateLeader(leaderId)
conn := s.BGsInfo[s.BGID].GetConnection(leaderId).GetConnectionPool()
return conn, true
}
logger.Warningf("Unhandled ERROR: %v", err)
return nil, false
}
// list of monitors (leaders of superleafs)
func (s *Server) GetBGMembers() []string {
leaders := make([]string, 0)
for i := 0; i < s.BGsInfo[s.BGID].GetSuperLeafNum(); i++ {
l := s.BGsInfo[s.BGID].GetLeader(i)
if l == s.GetServerAddr() {
continue
}
leaders = append(leaders, l)
}
return leaders
}
func (s *Server) GetServerAddr() string {
return s.addr
}
// return consensusState for cycleId
func (s *Server) GetConsensusStateByCycleId(cycleId int) *ConsensusState {
s.logLock.Lock()
defer s.logLock.Unlock()
common.AssertTrue(cycleId >= 0, "cycleID cannot be negative!", logger)
common.AssertTrue(cycleId > s.round2LastFinishedCycle-common.MaxCycleNumberInStorage, "Cannot access stale cycles!", logger)
// cycle may not complete in order
index := cycleId % common.MaxCycleNumberInStorage
if s.log[index].cycleId != cycleId {
// name-removed: Here we overwrite stale records
s.log[index] = NewConsensusState(len(s.BGsInfo), cycleId, s)
}
return s.log[index]
}
// check if the server recieves the state response of bgId for cycleId
// return true: already received
// false: not received
func (s *Server) CheckBGState(cycleId int, bgId int) bool {
logger.Debugf("check bg state response for cycle %v bgId %v", cycleId, bgId)
cs := s.GetConsensusStateByCycleId(cycleId)
op := NewCheckBGStateOp(bgId, cs)
if !cs.AddOPandWait(op) {
logger.Fatal("Error: should return true")
}
logger.Debugf("state response for cycle %v bgId %v is %v",
cycleId, bgId, op.GetResult())
return op.GetResult()
}
func (s *Server) CheckStateRequest(sp *pb.StateRequestMessage) bool {
return s.checkRequestQC(sp)
}
func (s *Server) CheckStateResponse(sp *pb.StateResponseMessage) bool {
r := s.checkQC(sp)
if r {
// if QC check ok, then replicate to the Raft group
s.disseminationLayer.AddStateResponse(sp)
}
return r
}
// if the state response is ready, then get the state response
// otherwise return nil and put request to pending list
func (s *Server) GetOrWaitBG | {
return ""
} | conditional_block |
doom.py | 00
max_steps = 100
batch_size = 64
explore_start = 1.0
explore_stop = .01
decay_rate = 0.0001
gamma = 0.95
pretrain_length = batch_size
memory_size = 1000000
render = True
DQNetwork = None
stack_size = 4
saver = tf.train.Saver()
ap = argparse.ArgumentParser()
ap.add_argument("-m", "--model", required=False, type=bool, default=False,
help="Use the latest saved model")
args = vars(ap.parse_args())
def train(memory):
global stacked_frames
global saver
sess = tf.Session()
sess.run(tf.global_variables_initializer())
decay_step = 0
game.init()
for episode in range(total_episodes):
print("Training on episode: {}".format(episode))
step = 0
episode_rewards = []
game.new_episode()
state = game.get_state().screen_buffer
state, stacked_frames = stack_frames(stacked_frames, state, True)
while step < max_steps:
step += 1
decay_step += 1
action, explore_probability = predict_action(decay_step, possible_actions, state, sess)
reward = game.make_action(action)
done = game.is_episode_finished()
episode_rewards.append(reward)
if done:
# the episode ends so no next state
next_state = np.zeros((84, 84), dtype=np.int)
next_state, stacked_frames = stack_frames(stacked_frames, next_state, False)
# Set step = max_steps to end the episode
step = max_steps
# Get the total reward of the episode
total_reward = np.sum(episode_rewards)
print('Episode: {}'.format(episode),
'Total reward: {}'.format(total_reward),
'Training loss: {:.4f}'.format(loss),
'Explore P: {:.4f}'.format(explore_probability))
memory.add((state, action, reward, next_state, done))
else:
# Get the next state
next_state = game.get_state().screen_buffer
# Stack the frame of the next_state
next_state, stacked_frames = stack_frames(stacked_frames, next_state, False)
# Add experience to memory
memory.add((state, action, reward, next_state, done))
# st+1 is now our current state
state = next_state
### LEARNING PART
# Obtain random mini-batch from memory
batch = memory.sample(batch_size)
states_mb = np.array([each[0] for each in batch])
actions_mb = np.array([each[1] for each in batch])
rewards_mb = np.array([each[2] for each in batch])
next_states_mb = np.array([each[3] for each in batch])
dones_mb = np.array([each[4] for each in batch])
target_Qs_batch = []
# Get Q values for next_state
Qs_next_state = sess.run(DQNetwork.output, feed_dict={DQNetwork.inputs_: next_states_mb})
# Set Q_target = r if the episode ends at s+1, otherwise set Q_target = r + gamma*maxQ(s', a')
for i in range(0, len(batch)):
terminal = dones_mb[i]
# If we are in a terminal state, only equals reward
if terminal:
target_Qs_batch.append(rewards_mb[i])
else:
target = rewards_mb[i] + gamma * np.max(Qs_next_state[i])
target_Qs_batch.append(target)
targets_mb = np.array([each for each in target_Qs_batch])
loss, _ = sess.run([DQNetwork.loss, DQNetwork.optimizer],
feed_dict={DQNetwork.inputs_: states_mb,
DQNetwork.target_Q: targets_mb,
DQNetwork.actions_: actions_mb})
# Write TF Summaries
summary = sess.run(write_op, feed_dict={DQNetwork.inputs_: states_mb,
DQNetwork.target_Q: targets_mb,
DQNetwork.actions_: actions_mb})
writer.add_summary(summary, episode)
writer.flush()
# Save model every 5 episodes
if episode % 5 == 0:
save_path = saver.save(sess, "./models/model.ckpt")
print("Model Saved")
def infer():
global env
global saver
sess = tf.Session()
total_test_rewards = []
game, possible_actions = create_environment()
totalScore = 0
saver.restore(sess, "./models/model.ckpt")
game.init()
for episode in range(1):
total_rewards = 0
state = env.reset()
state, stacked_frames = stack_frames(stacked_frames, state, True)
print("**************************************")
print("EPISODSE ", episode)
while True:
# reshape the state
state = state.reshape((1, *state_size))
Qs = sess.run(DQNetwork.output, feed_dict={DQNetwork.inputs_: state})
# Take the biggest Q value
choice = np.argmax(Qs)
action = possible_actions[choice]
next_state, reward, done, _ = env.step(action)
env.render()
total_rewards += reward
if done:
print("Score", total_rewards)
total_test_rewards.append(total_rewards)
break
next_state, stacked_frames = stack_frames(stacked_frames, next_state, False)
state = next_state
env.close()
def predict_action(decay_step, possible_actions, state, sess):
global explore_start
global explore_stop
global decay_rate
exp_exp_tradeoff = np.random.rand()
explore_probability = explore_stop + (explore_start - explore_stop) * np.exp(-decay_rate * decay_step)
if explore_probability > exp_exp_tradeoff:
action = random.choice(possible_actions)
else:
Qs = sess.run(DQNetwork.output, feed_dict={DQNetwork.inputs_: state.reshape((1, *state.shape))})
choice = np.argmax(Qs)
action = possible_actions[int(choice)]
return action, explore_probability
def createnetwork():
global DQNetwork
global state_size
global action_size
global learning_rate
tf.reset_default_graph()
DQNetwork = dqn2.DQNetwork(state_size, action_size, learning_rate)
def preprocess_frame(frame):
cropped_frame = frame[30:-10, 30: -30]
normalized_frame = cropped_frame/255.0
preprocessed_frame = transform.resize(normalized_frame, [84, 84])
return preprocessed_frame
def stack_frames(frames2stack, state, is_new_episode):
frame = preprocess_frame(state)
if is_new_episode:
# Clear the stacked frames
frames2stack = deque([np.zeros((84, 84), dtype=np.int) for i in range(stack_size)], maxlen=4)
# this is new episode so copy the first and only frame 4 times
frames2stack.append(frame)
frames2stack.append(frame)
frames2stack.append(frame)
frames2stack.append(frame)
# stack the frames
stacked_state = np.stack(frames2stack, axis=2)
else:
# Append frame to deque
frames2stack.append(frame)
stacked_state = np.stack(frames2stack, axis=2)
return stacked_state, frames2stack
def create_environment():
global game
global state_size
global action_size
game = DoomGame()
game.load_config("basic.cfg")
game.set_doom_scenario_path("basic.wad")
game.init()
state_size = [84, 84, 4] # 4 Frames at 84x84 resolution
# Possible actions
left = [1, 0, 0]
right = [0, 1, 0]
shoot = [0, 0, 1]
possible_actions = [left, right, shoot]
action_size = 3
return game, possible_actions
def test_environment():
|
if not args['model']:
stacked_frames = deque([np.zeros((84, 84), dtype=np.int) for i in range(stack_size)], maxlen=4)
game, possible_actions = create_environment()
print("Game Environment created")
createnetwork()
print("network created")
memory = mem.Memory(max_size=memory_size)
game.new_episode()
for i in range(pretrain_length):
if i is 0:
state = game.get_state().screen_buffer
state, stacked_frames = stack_frames(stacked_frames, state, True)
# Take a random action
action = random.choice(possible_actions)
# rewards
reward = game.make_action(action)
done = | global game
episodes = 10
for i in range(episodes):
game.new_episode()
while not game.is_episode_finished():
state = game.get_state()
img = state.screen_buffer
misc = state.game_variables
action = random.choice(possible_actions)
print("Taking action: {}".format(action))
reward = game.make_action(action)
print("Reward for above action: {}".format(reward))
time.sleep(.2)
print("Episode final reward score: {}".format(game.get_total_reward()))
time.sleep(2)
game.close() | identifier_body |
doom.py | 500
max_steps = 100
batch_size = 64
explore_start = 1.0
explore_stop = .01
decay_rate = 0.0001
gamma = 0.95
pretrain_length = batch_size
memory_size = 1000000
render = True
DQNetwork = None
stack_size = 4
saver = tf.train.Saver()
ap = argparse.ArgumentParser()
ap.add_argument("-m", "--model", required=False, type=bool, default=False,
help="Use the latest saved model")
args = vars(ap.parse_args())
def train(memory):
global stacked_frames
global saver
sess = tf.Session()
sess.run(tf.global_variables_initializer())
decay_step = 0
game.init()
for episode in range(total_episodes):
print("Training on episode: {}".format(episode))
step = 0
episode_rewards = []
game.new_episode()
state = game.get_state().screen_buffer
state, stacked_frames = stack_frames(stacked_frames, state, True)
while step < max_steps:
step += 1
decay_step += 1
action, explore_probability = predict_action(decay_step, possible_actions, state, sess)
reward = game.make_action(action)
done = game.is_episode_finished()
episode_rewards.append(reward)
if done:
# the episode ends so no next state
next_state = np.zeros((84, 84), dtype=np.int)
next_state, stacked_frames = stack_frames(stacked_frames, next_state, False)
# Set step = max_steps to end the episode
step = max_steps
# Get the total reward of the episode
total_reward = np.sum(episode_rewards)
print('Episode: {}'.format(episode),
'Total reward: {}'.format(total_reward),
'Training loss: {:.4f}'.format(loss),
'Explore P: {:.4f}'.format(explore_probability))
memory.add((state, action, reward, next_state, done))
else:
# Get the next state
next_state = game.get_state().screen_buffer
# Stack the frame of the next_state
next_state, stacked_frames = stack_frames(stacked_frames, next_state, False)
# Add experience to memory
memory.add((state, action, reward, next_state, done))
# st+1 is now our current state
state = next_state
### LEARNING PART
# Obtain random mini-batch from memory
batch = memory.sample(batch_size)
states_mb = np.array([each[0] for each in batch])
actions_mb = np.array([each[1] for each in batch])
rewards_mb = np.array([each[2] for each in batch])
next_states_mb = np.array([each[3] for each in batch])
dones_mb = np.array([each[4] for each in batch])
target_Qs_batch = []
# Get Q values for next_state
Qs_next_state = sess.run(DQNetwork.output, feed_dict={DQNetwork.inputs_: next_states_mb})
# Set Q_target = r if the episode ends at s+1, otherwise set Q_target = r + gamma*maxQ(s', a')
for i in range(0, len(batch)):
terminal = dones_mb[i]
# If we are in a terminal state, only equals reward
if terminal:
target_Qs_batch.append(rewards_mb[i])
else:
target = rewards_mb[i] + gamma * np.max(Qs_next_state[i])
target_Qs_batch.append(target)
targets_mb = np.array([each for each in target_Qs_batch])
loss, _ = sess.run([DQNetwork.loss, DQNetwork.optimizer],
feed_dict={DQNetwork.inputs_: states_mb,
DQNetwork.target_Q: targets_mb,
DQNetwork.actions_: actions_mb})
# Write TF Summaries
summary = sess.run(write_op, feed_dict={DQNetwork.inputs_: states_mb,
DQNetwork.target_Q: targets_mb,
DQNetwork.actions_: actions_mb})
writer.add_summary(summary, episode)
writer.flush()
# Save model every 5 episodes
if episode % 5 == 0:
save_path = saver.save(sess, "./models/model.ckpt")
print("Model Saved")
def infer():
global env
global saver
sess = tf.Session()
total_test_rewards = []
game, possible_actions = create_environment()
totalScore = 0
saver.restore(sess, "./models/model.ckpt")
game.init()
for episode in range(1):
total_rewards = 0
state = env.reset()
state, stacked_frames = stack_frames(stacked_frames, state, True)
print("**************************************")
print("EPISODSE ", episode)
while True:
# reshape the state
state = state.reshape((1, *state_size))
Qs = sess.run(DQNetwork.output, feed_dict={DQNetwork.inputs_: state})
# Take the biggest Q value
choice = np.argmax(Qs)
action = possible_actions[choice]
next_state, reward, done, _ = env.step(action)
env.render()
total_rewards += reward
if done:
|
next_state, stacked_frames = stack_frames(stacked_frames, next_state, False)
state = next_state
env.close()
def predict_action(decay_step, possible_actions, state, sess):
global explore_start
global explore_stop
global decay_rate
exp_exp_tradeoff = np.random.rand()
explore_probability = explore_stop + (explore_start - explore_stop) * np.exp(-decay_rate * decay_step)
if explore_probability > exp_exp_tradeoff:
action = random.choice(possible_actions)
else:
Qs = sess.run(DQNetwork.output, feed_dict={DQNetwork.inputs_: state.reshape((1, *state.shape))})
choice = np.argmax(Qs)
action = possible_actions[int(choice)]
return action, explore_probability
def createnetwork():
global DQNetwork
global state_size
global action_size
global learning_rate
tf.reset_default_graph()
DQNetwork = dqn2.DQNetwork(state_size, action_size, learning_rate)
def preprocess_frame(frame):
cropped_frame = frame[30:-10, 30: -30]
normalized_frame = cropped_frame/255.0
preprocessed_frame = transform.resize(normalized_frame, [84, 84])
return preprocessed_frame
def stack_frames(frames2stack, state, is_new_episode):
frame = preprocess_frame(state)
if is_new_episode:
# Clear the stacked frames
frames2stack = deque([np.zeros((84, 84), dtype=np.int) for i in range(stack_size)], maxlen=4)
# this is new episode so copy the first and only frame 4 times
frames2stack.append(frame)
frames2stack.append(frame)
frames2stack.append(frame)
frames2stack.append(frame)
# stack the frames
stacked_state = np.stack(frames2stack, axis=2)
else:
# Append frame to deque
frames2stack.append(frame)
stacked_state = np.stack(frames2stack, axis=2)
return stacked_state, frames2stack
def create_environment():
global game
global state_size
global action_size
game = DoomGame()
game.load_config("basic.cfg")
game.set_doom_scenario_path("basic.wad")
game.init()
state_size = [84, 84, 4] # 4 Frames at 84x84 resolution
# Possible actions
left = [1, 0, 0]
right = [0, 1, 0]
shoot = [0, 0, 1]
possible_actions = [left, right, shoot]
action_size = 3
return game, possible_actions
def test_environment():
global game
episodes = 10
for i in range(episodes):
game.new_episode()
while not game.is_episode_finished():
state = game.get_state()
img = state.screen_buffer
misc = state.game_variables
action = random.choice(possible_actions)
print("Taking action: {}".format(action))
reward = game.make_action(action)
print("Reward for above action: {}".format(reward))
time.sleep(.2)
print("Episode final reward score: {}".format(game.get_total_reward()))
time.sleep(2)
game.close()
if not args['model']:
stacked_frames = deque([np.zeros((84, 84), dtype=np.int) for i in range(stack_size)], maxlen=4)
game, possible_actions = create_environment()
print("Game Environment created")
createnetwork()
print("network created")
memory = mem.Memory(max_size=memory_size)
game.new_episode()
for i in range(pretrain_length):
if i is 0:
state = game.get_state().screen_buffer
state, stacked_frames = stack_frames(stacked_frames, state, True)
# Take a random action
action = random.choice(possible_actions)
# rewards
reward = game.make_action(action)
done = | print("Score", total_rewards)
total_test_rewards.append(total_rewards)
break | conditional_block |
doom.py | 00
max_steps = 100
batch_size = 64
explore_start = 1.0
explore_stop = .01
decay_rate = 0.0001
gamma = 0.95
pretrain_length = batch_size
memory_size = 1000000
render = True
DQNetwork = None
stack_size = 4
saver = tf.train.Saver()
ap = argparse.ArgumentParser()
ap.add_argument("-m", "--model", required=False, type=bool, default=False,
help="Use the latest saved model")
args = vars(ap.parse_args())
def | (memory):
global stacked_frames
global saver
sess = tf.Session()
sess.run(tf.global_variables_initializer())
decay_step = 0
game.init()
for episode in range(total_episodes):
print("Training on episode: {}".format(episode))
step = 0
episode_rewards = []
game.new_episode()
state = game.get_state().screen_buffer
state, stacked_frames = stack_frames(stacked_frames, state, True)
while step < max_steps:
step += 1
decay_step += 1
action, explore_probability = predict_action(decay_step, possible_actions, state, sess)
reward = game.make_action(action)
done = game.is_episode_finished()
episode_rewards.append(reward)
if done:
# the episode ends so no next state
next_state = np.zeros((84, 84), dtype=np.int)
next_state, stacked_frames = stack_frames(stacked_frames, next_state, False)
# Set step = max_steps to end the episode
step = max_steps
# Get the total reward of the episode
total_reward = np.sum(episode_rewards)
print('Episode: {}'.format(episode),
'Total reward: {}'.format(total_reward),
'Training loss: {:.4f}'.format(loss),
'Explore P: {:.4f}'.format(explore_probability))
memory.add((state, action, reward, next_state, done))
else:
# Get the next state
next_state = game.get_state().screen_buffer
# Stack the frame of the next_state
next_state, stacked_frames = stack_frames(stacked_frames, next_state, False)
# Add experience to memory
memory.add((state, action, reward, next_state, done))
# st+1 is now our current state
state = next_state
### LEARNING PART
# Obtain random mini-batch from memory
batch = memory.sample(batch_size)
states_mb = np.array([each[0] for each in batch])
actions_mb = np.array([each[1] for each in batch])
rewards_mb = np.array([each[2] for each in batch])
next_states_mb = np.array([each[3] for each in batch])
dones_mb = np.array([each[4] for each in batch])
target_Qs_batch = []
# Get Q values for next_state
Qs_next_state = sess.run(DQNetwork.output, feed_dict={DQNetwork.inputs_: next_states_mb})
# Set Q_target = r if the episode ends at s+1, otherwise set Q_target = r + gamma*maxQ(s', a')
for i in range(0, len(batch)):
terminal = dones_mb[i]
# If we are in a terminal state, only equals reward
if terminal:
target_Qs_batch.append(rewards_mb[i])
else:
target = rewards_mb[i] + gamma * np.max(Qs_next_state[i])
target_Qs_batch.append(target)
targets_mb = np.array([each for each in target_Qs_batch])
loss, _ = sess.run([DQNetwork.loss, DQNetwork.optimizer],
feed_dict={DQNetwork.inputs_: states_mb,
DQNetwork.target_Q: targets_mb,
DQNetwork.actions_: actions_mb})
# Write TF Summaries
summary = sess.run(write_op, feed_dict={DQNetwork.inputs_: states_mb,
DQNetwork.target_Q: targets_mb,
DQNetwork.actions_: actions_mb})
writer.add_summary(summary, episode)
writer.flush()
# Save model every 5 episodes
if episode % 5 == 0:
save_path = saver.save(sess, "./models/model.ckpt")
print("Model Saved")
def infer():
global env
global saver
sess = tf.Session()
total_test_rewards = []
game, possible_actions = create_environment()
totalScore = 0
saver.restore(sess, "./models/model.ckpt")
game.init()
for episode in range(1):
total_rewards = 0
state = env.reset()
state, stacked_frames = stack_frames(stacked_frames, state, True)
print("**************************************")
print("EPISODSE ", episode)
while True:
# reshape the state
state = state.reshape((1, *state_size))
Qs = sess.run(DQNetwork.output, feed_dict={DQNetwork.inputs_: state})
# Take the biggest Q value
choice = np.argmax(Qs)
action = possible_actions[choice]
next_state, reward, done, _ = env.step(action)
env.render()
total_rewards += reward
if done:
print("Score", total_rewards)
total_test_rewards.append(total_rewards)
break
next_state, stacked_frames = stack_frames(stacked_frames, next_state, False)
state = next_state
env.close()
def predict_action(decay_step, possible_actions, state, sess):
global explore_start
global explore_stop
global decay_rate
exp_exp_tradeoff = np.random.rand()
explore_probability = explore_stop + (explore_start - explore_stop) * np.exp(-decay_rate * decay_step)
if explore_probability > exp_exp_tradeoff:
action = random.choice(possible_actions)
else:
Qs = sess.run(DQNetwork.output, feed_dict={DQNetwork.inputs_: state.reshape((1, *state.shape))})
choice = np.argmax(Qs)
action = possible_actions[int(choice)]
return action, explore_probability
def createnetwork():
global DQNetwork
global state_size
global action_size
global learning_rate
tf.reset_default_graph()
DQNetwork = dqn2.DQNetwork(state_size, action_size, learning_rate)
def preprocess_frame(frame):
cropped_frame = frame[30:-10, 30: -30]
normalized_frame = cropped_frame/255.0
preprocessed_frame = transform.resize(normalized_frame, [84, 84])
return preprocessed_frame
def stack_frames(frames2stack, state, is_new_episode):
frame = preprocess_frame(state)
if is_new_episode:
# Clear the stacked frames
frames2stack = deque([np.zeros((84, 84), dtype=np.int) for i in range(stack_size)], maxlen=4)
# this is new episode so copy the first and only frame 4 times
frames2stack.append(frame)
frames2stack.append(frame)
frames2stack.append(frame)
frames2stack.append(frame)
# stack the frames
stacked_state = np.stack(frames2stack, axis=2)
else:
# Append frame to deque
frames2stack.append(frame)
stacked_state = np.stack(frames2stack, axis=2)
return stacked_state, frames2stack
def create_environment():
global game
global state_size
global action_size
game = DoomGame()
game.load_config("basic.cfg")
game.set_doom_scenario_path("basic.wad")
game.init()
state_size = [84, 84, 4] # 4 Frames at 84x84 resolution
# Possible actions
left = [1, 0, 0]
right = [0, 1, 0]
shoot = [0, 0, 1]
possible_actions = [left, right, shoot]
action_size = 3
return game, possible_actions
def test_environment():
global game
episodes = 10
for i in range(episodes):
game.new_episode()
while not game.is_episode_finished():
state = game.get_state()
img = state.screen_buffer
misc = state.game_variables
action = random.choice(possible_actions)
print("Taking action: {}".format(action))
reward = game.make_action(action)
print("Reward for above action: {}".format(reward))
time.sleep(.2)
print("Episode final reward score: {}".format(game.get_total_reward()))
time.sleep(2)
game.close()
if not args['model']:
stacked_frames = deque([np.zeros((84, 84), dtype=np.int) for i in range(stack_size)], maxlen=4)
game, possible_actions = create_environment()
print("Game Environment created")
createnetwork()
print("network created")
memory = mem.Memory(max_size=memory_size)
game.new_episode()
for i in range(pretrain_length):
if i is 0:
state = game.get_state().screen_buffer
state, stacked_frames = stack_frames(stacked_frames, state, True)
# Take a random action
action = random.choice(possible_actions)
# rewards
reward = game.make_action(action)
done = | train | identifier_name |
doom.py | 500
max_steps = 100
batch_size = 64
explore_start = 1.0
explore_stop = .01
decay_rate = 0.0001
gamma = 0.95
pretrain_length = batch_size
memory_size = 1000000
render = True
DQNetwork = None
stack_size = 4
saver = tf.train.Saver()
ap = argparse.ArgumentParser()
ap.add_argument("-m", "--model", required=False, type=bool, default=False,
help="Use the latest saved model")
args = vars(ap.parse_args())
def train(memory):
global stacked_frames
global saver
sess = tf.Session()
sess.run(tf.global_variables_initializer())
decay_step = 0
game.init()
for episode in range(total_episodes):
print("Training on episode: {}".format(episode))
step = 0
episode_rewards = []
game.new_episode()
state = game.get_state().screen_buffer
state, stacked_frames = stack_frames(stacked_frames, state, True)
while step < max_steps:
step += 1
decay_step += 1
action, explore_probability = predict_action(decay_step, possible_actions, state, sess)
reward = game.make_action(action)
done = game.is_episode_finished()
episode_rewards.append(reward)
if done:
# the episode ends so no next state
next_state = np.zeros((84, 84), dtype=np.int)
next_state, stacked_frames = stack_frames(stacked_frames, next_state, False)
# Set step = max_steps to end the episode
step = max_steps
# Get the total reward of the episode
total_reward = np.sum(episode_rewards)
print('Episode: {}'.format(episode),
'Total reward: {}'.format(total_reward),
'Training loss: {:.4f}'.format(loss),
'Explore P: {:.4f}'.format(explore_probability))
memory.add((state, action, reward, next_state, done))
else:
# Get the next state
next_state = game.get_state().screen_buffer
# Stack the frame of the next_state
next_state, stacked_frames = stack_frames(stacked_frames, next_state, False)
# Add experience to memory
memory.add((state, action, reward, next_state, done))
# st+1 is now our current state
state = next_state
### LEARNING PART
# Obtain random mini-batch from memory
batch = memory.sample(batch_size)
states_mb = np.array([each[0] for each in batch])
actions_mb = np.array([each[1] for each in batch])
rewards_mb = np.array([each[2] for each in batch])
next_states_mb = np.array([each[3] for each in batch])
dones_mb = np.array([each[4] for each in batch])
target_Qs_batch = []
# Get Q values for next_state
Qs_next_state = sess.run(DQNetwork.output, feed_dict={DQNetwork.inputs_: next_states_mb})
# Set Q_target = r if the episode ends at s+1, otherwise set Q_target = r + gamma*maxQ(s', a')
for i in range(0, len(batch)):
terminal = dones_mb[i]
# If we are in a terminal state, only equals reward
if terminal:
target_Qs_batch.append(rewards_mb[i])
else:
target = rewards_mb[i] + gamma * np.max(Qs_next_state[i])
target_Qs_batch.append(target)
targets_mb = np.array([each for each in target_Qs_batch]) | loss, _ = sess.run([DQNetwork.loss, DQNetwork.optimizer],
feed_dict={DQNetwork.inputs_: states_mb,
DQNetwork.target_Q: targets_mb,
DQNetwork.actions_: actions_mb})
# Write TF Summaries
summary = sess.run(write_op, feed_dict={DQNetwork.inputs_: states_mb,
DQNetwork.target_Q: targets_mb,
DQNetwork.actions_: actions_mb})
writer.add_summary(summary, episode)
writer.flush()
# Save model every 5 episodes
if episode % 5 == 0:
save_path = saver.save(sess, "./models/model.ckpt")
print("Model Saved")
def infer():
global env
global saver
sess = tf.Session()
total_test_rewards = []
game, possible_actions = create_environment()
totalScore = 0
saver.restore(sess, "./models/model.ckpt")
game.init()
for episode in range(1):
total_rewards = 0
state = env.reset()
state, stacked_frames = stack_frames(stacked_frames, state, True)
print("**************************************")
print("EPISODSE ", episode)
while True:
# reshape the state
state = state.reshape((1, *state_size))
Qs = sess.run(DQNetwork.output, feed_dict={DQNetwork.inputs_: state})
# Take the biggest Q value
choice = np.argmax(Qs)
action = possible_actions[choice]
next_state, reward, done, _ = env.step(action)
env.render()
total_rewards += reward
if done:
print("Score", total_rewards)
total_test_rewards.append(total_rewards)
break
next_state, stacked_frames = stack_frames(stacked_frames, next_state, False)
state = next_state
env.close()
def predict_action(decay_step, possible_actions, state, sess):
global explore_start
global explore_stop
global decay_rate
exp_exp_tradeoff = np.random.rand()
explore_probability = explore_stop + (explore_start - explore_stop) * np.exp(-decay_rate * decay_step)
if explore_probability > exp_exp_tradeoff:
action = random.choice(possible_actions)
else:
Qs = sess.run(DQNetwork.output, feed_dict={DQNetwork.inputs_: state.reshape((1, *state.shape))})
choice = np.argmax(Qs)
action = possible_actions[int(choice)]
return action, explore_probability
def createnetwork():
global DQNetwork
global state_size
global action_size
global learning_rate
tf.reset_default_graph()
DQNetwork = dqn2.DQNetwork(state_size, action_size, learning_rate)
def preprocess_frame(frame):
cropped_frame = frame[30:-10, 30: -30]
normalized_frame = cropped_frame/255.0
preprocessed_frame = transform.resize(normalized_frame, [84, 84])
return preprocessed_frame
def stack_frames(frames2stack, state, is_new_episode):
frame = preprocess_frame(state)
if is_new_episode:
# Clear the stacked frames
frames2stack = deque([np.zeros((84, 84), dtype=np.int) for i in range(stack_size)], maxlen=4)
# this is new episode so copy the first and only frame 4 times
frames2stack.append(frame)
frames2stack.append(frame)
frames2stack.append(frame)
frames2stack.append(frame)
# stack the frames
stacked_state = np.stack(frames2stack, axis=2)
else:
# Append frame to deque
frames2stack.append(frame)
stacked_state = np.stack(frames2stack, axis=2)
return stacked_state, frames2stack
def create_environment():
global game
global state_size
global action_size
game = DoomGame()
game.load_config("basic.cfg")
game.set_doom_scenario_path("basic.wad")
game.init()
state_size = [84, 84, 4] # 4 Frames at 84x84 resolution
# Possible actions
left = [1, 0, 0]
right = [0, 1, 0]
shoot = [0, 0, 1]
possible_actions = [left, right, shoot]
action_size = 3
return game, possible_actions
def test_environment():
global game
episodes = 10
for i in range(episodes):
game.new_episode()
while not game.is_episode_finished():
state = game.get_state()
img = state.screen_buffer
misc = state.game_variables
action = random.choice(possible_actions)
print("Taking action: {}".format(action))
reward = game.make_action(action)
print("Reward for above action: {}".format(reward))
time.sleep(.2)
print("Episode final reward score: {}".format(game.get_total_reward()))
time.sleep(2)
game.close()
if not args['model']:
stacked_frames = deque([np.zeros((84, 84), dtype=np.int) for i in range(stack_size)], maxlen=4)
game, possible_actions = create_environment()
print("Game Environment created")
createnetwork()
print("network created")
memory = mem.Memory(max_size=memory_size)
game.new_episode()
for i in range(pretrain_length):
if i is 0:
state = game.get_state().screen_buffer
state, stacked_frames = stack_frames(stacked_frames, state, True)
# Take a random action
action = random.choice(possible_actions)
# rewards
reward = game.make_action(action)
done = | random_line_split |
|
dataset3.js | 418531', y: '3.17806176', label: 'C1' },
{ x: '4.28296762', y: '9.39110461', label: 'C2' },
{ x: '0.54567095', y: '1.74704601', label: 'C2' },
{ x: '6.47297586', y: '80.0737218', label: 'C2' },
{ x: '3.56196783', y: '3.46501229', label: 'C1' },
{ x: '4.92110965', y: '3.21782503', label: 'C1' },
{ x: '7.08960234', y: '3.31001019', label: 'C1' },
{ x: '6.6400647', y: '11.4021873', label: 'C2' },
{ x: '3.04376354', y: '3.11156957', label: 'C1' },
{ x: '3.29279458', y: '3.20280598', label: 'C1' },
{ x: '11.1281167', y: '3.35227009', label: 'C1' },
{ x: '0.62532426', y: '2.6768092', label: 'C1' },
{ x: '14.1219825', y: '3.47557778', label: 'C1' },
{ x: '5.97399823', y: '7.92169457', label: 'C2' },
{ x: '11.4297922', y: '21.5417738', label: 'C2' },
{ x: '14.5248268', y: '27.7087372', label: 'C2' },
{ x: '3.09732353', y: '3.26640523', label: 'C1' },
{ x: '9.90401929', y: '62.2373692', label: 'C2' },
{ x: '8.49818415', y: '69.0314842', label: 'C2' },
{ x: '4.60844752', y: '3.15543562', label: 'C1' },
{ x: '14.7514267', y: '3.45141488', label: 'C1' },
{ x: '10.5700858', y: '3.42392187', label: 'C1' },
{ x: '11.8527787', y: '3.38192599', label: 'C1' }, | { x: '4.06277657', y: '3.30054593', label: 'C1' },
{ x: '10.1004059', y: '73.8328052', label: 'C2' },
{ x: '11.9150365', y: '3.37543243', label: 'C1' },
{ x: '2.67113254', y: '5.12725717', label: 'C2' },
{ x: '9.21185949', y: '14.5993617', label: 'C2' },
{ x: '2.59559923', y: '5.1530681', label: 'C2' },
{ x: '11.2512589', y: '34.1527354', label: 'C2' },
{ x: '3.46217231', y: '4.64449256', label: 'C2' },
{ x: '10.1713614', y: '287.740405', label: 'C2' },
{ x: '12.7060762', y: '18.9983532', label: 'C2' },
{ x: '0.24033326', y: '2.65936106', label: 'C1' },
{ x: '7.21594897', y: '3.2822869', label: 'C1' },
{ x: '11.1487167', y: '16.5251589', label: 'C2' },
{ x: '11.31077', y: '36.740755', label: 'C2' },
{ x: '8.7980285', y: '150.739092', label: 'C2' },
{ x: '12.3128452', y: '3.46424609', label: 'C1' },
{ x: '14.3563646', y: '27.2356295', label: 'C2' },
{ x: '7.55607776', y: '3.27713683', label: 'C1' },
{ x: '1.16779077', y: '2.59038073', label: 'C1' },
{ x: '1.15726188', y: '2.18638168', label: 'C2' },
{ x: '4.74303415', y: '3.28008028', label: 'C1' },
{ x: '6.85750908', y: '11.8555794', label: 'C2' },
{ x: '4.57319864', y: '3.14574577', label: 'C1' },
{ x: '10.458388', y: '3.45509926', label: 'C1' },
{ x: '7.18939787', y: '3.35470766', label: 'C1' },
{ x: '7.59407331', y: '14.5317836', label: 'C2' },
{ x: '1.32847072', y: '2.86762537', label: 'C2' },
{ x: '4.82451857', y: '3.23219081', label: 'C1' },
{ x: '11.2828455', y: '58.5434514', label: ' | { x: '0.51454207', y: '2.55030179', label: 'C1' },
{ x: '0.1780755', y: '1.52949356', label: 'C2' },
{ x: '0.76631977', y: '2.5589982', label: 'C1' },
{ x: '14.2249825', y: '19.6862421', label: 'C2' }, | random_line_split |
covid19co_pipe.py | -Social/Casos-positivos-de-COVID-19-en-Colombia/gt2j-8ykr)
#
# The number of new cases are increasing day by day around the world.
# This dataset has information about reported cases from 32 Colombia departments.
#
# Also you can get the dataset Google COVID-19 Community Mobility Reports - Colombia.
#
# You can view and collaborate to the analysis here:
# [colombia_covid_19_analysis](https://www.kaggle.com/sebaxtian/colombia-covid-19-analysis) Kaggle Notebook Kernel.
# %% [markdown]
# ---
# %% [markdown]
# ## Data Sources
# %%
# Input data files are available in the "../input/" directory.
INPUT_DIR = './'
if os.path.split(os.path.abspath('.'))[-1] == 'src':
INPUT_DIR = '../input'
# Output data files are available in the "../output/" directory.
OUTPUT_DIR = './'
if os.path.split(os.path.abspath('.'))[-1] == 'src':
OUTPUT_DIR = '../output'
# Official Daily Report Until Now
URL_OFFICIAL_DATASET = 'https://www.datos.gov.co/api/views/gt2j-8ykr/rows.csv?accessType=DOWNLOAD'
# Official Daily Samples Processed
URL_SAMPLES_PROCESSED = 'https://www.datos.gov.co/api/views/8835-5baf/rows.csv?accessType=DOWNLOAD'
# %% [markdown]
# ---
# %% [markdown]
# ## Official Covid19 Colombia Daily Report
# %%
# Official Daily Report Until Now
with requests.get(URL_OFFICIAL_DATASET) as official_dataset:
with open(os.path.join(INPUT_DIR, 'covid19co_official.csv'), 'wb') as dataset_file:
dataset_file.write(official_dataset.content)
# %%
# Open Official Daily Report
covid19co = pd.read_csv(os.path.join(INPUT_DIR, 'covid19co_official.csv'))
# Total Daily Report
covid19co.shape
# %%
# Show dataframe
covid19co.tail()
# %%
# Show attributes
list(covid19co.columns.values)
# %%
# Update Name Columns
# Remove Accents and Uppercase
covid19co.columns = [unidecode.unidecode(value).upper() for value in covid19co.columns]
# Show dataframe
covid19co.head()
| covid19co[attr] = covid19co[attr].transform(lambda value: str(value).title())
# Show dataframe
covid19co.head()
# %%
# Fill NaN Values
if covid19co.isna().sum().sum() > 0:
covid19co.fillna(value='-', inplace=True)
# Show dataframe
covid19co.head()
# %%
# Setup Date Format
def setup_date(value):
try:
value = value.split('T')[0].split('-')
if len(value) == 3:
value = value[2] + '/' + value[1] + '/' + value[0]
else:
value = '-'
except IndexError:
value = '-'
if len(value) != 10 and len(value) != 1:
value = '-'
return value
# Date Columns
date_columns = list(filter(lambda value: value.find('FECHA') != -1 or value.find('FIS') != -1, covid19co.columns))
# For each date column
for date_column in date_columns:
covid19co[date_column] = covid19co[date_column].transform(lambda value: setup_date(value))
# Show dataframe
covid19co.head()
# %%
# Add Day, Month, Year, Month Name and Day Name for each Date
# Spanish
nombre_mes = ['Enero', 'Febrero', 'Marzo', 'Abril', 'Mayo', 'Junio', 'Julio', 'Agosto', 'Septiembre', 'Octubre', 'Noviembre', 'Diciembre']
nombre_dia = ['Lunes', 'Martes', 'Miércoles', 'Jueves', 'Viernes', 'Sábado', 'Domingo']
# Get day
def get_day(value):
if value not in '-':
return value.split('/')[0]
return value
# Get month
def get_month(value):
if value not in '-':
return value.split('/')[1]
return value
# Get year
def get_year(value):
if value not in '-':
return value.split('/')[2]
return value
# Get month name
def get_month_name(value):
if value not in '-':
return nombre_mes[int(value.split('/')[1]) - 1]
return value
# Get weekday
def get_weekday(value):
if value not in '-':
return nombre_dia[datetime.date(int(value.split('/')[2]), int(value.split('/')[1]), int(value.split('/')[0])).weekday()]
return value
# For each date column
for date_column in date_columns:
covid19co[date_column + ' DIA'] = covid19co[date_column].transform(lambda value: get_day(value))
covid19co[date_column + ' MES'] = covid19co[date_column].transform(lambda value: get_month(value))
covid19co[date_column + ' ANIO'] = covid19co[date_column].transform(lambda value: get_year(value))
covid19co[date_column + ' NOMBRE MES'] = covid19co[date_column].transform(lambda value: get_month_name(value))
covid19co[date_column + ' DIA SEMANA'] = covid19co[date_column].transform(lambda value: get_weekday(value))
# Show dataframe
covid19co.head()
# %% [markdown]
# ## Covid19 Colombia Dataset
# > ***Output file***: covid19co.csv
# %%
# Save dataframe
covid19co.to_csv(os.path.join(OUTPUT_DIR, 'covid19co.csv'), index=False)
# %% [markdown]
# ---
# %% [markdown]
# ## Official Covid19 Colombia Samples Processed
# %%
# Official Samples Processed Until Now
with requests.get(URL_SAMPLES_PROCESSED) as official_dataset:
with open(os.path.join(INPUT_DIR, 'covid19co_samples_processed_official.csv'), 'wb') as dataset_file:
dataset_file.write(official_dataset.content)
# %%
# Open Official Samples Processed
covid19co_samples_processed = pd.read_csv(os.path.join(INPUT_DIR, 'covid19co_samples_processed_official.csv'))
# Total Daily Report
covid19co_samples_processed.shape
# %%
# Show dataframe
covid19co_samples_processed.head()
# %%
# Update Name Columns
# Remove Accents and Uppercase
covid19co_samples_processed.columns = [unidecode.unidecode(value).upper() for value in covid19co_samples_processed.columns]
# Show dataframe
covid19co_samples_processed.head()
# %%
# Setup Date Format
covid19co_samples_processed['FECHA'] = covid19co_samples_processed['FECHA'].transform(lambda value: setup_date(value))
# Show dataframe
covid19co_samples_processed.head()
# %%
# Select Columns
covid19co_samples_processed = covid19co_samples_processed[['FECHA', 'ACUMULADAS']]
# Show dataframe
covid19co_samples_processed.tail()
# %% [markdown]
# ## Covid19 Colombia Samples Processed Dataset
# > ***Output file***: covid19co_samples_processed.csv
# %%
# Save dataframe
covid19co_samples_processed.to_csv(os.path.join(OUTPUT_DIR, 'covid19co_samples_processed.csv'), index=False)
# %% [markdown]
# ---
# %% [markdown]
# ## Google Community Mobility Reports - Colombia
# %%
# Open Official Google Community Mobility - Colombia
google_community_mobility_reports = pd.read_csv(os.path.join(OUTPUT_DIR, 'google_community_mobility_reports.csv'))
# Total Google Community Mobility - Colombia
google_community_mobility_reports.shape
# %%
# Show dataframe
google_community_mobility_reports.tail()
# %%
# Update Google Community Mobility Reports - Colombia
date_last_report = google_community_mobility_reports['date'].values[-1]
#print(date_last_report)
# 13/05/2020
date_last_report = date_last_report.split('/')
date_last_report = date_last_report[2] + '-' + date_last_report[1] + '-' + date_last_report[0]
#print(date_last_report)
# 2020-05-13
new_reports = pd.DataFrame(columns=['date', 'country', 'file', 'url'])
new_reports['date'] = [dti.strftime('%Y-%m-%d') for dti in pd.date_range(start=date_last_report, end=datetime.date.today().isoformat(), freq='D')]
new_reports['country'] = 'Colombia'
new_reports['file'] = [date + '_CO_Mobility_Report_en.pdf' for date in new_reports['date'].values]
# Get URL report
def get_report_url(file):
with requests.get('https://www.gstatic.com/covid19/mobility/' + file) as community_mobility_report:
if community_mobility_report.status_code == 200:
return community_mobility_report.url
else:
return np.nan
# Get URL |
# %%
# Update texto to title text format
for attr in covid19co.columns:
if covid19co[attr].dtypes == 'object': | random_line_split |
covid19co_pipe.py | and collaborate to the analysis here:
# [colombia_covid_19_analysis](https://www.kaggle.com/sebaxtian/colombia-covid-19-analysis) Kaggle Notebook Kernel.
# %% [markdown]
# ---
# %% [markdown]
# ## Data Sources
# %%
# Input data files are available in the "../input/" directory.
INPUT_DIR = './'
if os.path.split(os.path.abspath('.'))[-1] == 'src':
INPUT_DIR = '../input'
# Output data files are available in the "../output/" directory.
OUTPUT_DIR = './'
if os.path.split(os.path.abspath('.'))[-1] == 'src':
OUTPUT_DIR = '../output'
# Official Daily Report Until Now
URL_OFFICIAL_DATASET = 'https://www.datos.gov.co/api/views/gt2j-8ykr/rows.csv?accessType=DOWNLOAD'
# Official Daily Samples Processed
URL_SAMPLES_PROCESSED = 'https://www.datos.gov.co/api/views/8835-5baf/rows.csv?accessType=DOWNLOAD'
# %% [markdown]
# ---
# %% [markdown]
# ## Official Covid19 Colombia Daily Report
# %%
# Official Daily Report Until Now
with requests.get(URL_OFFICIAL_DATASET) as official_dataset:
with open(os.path.join(INPUT_DIR, 'covid19co_official.csv'), 'wb') as dataset_file:
dataset_file.write(official_dataset.content)
# %%
# Open Official Daily Report
covid19co = pd.read_csv(os.path.join(INPUT_DIR, 'covid19co_official.csv'))
# Total Daily Report
covid19co.shape
# %%
# Show dataframe
covid19co.tail()
# %%
# Show attributes
list(covid19co.columns.values)
# %%
# Update Name Columns
# Remove Accents and Uppercase
covid19co.columns = [unidecode.unidecode(value).upper() for value in covid19co.columns]
# Show dataframe
covid19co.head()
# %%
# Update texto to title text format
for attr in covid19co.columns:
if covid19co[attr].dtypes == 'object':
covid19co[attr] = covid19co[attr].transform(lambda value: str(value).title())
# Show dataframe
covid19co.head()
# %%
# Fill NaN Values
if covid19co.isna().sum().sum() > 0:
covid19co.fillna(value='-', inplace=True)
# Show dataframe
covid19co.head()
# %%
# Setup Date Format
def setup_date(value):
try:
value = value.split('T')[0].split('-')
if len(value) == 3:
value = value[2] + '/' + value[1] + '/' + value[0]
else:
value = '-'
except IndexError:
value = '-'
if len(value) != 10 and len(value) != 1:
value = '-'
return value
# Date Columns
date_columns = list(filter(lambda value: value.find('FECHA') != -1 or value.find('FIS') != -1, covid19co.columns))
# For each date column
for date_column in date_columns:
covid19co[date_column] = covid19co[date_column].transform(lambda value: setup_date(value))
# Show dataframe
covid19co.head()
# %%
# Add Day, Month, Year, Month Name and Day Name for each Date
# Spanish
nombre_mes = ['Enero', 'Febrero', 'Marzo', 'Abril', 'Mayo', 'Junio', 'Julio', 'Agosto', 'Septiembre', 'Octubre', 'Noviembre', 'Diciembre']
nombre_dia = ['Lunes', 'Martes', 'Miércoles', 'Jueves', 'Viernes', 'Sábado', 'Domingo']
# Get day
def get_day(value):
if value not in '-':
return value.split('/')[0]
return value
# Get month
def get_month(value):
if value not in '-':
return value.split('/')[1]
return value
# Get year
def get_year(value):
if value not in '-':
return value.split('/')[2]
return value
# Get month name
def get_month_name(value):
if value not in '-':
return nombre_mes[int(value.split('/')[1]) - 1]
return value
# Get weekday
def get_weekday(value):
if value not in '-':
return nombre_dia[datetime.date(int(value.split('/')[2]), int(value.split('/')[1]), int(value.split('/')[0])).weekday()]
return value
# For each date column
for date_column in date_columns:
covid19co[date_column + ' DIA'] = covid19co[date_column].transform(lambda value: get_day(value))
covid19co[date_column + ' MES'] = covid19co[date_column].transform(lambda value: get_month(value))
covid19co[date_column + ' ANIO'] = covid19co[date_column].transform(lambda value: get_year(value))
covid19co[date_column + ' NOMBRE MES'] = covid19co[date_column].transform(lambda value: get_month_name(value))
covid19co[date_column + ' DIA SEMANA'] = covid19co[date_column].transform(lambda value: get_weekday(value))
# Show dataframe
covid19co.head()
# %% [markdown]
# ## Covid19 Colombia Dataset
# > ***Output file***: covid19co.csv
# %%
# Save dataframe
covid19co.to_csv(os.path.join(OUTPUT_DIR, 'covid19co.csv'), index=False)
# %% [markdown]
# ---
# %% [markdown]
# ## Official Covid19 Colombia Samples Processed
# %%
# Official Samples Processed Until Now
with requests.get(URL_SAMPLES_PROCESSED) as official_dataset:
with open(os.path.join(INPUT_DIR, 'covid19co_samples_processed_official.csv'), 'wb') as dataset_file:
dataset_file.write(official_dataset.content)
# %%
# Open Official Samples Processed
covid19co_samples_processed = pd.read_csv(os.path.join(INPUT_DIR, 'covid19co_samples_processed_official.csv'))
# Total Daily Report
covid19co_samples_processed.shape
# %%
# Show dataframe
covid19co_samples_processed.head()
# %%
# Update Name Columns
# Remove Accents and Uppercase
covid19co_samples_processed.columns = [unidecode.unidecode(value).upper() for value in covid19co_samples_processed.columns]
# Show dataframe
covid19co_samples_processed.head()
# %%
# Setup Date Format
covid19co_samples_processed['FECHA'] = covid19co_samples_processed['FECHA'].transform(lambda value: setup_date(value))
# Show dataframe
covid19co_samples_processed.head()
# %%
# Select Columns
covid19co_samples_processed = covid19co_samples_processed[['FECHA', 'ACUMULADAS']]
# Show dataframe
covid19co_samples_processed.tail()
# %% [markdown]
# ## Covid19 Colombia Samples Processed Dataset
# > ***Output file***: covid19co_samples_processed.csv
# %%
# Save dataframe
covid19co_samples_processed.to_csv(os.path.join(OUTPUT_DIR, 'covid19co_samples_processed.csv'), index=False)
# %% [markdown]
# ---
# %% [markdown]
# ## Google Community Mobility Reports - Colombia
# %%
# Open Official Google Community Mobility - Colombia
google_community_mobility_reports = pd.read_csv(os.path.join(OUTPUT_DIR, 'google_community_mobility_reports.csv'))
# Total Google Community Mobility - Colombia
google_community_mobility_reports.shape
# %%
# Show dataframe
google_community_mobility_reports.tail()
# %%
# Update Google Community Mobility Reports - Colombia
date_last_report = google_community_mobility_reports['date'].values[-1]
#print(date_last_report)
# 13/05/2020
date_last_report = date_last_report.split('/')
date_last_report = date_last_report[2] + '-' + date_last_report[1] + '-' + date_last_report[0]
#print(date_last_report)
# 2020-05-13
new_reports = pd.DataFrame(columns=['date', 'country', 'file', 'url'])
new_reports['date'] = [dti.strftime('%Y-%m-%d') for dti in pd.date_range(start=date_last_report, end=datetime.date.today().isoformat(), freq='D')]
new_reports['country'] = 'Colombia'
new_reports['file'] = [date + '_CO_Mobility_Report_en.pdf' for date in new_reports['date'].values]
# Get URL report
def get_report_url(file):
with requests.get('https://www.gstatic.com/covid19/mobility/' + file) as community_mobility_report:
if community_mobility_report.status_code == 200:
return community_mobility_report.url
else:
return np.nan
# Get URL report
new_reports['url'] = new_reports['file'].transform(lambda value: get_report_url(value))
# Drop any report without URL
new_reports.dropna(inplace=True)
# Reset index
new_reports.reset_index(inplace=True, drop=True)
# Only new reports
new_reports = new_reports.iloc[1:]
# Show dataframe
new_reports.head()
# %%
# Get/Add Mobility Changes
def ge | t_mobility_changes(U | identifier_name |
|
covid19co_pipe.py | ocial/Casos-positivos-de-COVID-19-en-Colombia/gt2j-8ykr)
#
# The number of new cases are increasing day by day around the world.
# This dataset has information about reported cases from 32 Colombia departments.
#
# Also you can get the dataset Google COVID-19 Community Mobility Reports - Colombia.
#
# You can view and collaborate to the analysis here:
# [colombia_covid_19_analysis](https://www.kaggle.com/sebaxtian/colombia-covid-19-analysis) Kaggle Notebook Kernel.
# %% [markdown]
# ---
# %% [markdown]
# ## Data Sources
# %%
# Input data files are available in the "../input/" directory.
INPUT_DIR = './'
if os.path.split(os.path.abspath('.'))[-1] == 'src':
INPUT_DIR = '../input'
# Output data files are available in the "../output/" directory.
OUTPUT_DIR = './'
if os.path.split(os.path.abspath('.'))[-1] == 'src':
OUTPUT_DIR = '../output'
# Official Daily Report Until Now
URL_OFFICIAL_DATASET = 'https://www.datos.gov.co/api/views/gt2j-8ykr/rows.csv?accessType=DOWNLOAD'
# Official Daily Samples Processed
URL_SAMPLES_PROCESSED = 'https://www.datos.gov.co/api/views/8835-5baf/rows.csv?accessType=DOWNLOAD'
# %% [markdown]
# ---
# %% [markdown]
# ## Official Covid19 Colombia Daily Report
# %%
# Official Daily Report Until Now
with requests.get(URL_OFFICIAL_DATASET) as official_dataset:
with open(os.path.join(INPUT_DIR, 'covid19co_official.csv'), 'wb') as dataset_file:
dataset_file.write(official_dataset.content)
# %%
# Open Official Daily Report
covid19co = pd.read_csv(os.path.join(INPUT_DIR, 'covid19co_official.csv'))
# Total Daily Report
covid19co.shape
# %%
# Show dataframe
covid19co.tail()
# %%
# Show attributes
list(covid19co.columns.values)
# %%
# Update Name Columns
# Remove Accents and Uppercase
covid19co.columns = [unidecode.unidecode(value).upper() for value in covid19co.columns]
# Show dataframe
covid19co.head()
# %%
# Update texto to title text format
for attr in covid19co.columns:
if covid19co[attr].dtypes == 'object':
covid19co[attr] = covid19co[attr].transform(lambda value: str(value).title())
# Show dataframe
covid19co.head()
# %%
# Fill NaN Values
if covid19co.isna().sum().sum() > 0:
covid19co.fillna(value='-', inplace=True)
# Show dataframe
covid19co.head()
# %%
# Setup Date Format
def setup_date(value):
try:
value = value.split('T')[0].split('-')
if len(value) == 3:
value = value[2] + '/' + value[1] + '/' + value[0]
else:
value = '-'
except IndexError:
value = '-'
if len(value) != 10 and len(value) != 1:
value = '-'
return value
# Date Columns
date_columns = list(filter(lambda value: value.find('FECHA') != -1 or value.find('FIS') != -1, covid19co.columns))
# For each date column
for date_column in date_columns:
covid19co[date_column] = covid19co[date_column].transform(lambda value: setup_date(value))
# Show dataframe
covid19co.head()
# %%
# Add Day, Month, Year, Month Name and Day Name for each Date
# Spanish
nombre_mes = ['Enero', 'Febrero', 'Marzo', 'Abril', 'Mayo', 'Junio', 'Julio', 'Agosto', 'Septiembre', 'Octubre', 'Noviembre', 'Diciembre']
nombre_dia = ['Lunes', 'Martes', 'Miércoles', 'Jueves', 'Viernes', 'Sábado', 'Domingo']
# Get day
def get_day(value):
if value not in '-':
return value.split('/')[0]
return value
# Get month
def get_month(value):
if value not in '-':
return value.split('/')[1]
return value
# Get year
def get_year(value):
if value not in '-':
return value.split('/')[2]
return value
# Get month name
def get_month_name(value):
if value not in '-':
return nombre_mes[int(value.split('/')[1]) - 1]
return value
# Get weekday
def get_weekday(value):
if value not in '-':
return nombre_dia[datetime.date(int(value.split('/')[2]), int(value.split('/')[1]), int(value.split('/')[0])).weekday()]
return value
# For each date column
for date_column in date_columns:
covid19co[date_column + ' DIA'] = covid19co[date_column].transform(lambda value: get_day(value))
covid19co[date_column + ' MES'] = covid19co[date_column].transform(lambda value: get_month(value))
covid19co[date_column + ' ANIO'] = covid19co[date_column].transform(lambda value: get_year(value))
covid19co[date_column + ' NOMBRE MES'] = covid19co[date_column].transform(lambda value: get_month_name(value))
covid19co[date_column + ' DIA SEMANA'] = covid19co[date_column].transform(lambda value: get_weekday(value))
# Show dataframe
covid19co.head()
# %% [markdown]
# ## Covid19 Colombia Dataset
# > ***Output file***: covid19co.csv
# %%
# Save dataframe
covid19co.to_csv(os.path.join(OUTPUT_DIR, 'covid19co.csv'), index=False)
# %% [markdown]
# ---
# %% [markdown]
# ## Official Covid19 Colombia Samples Processed
# %%
# Official Samples Processed Until Now
with requests.get(URL_SAMPLES_PROCESSED) as official_dataset:
with open(os.path.join(INPUT_DIR, 'covid19co_samples_processed_official.csv'), 'wb') as dataset_file:
dataset_file.write(official_dataset.content)
# %%
# Open Official Samples Processed
covid19co_samples_processed = pd.read_csv(os.path.join(INPUT_DIR, 'covid19co_samples_processed_official.csv'))
# Total Daily Report
covid19co_samples_processed.shape
# %%
# Show dataframe
covid19co_samples_processed.head()
# %%
# Update Name Columns
# Remove Accents and Uppercase
covid19co_samples_processed.columns = [unidecode.unidecode(value).upper() for value in covid19co_samples_processed.columns]
# Show dataframe
covid19co_samples_processed.head()
# %%
# Setup Date Format
covid19co_samples_processed['FECHA'] = covid19co_samples_processed['FECHA'].transform(lambda value: setup_date(value))
# Show dataframe
covid19co_samples_processed.head()
# %%
# Select Columns
covid19co_samples_processed = covid19co_samples_processed[['FECHA', 'ACUMULADAS']]
# Show dataframe
covid19co_samples_processed.tail()
# %% [markdown]
# ## Covid19 Colombia Samples Processed Dataset
# > ***Output file***: covid19co_samples_processed.csv
# %%
# Save dataframe
covid19co_samples_processed.to_csv(os.path.join(OUTPUT_DIR, 'covid19co_samples_processed.csv'), index=False)
# %% [markdown]
# ---
# %% [markdown]
# ## Google Community Mobility Reports - Colombia
# %%
# Open Official Google Community Mobility - Colombia
google_community_mobility_reports = pd.read_csv(os.path.join(OUTPUT_DIR, 'google_community_mobility_reports.csv'))
# Total Google Community Mobility - Colombia
google_community_mobility_reports.shape
# %%
# Show dataframe
google_community_mobility_reports.tail()
# %%
# Update Google Community Mobility Reports - Colombia
date_last_report = google_community_mobility_reports['date'].values[-1]
#print(date_last_report)
# 13/05/2020
date_last_report = date_last_report.split('/')
date_last_report = date_last_report[2] + '-' + date_last_report[1] + '-' + date_last_report[0]
#print(date_last_report)
# 2020-05-13
new_reports = pd.DataFrame(columns=['date', 'country', 'file', 'url'])
new_reports['date'] = [dti.strftime('%Y-%m-%d') for dti in pd.date_range(start=date_last_report, end=datetime.date.today().isoformat(), freq='D')]
new_reports['country'] = 'Colombia'
new_reports['file'] = [date + '_CO_Mobility_Report_en.pdf' for date in new_reports['date'].values]
# Get URL report
def get_report_url(file):
wi | Get URL report | th requests.get('https://www.gstatic.com/covid19/mobility/' + file) as community_mobility_report:
if community_mobility_report.status_code == 200:
return community_mobility_report.url
else:
return np.nan
# | identifier_body |
covid19co_pipe.py | ocial/Casos-positivos-de-COVID-19-en-Colombia/gt2j-8ykr)
#
# The number of new cases are increasing day by day around the world.
# This dataset has information about reported cases from 32 Colombia departments.
#
# Also you can get the dataset Google COVID-19 Community Mobility Reports - Colombia.
#
# You can view and collaborate to the analysis here:
# [colombia_covid_19_analysis](https://www.kaggle.com/sebaxtian/colombia-covid-19-analysis) Kaggle Notebook Kernel.
# %% [markdown]
# ---
# %% [markdown]
# ## Data Sources
# %%
# Input data files are available in the "../input/" directory.
INPUT_DIR = './'
if os.path.split(os.path.abspath('.'))[-1] == 'src':
INPUT_DIR = '../input'
# Output data files are available in the "../output/" directory.
OUTPUT_DIR = './'
if os.path.split(os.path.abspath('.'))[-1] == 'src':
OUTPUT_DIR = '../output'
# Official Daily Report Until Now
URL_OFFICIAL_DATASET = 'https://www.datos.gov.co/api/views/gt2j-8ykr/rows.csv?accessType=DOWNLOAD'
# Official Daily Samples Processed
URL_SAMPLES_PROCESSED = 'https://www.datos.gov.co/api/views/8835-5baf/rows.csv?accessType=DOWNLOAD'
# %% [markdown]
# ---
# %% [markdown]
# ## Official Covid19 Colombia Daily Report
# %%
# Official Daily Report Until Now
with requests.get(URL_OFFICIAL_DATASET) as official_dataset:
with open(os.path.join(INPUT_DIR, 'covid19co_official.csv'), 'wb') as dataset_file:
dataset_file.write(official_dataset.content)
# %%
# Open Official Daily Report
covid19co = pd.read_csv(os.path.join(INPUT_DIR, 'covid19co_official.csv'))
# Total Daily Report
covid19co.shape
# %%
# Show dataframe
covid19co.tail()
# %%
# Show attributes
list(covid19co.columns.values)
# %%
# Update Name Columns
# Remove Accents and Uppercase
covid19co.columns = [unidecode.unidecode(value).upper() for value in covid19co.columns]
# Show dataframe
covid19co.head()
# %%
# Update texto to title text format
for attr in covid19co.columns:
if covid19co[attr].dtypes == 'object':
covid19co[attr] = covid19co[attr].transform(lambda value: str(value).title())
# Show dataframe
covid19co.head()
# %%
# Fill NaN Values
if covid19co.isna().sum().sum() > 0:
covid19co.fillna(value='-', inplace=True)
# Show dataframe
covid19co.head()
# %%
# Setup Date Format
def setup_date(value):
try:
value = value.split('T')[0].split('-')
if len(value) == 3:
|
else:
value = '-'
except IndexError:
value = '-'
if len(value) != 10 and len(value) != 1:
value = '-'
return value
# Date Columns
date_columns = list(filter(lambda value: value.find('FECHA') != -1 or value.find('FIS') != -1, covid19co.columns))
# For each date column
for date_column in date_columns:
covid19co[date_column] = covid19co[date_column].transform(lambda value: setup_date(value))
# Show dataframe
covid19co.head()
# %%
# Add Day, Month, Year, Month Name and Day Name for each Date
# Spanish
nombre_mes = ['Enero', 'Febrero', 'Marzo', 'Abril', 'Mayo', 'Junio', 'Julio', 'Agosto', 'Septiembre', 'Octubre', 'Noviembre', 'Diciembre']
nombre_dia = ['Lunes', 'Martes', 'Miércoles', 'Jueves', 'Viernes', 'Sábado', 'Domingo']
# Get day
def get_day(value):
if value not in '-':
return value.split('/')[0]
return value
# Get month
def get_month(value):
if value not in '-':
return value.split('/')[1]
return value
# Get year
def get_year(value):
if value not in '-':
return value.split('/')[2]
return value
# Get month name
def get_month_name(value):
if value not in '-':
return nombre_mes[int(value.split('/')[1]) - 1]
return value
# Get weekday
def get_weekday(value):
if value not in '-':
return nombre_dia[datetime.date(int(value.split('/')[2]), int(value.split('/')[1]), int(value.split('/')[0])).weekday()]
return value
# For each date column
for date_column in date_columns:
covid19co[date_column + ' DIA'] = covid19co[date_column].transform(lambda value: get_day(value))
covid19co[date_column + ' MES'] = covid19co[date_column].transform(lambda value: get_month(value))
covid19co[date_column + ' ANIO'] = covid19co[date_column].transform(lambda value: get_year(value))
covid19co[date_column + ' NOMBRE MES'] = covid19co[date_column].transform(lambda value: get_month_name(value))
covid19co[date_column + ' DIA SEMANA'] = covid19co[date_column].transform(lambda value: get_weekday(value))
# Show dataframe
covid19co.head()
# %% [markdown]
# ## Covid19 Colombia Dataset
# > ***Output file***: covid19co.csv
# %%
# Save dataframe
covid19co.to_csv(os.path.join(OUTPUT_DIR, 'covid19co.csv'), index=False)
# %% [markdown]
# ---
# %% [markdown]
# ## Official Covid19 Colombia Samples Processed
# %%
# Official Samples Processed Until Now
with requests.get(URL_SAMPLES_PROCESSED) as official_dataset:
with open(os.path.join(INPUT_DIR, 'covid19co_samples_processed_official.csv'), 'wb') as dataset_file:
dataset_file.write(official_dataset.content)
# %%
# Open Official Samples Processed
covid19co_samples_processed = pd.read_csv(os.path.join(INPUT_DIR, 'covid19co_samples_processed_official.csv'))
# Total Daily Report
covid19co_samples_processed.shape
# %%
# Show dataframe
covid19co_samples_processed.head()
# %%
# Update Name Columns
# Remove Accents and Uppercase
covid19co_samples_processed.columns = [unidecode.unidecode(value).upper() for value in covid19co_samples_processed.columns]
# Show dataframe
covid19co_samples_processed.head()
# %%
# Setup Date Format
covid19co_samples_processed['FECHA'] = covid19co_samples_processed['FECHA'].transform(lambda value: setup_date(value))
# Show dataframe
covid19co_samples_processed.head()
# %%
# Select Columns
covid19co_samples_processed = covid19co_samples_processed[['FECHA', 'ACUMULADAS']]
# Show dataframe
covid19co_samples_processed.tail()
# %% [markdown]
# ## Covid19 Colombia Samples Processed Dataset
# > ***Output file***: covid19co_samples_processed.csv
# %%
# Save dataframe
covid19co_samples_processed.to_csv(os.path.join(OUTPUT_DIR, 'covid19co_samples_processed.csv'), index=False)
# %% [markdown]
# ---
# %% [markdown]
# ## Google Community Mobility Reports - Colombia
# %%
# Open Official Google Community Mobility - Colombia
google_community_mobility_reports = pd.read_csv(os.path.join(OUTPUT_DIR, 'google_community_mobility_reports.csv'))
# Total Google Community Mobility - Colombia
google_community_mobility_reports.shape
# %%
# Show dataframe
google_community_mobility_reports.tail()
# %%
# Update Google Community Mobility Reports - Colombia
date_last_report = google_community_mobility_reports['date'].values[-1]
#print(date_last_report)
# 13/05/2020
date_last_report = date_last_report.split('/')
date_last_report = date_last_report[2] + '-' + date_last_report[1] + '-' + date_last_report[0]
#print(date_last_report)
# 2020-05-13
new_reports = pd.DataFrame(columns=['date', 'country', 'file', 'url'])
new_reports['date'] = [dti.strftime('%Y-%m-%d') for dti in pd.date_range(start=date_last_report, end=datetime.date.today().isoformat(), freq='D')]
new_reports['country'] = 'Colombia'
new_reports['file'] = [date + '_CO_Mobility_Report_en.pdf' for date in new_reports['date'].values]
# Get URL report
def get_report_url(file):
with requests.get('https://www.gstatic.com/covid19/mobility/' + file) as community_mobility_report:
if community_mobility_report.status_code == 200:
return community_mobility_report.url
else:
return np.nan
# Get URL report | value = value[2] + '/' + value[1] + '/' + value[0] | conditional_block |
peermanager.go | .conf.NPAddPeers {
// go-multiaddr implementation does not support recent p2p protocol yet, but deprecated name ipfs.
// This adhoc will be removed when go-multiaddr is patched.
target = strings.Replace(target, "/p2p/", "/ipfs/", 1)
targetAddr, err := ma.NewMultiaddr(target)
if err != nil {
pm.logger.Warn().Err(err).Str("target", target).Msg("invalid NPAddPeer address")
continue
}
splitted := strings.Split(targetAddr.String(), "/")
if len(splitted) != 7 {
pm.logger.Warn().Str("target", target).Msg("invalid NPAddPeer address")
continue
}
peerAddrString := splitted[2]
peerPortString := splitted[4]
peerPort, err := strconv.Atoi(peerPortString)
if err != nil {
pm.logger.Warn().Str("port", peerPortString).Msg("invalid Peer port")
continue
}
peerIDString := splitted[6]
peerID, err := peer.IDB58Decode(peerIDString)
if err != nil {
pm.logger.Warn().Str(LogPeerID, peerIDString).Msg("invalid PeerID")
continue
}
peerMeta := PeerMeta{
ID: peerID,
Port: uint32(peerPort),
IPAddress: peerAddrString,
Designated: true,
Outbound: true,
}
pm.logger.Info().Str(LogPeerID, peerID.Pretty()).Str("addr", peerAddrString).Int("port", peerPort).Msg("Adding Designated peer")
pm.designatedPeers[peerID] = peerMeta
}
}
func (pm *peerManager) runManagePeers() {
addrDuration := time.Minute * 3
addrTicker := time.NewTicker(addrDuration)
// reconnectRunners := make(map[peer.ID]*reconnectRunner)
MANLOOP:
for {
select {
case meta := <-pm.addPeerChannel:
if pm.addOutboundPeer(meta) {
if _, found := pm.designatedPeers[meta.ID]; found {
pm.rm.CancelJob(meta.ID)
}
}
case id := <-pm.removePeerChannel:
if pm.removePeer(id) {
if meta, found := pm.designatedPeers[id]; found {
pm.rm.AddJob(meta)
}
}
case <-addrTicker.C:
pm.checkAndCollectPeerListFromAll()
pm.logPeerMetrics()
case peerMetas := <-pm.fillPoolChannel:
pm.tryFillPool(&peerMetas)
case <-pm.finishChannel:
addrTicker.Stop()
pm.rm.Stop()
// TODO need to keep loop till all remote peer objects are removed, otherwise panic or channel deadlock can come.
break MANLOOP
}
}
// cleanup peers
for peerID := range pm.remotePeers {
pm.removePeer(peerID)
}
}
func (pm *peerManager) logPeerMetrics() {
if pm.logger.IsDebugEnabled() {
pm.logger.Debug().Msg(pm.mm.Summary())
}
}
// addOutboundPeer try to connect and handshake to remote peer. it can be called after peermanager is inited.
// It return true if peer is added or already exist, or return false if failed to add peer.
func (pm *peerManager) addOutboundPeer(meta PeerMeta) bool {
addrString := fmt.Sprintf("/ip4/%s/tcp/%d", meta.IPAddress, meta.Port)
var peerAddr, err = ma.NewMultiaddr(addrString)
if err != nil {
pm.logger.Warn().Err(err).Str("addr", addrString).Msg("invalid NPAddPeer address")
return false
}
var peerID = meta.ID
pm.mutex.Lock()
inboundPeer, ok := pm.remotePeers[peerID]
if ok {
// peer is already exist (and maybe inbound peer)
pm.logger.Info().Str(LogPeerID, inboundPeer.meta.ID.Pretty()).Msg("Peer is already managed by peermanager")
if meta.Designated {
// If remote peer was connected first. designated flag is not set yet.
inboundPeer.meta.Designated = true
}
pm.mutex.Unlock()
return true
}
pm.mutex.Unlock()
pm.Peerstore().AddAddr(peerID, peerAddr, meta.TTL())
ctx := context.Background()
s, err := pm.NewStream(ctx, meta.ID, aergoP2PSub)
if err != nil {
pm.logger.Info().Err(err).Str("addr", addrString).Str(LogPeerID, meta.ID.Pretty()).Str(LogProtoID, string(aergoP2PSub)).Msg("Error while get stream")
return false
}
rd := metric.NewReader(s)
wt := metric.NewWriter(s)
h := newHandshaker(pm, pm.actorServ, pm.logger, peerID)
rw, remoteStatus, err := h.handshakeOutboundPeerTimeout(rd, wt, defaultHandshakeTTL)
if err != nil {
pm.logger.Debug().Err(err).Str(LogPeerID, meta.ID.Pretty()).Msg("Failed to handshake")
//pm.sendGoAway(rw, "Failed to handshake")
s.Close()
return false
}
pm.mutex.Lock()
inboundPeer, ok = pm.remotePeers[peerID]
if ok {
if ComparePeerID(pm.selfMeta.ID, meta.ID) <= 0 {
pm.logger.Info().Str(LogPeerID, inboundPeer.meta.ID.Pretty()).Msg("Inbound connection was already handshaked while handshaking outbound connection, and remote peer is higher priority so closing this outbound connection.")
pm.mutex.Unlock()
pm.sendGoAway(rw, "Already handshaked")
s.Close()
return true
} else {
pm.logger.Info().Str(LogPeerID, inboundPeer.meta.ID.Pretty()).Msg("Inbound connection was already handshaked while handshaking outbound connection, but local peer is higher priority so closing that inbound connection")
// disconnect lower valued connection
pm.deletePeer(meta.ID)
inboundPeer.stop()
}
}
// update peer info to remote sent infor
meta = FromPeerAddress(remoteStatus.Sender)
outboundPeer := newRemotePeer(meta, pm, pm.actorServ, pm.logger, pm.mf, pm.signer, rw)
// insert Handlers
pm.handlerFactory.insertHandlers(outboundPeer)
go outboundPeer.runPeer()
pm.insertPeer(peerID, outboundPeer)
pm.logger.Info().Str(LogPeerID, peerID.Pretty()).Str("addr", net.ParseIP(meta.IPAddress).String()+":"+strconv.Itoa(int(meta.Port))).Msg("Outbound peer is added to peerService")
outboundPeer.metric = pm.mm.Add(peerID, rd, wt)
pm.mutex.Unlock()
addrs := pm.Peerstore().Addrs(peerID)
addrStrs := make([]string, len(addrs))
for i, addr := range addrs {
addrStrs[i] = addr.String()
}
pm.logger.Debug().Strs("addrs", addrStrs).Str(LogPeerID, outboundPeer.meta.ID.Pretty()).Msg("addresses of peer")
// peer is ready
h.doInitialSync()
// notice to p2pmanager that handshaking is finished
pm.NotifyPeerHandshake(peerID)
return true
}
func (pm *peerManager) sendGoAway(rw MsgReadWriter, msg string) {
goMsg := &types.GoAwayNotice{Message: msg}
// TODO code smell. non safe casting.
mo := pm.mf.newMsgRequestOrder(false, GoAway, goMsg).(*pbRequestOrder)
container := mo.message
rw.WriteMsg(container)
}
func (pm *peerManager) checkInPeerstore(peerID peer.ID) bool {
found := false
for _, existingPeerID := range pm.Peerstore().Peers() {
if existingPeerID == peerID {
found = true
break
}
}
return found
}
func (pm *peerManager) AddNewPeer(peer PeerMeta) {
pm.addPeerChannel <- peer
}
func (pm *peerManager) RemovePeer(peerID peer.ID) {
pm.removePeerChannel <- peerID
}
func (pm *peerManager) NotifyPeerHandshake(peerID peer.ID) {
pm.checkAndCollectPeerList(peerID)
}
func (pm *peerManager) NotifyPeerAddressReceived(metas []PeerMeta) {
pm.fillPoolChannel <- metas
}
// removePeer remove and disconnect managed remote peer connection
// It return true if peer is exist and managed by peermanager
func (pm *peerManager) removePeer(peerID peer.ID) bool {
pm.mutex.Lock()
target, ok := pm.remotePeers[peerID]
if !ok {
pm.mutex.Unlock()
return false
}
pm.deletePeer(peerID)
// No internal module access this peer anymore, but remote message can be received.
target.stop()
pm.mutex.Unlock()
// also disconnect connection
for _, existingPeerID := range pm.Peerstore().Peers() {
if existingPeerID == peerID {
for _, listener := range pm.eventListeners {
listener.OnRemovePeer(peerID)
}
pm.Network().ClosePeer(peerID)
return true
}
}
return true
}
func (pm *peerManager) Peerstore() pstore.Peerstore | {
return pm.Host.Peerstore()
} | identifier_body |
|
peermanager.go | }
func (pm *peerManager) SelfMeta() PeerMeta {
return pm.selfMeta
}
func (pm *peerManager) SelfNodeID() peer.ID {
return pm.selfMeta.ID
}
func (pm *peerManager) RegisterEventListener(listener PeerEventListener) {
pm.mutex.Lock()
defer pm.mutex.Unlock()
pm.eventListeners = append(pm.eventListeners, listener)
}
func (pm *peerManager) init() {
// check Key and address
priv := NodePrivKey()
pub := NodePubKey()
pid := NodeID()
pm.privateKey = priv
pm.publicKey = pub
// init address and port
// if not set, it look up ip addresses of machine and choose suitable one (but not so smart) and default port 7845
peerAddr, peerPort := pm.getProtocolAddrs()
pm.selfMeta.IPAddress = peerAddr.String()
pm.selfMeta.Port = uint32(peerPort)
pm.selfMeta.ID = pid
// if bindAddress or bindPort is not set, it will be same as NetProtocolAddr or NetProtocolPort
if len(pm.conf.NPBindAddr) > 0 {
bindAddr := net.ParseIP(pm.conf.NPBindAddr)
if bindAddr == nil {
panic("invalid NPBindAddr " + pm.conf.NPBindAddr)
}
pm.bindAddress = bindAddr
} else {
pm.bindAddress = peerAddr
}
if pm.conf.NPBindPort > 0 {
pm.bindPort = pm.conf.NPBindPort
} else {
pm.bindPort = peerPort
}
// set meta info
// TODO more survey libp2p NAT configuration
// set designated peers
pm.addDesignatedPeers()
}
func (pm *peerManager) getProtocolAddrs() (protocolAddr net.IP, protocolPort int) {
if len(pm.conf.NetProtocolAddr) != 0 {
protocolAddr = net.ParseIP(pm.conf.NetProtocolAddr)
if protocolAddr == nil {
panic("invalid NetProtocolAddr " + pm.conf.NetProtocolAddr)
}
if protocolAddr.IsUnspecified() {
panic("NetProtocolAddr should be a specified IP address, not 0.0.0.0")
}
} else {
extIP, err := externalIP()
if err != nil {
panic("error while finding IP address: " + err.Error())
}
protocolAddr = extIP
}
protocolPort = pm.conf.NetProtocolPort
if protocolPort <= 0 {
panic("invalid NetProtocolPort " + strconv.Itoa(pm.conf.NetProtocolPort))
}
return
}
func (pm *peerManager) run() {
go pm.runManagePeers()
// need to start listen after chainservice is read to init
// FIXME: adhoc code
go func() {
time.Sleep(time.Second * 3)
pm.startListener()
// addition should start after all modules are started
go func() {
time.Sleep(time.Second * 2)
for _, meta := range pm.designatedPeers {
pm.addPeerChannel <- meta
}
}()
}()
}
func (pm *peerManager) addDesignatedPeers() {
// add remote node from config
for _, target := range pm.conf.NPAddPeers {
// go-multiaddr implementation does not support recent p2p protocol yet, but deprecated name ipfs.
// This adhoc will be removed when go-multiaddr is patched.
target = strings.Replace(target, "/p2p/", "/ipfs/", 1)
targetAddr, err := ma.NewMultiaddr(target)
if err != nil {
pm.logger.Warn().Err(err).Str("target", target).Msg("invalid NPAddPeer address")
continue
}
splitted := strings.Split(targetAddr.String(), "/")
if len(splitted) != 7 {
pm.logger.Warn().Str("target", target).Msg("invalid NPAddPeer address")
continue
}
peerAddrString := splitted[2]
peerPortString := splitted[4]
peerPort, err := strconv.Atoi(peerPortString)
if err != nil {
pm.logger.Warn().Str("port", peerPortString).Msg("invalid Peer port")
continue
}
peerIDString := splitted[6]
peerID, err := peer.IDB58Decode(peerIDString)
if err != nil {
pm.logger.Warn().Str(LogPeerID, peerIDString).Msg("invalid PeerID")
continue
}
peerMeta := PeerMeta{
ID: peerID,
Port: uint32(peerPort),
IPAddress: peerAddrString,
Designated: true,
Outbound: true,
}
pm.logger.Info().Str(LogPeerID, peerID.Pretty()).Str("addr", peerAddrString).Int("port", peerPort).Msg("Adding Designated peer")
pm.designatedPeers[peerID] = peerMeta
}
}
func (pm *peerManager) runManagePeers() {
addrDuration := time.Minute * 3
addrTicker := time.NewTicker(addrDuration)
// reconnectRunners := make(map[peer.ID]*reconnectRunner)
MANLOOP:
for {
select {
case meta := <-pm.addPeerChannel:
if pm.addOutboundPeer(meta) |
case id := <-pm.removePeerChannel:
if pm.removePeer(id) {
if meta, found := pm.designatedPeers[id]; found {
pm.rm.AddJob(meta)
}
}
case <-addrTicker.C:
pm.checkAndCollectPeerListFromAll()
pm.logPeerMetrics()
case peerMetas := <-pm.fillPoolChannel:
pm.tryFillPool(&peerMetas)
case <-pm.finishChannel:
addrTicker.Stop()
pm.rm.Stop()
// TODO need to keep loop till all remote peer objects are removed, otherwise panic or channel deadlock can come.
break MANLOOP
}
}
// cleanup peers
for peerID := range pm.remotePeers {
pm.removePeer(peerID)
}
}
func (pm *peerManager) logPeerMetrics() {
if pm.logger.IsDebugEnabled() {
pm.logger.Debug().Msg(pm.mm.Summary())
}
}
// addOutboundPeer try to connect and handshake to remote peer. it can be called after peermanager is inited.
// It return true if peer is added or already exist, or return false if failed to add peer.
func (pm *peerManager) addOutboundPeer(meta PeerMeta) bool {
addrString := fmt.Sprintf("/ip4/%s/tcp/%d", meta.IPAddress, meta.Port)
var peerAddr, err = ma.NewMultiaddr(addrString)
if err != nil {
pm.logger.Warn().Err(err).Str("addr", addrString).Msg("invalid NPAddPeer address")
return false
}
var peerID = meta.ID
pm.mutex.Lock()
inboundPeer, ok := pm.remotePeers[peerID]
if ok {
// peer is already exist (and maybe inbound peer)
pm.logger.Info().Str(LogPeerID, inboundPeer.meta.ID.Pretty()).Msg("Peer is already managed by peermanager")
if meta.Designated {
// If remote peer was connected first. designated flag is not set yet.
inboundPeer.meta.Designated = true
}
pm.mutex.Unlock()
return true
}
pm.mutex.Unlock()
pm.Peerstore().AddAddr(peerID, peerAddr, meta.TTL())
ctx := context.Background()
s, err := pm.NewStream(ctx, meta.ID, aergoP2PSub)
if err != nil {
pm.logger.Info().Err(err).Str("addr", addrString).Str(LogPeerID, meta.ID.Pretty()).Str(LogProtoID, string(aergoP2PSub)).Msg("Error while get stream")
return false
}
rd := metric.NewReader(s)
wt := metric.NewWriter(s)
h := newHandshaker(pm, pm.actorServ, pm.logger, peerID)
rw, remoteStatus, err := h.handshakeOutboundPeerTimeout(rd, wt, defaultHandshakeTTL)
if err != nil {
pm.logger.Debug().Err(err).Str(LogPeerID, meta.ID.Pretty()).Msg("Failed to handshake")
//pm.sendGoAway(rw, "Failed to handshake")
s.Close()
return false
}
pm.mutex.Lock()
inboundPeer, ok = pm.remotePeers[peerID]
if ok {
if ComparePeerID(pm.selfMeta.ID, meta.ID) <= 0 {
pm.logger.Info().Str(LogPeerID, inboundPeer.meta.ID.Pretty()).Msg("Inbound connection was already handshaked while handshaking outbound connection, and remote peer is higher priority so closing this outbound connection.")
pm.mutex.Unlock()
pm.sendGoAway(rw, "Already handshaked")
s.Close()
return true
} else {
pm.logger.Info().Str(LogPeerID, inboundPeer.meta.ID.Pretty()).Msg("Inbound connection was already handshaked while handshaking outbound connection, but local peer is higher priority so closing that inbound connection")
// disconnect lower valued connection
pm.deletePeer(meta.ID)
inboundPeer.stop()
| {
if _, found := pm.designatedPeers[meta.ID]; found {
pm.rm.CancelJob(meta.ID)
}
} | conditional_block |
peermanager.go | {
// add remote node from config
for _, target := range pm.conf.NPAddPeers {
// go-multiaddr implementation does not support recent p2p protocol yet, but deprecated name ipfs.
// This adhoc will be removed when go-multiaddr is patched.
target = strings.Replace(target, "/p2p/", "/ipfs/", 1)
targetAddr, err := ma.NewMultiaddr(target)
if err != nil {
pm.logger.Warn().Err(err).Str("target", target).Msg("invalid NPAddPeer address")
continue
}
splitted := strings.Split(targetAddr.String(), "/")
if len(splitted) != 7 {
pm.logger.Warn().Str("target", target).Msg("invalid NPAddPeer address")
continue
}
peerAddrString := splitted[2]
peerPortString := splitted[4]
peerPort, err := strconv.Atoi(peerPortString)
if err != nil {
pm.logger.Warn().Str("port", peerPortString).Msg("invalid Peer port")
continue
}
peerIDString := splitted[6]
peerID, err := peer.IDB58Decode(peerIDString)
if err != nil {
pm.logger.Warn().Str(LogPeerID, peerIDString).Msg("invalid PeerID")
continue
}
peerMeta := PeerMeta{
ID: peerID,
Port: uint32(peerPort),
IPAddress: peerAddrString,
Designated: true,
Outbound: true,
}
pm.logger.Info().Str(LogPeerID, peerID.Pretty()).Str("addr", peerAddrString).Int("port", peerPort).Msg("Adding Designated peer")
pm.designatedPeers[peerID] = peerMeta
}
}
func (pm *peerManager) runManagePeers() {
addrDuration := time.Minute * 3
addrTicker := time.NewTicker(addrDuration)
// reconnectRunners := make(map[peer.ID]*reconnectRunner)
MANLOOP:
for {
select {
case meta := <-pm.addPeerChannel:
if pm.addOutboundPeer(meta) {
if _, found := pm.designatedPeers[meta.ID]; found {
pm.rm.CancelJob(meta.ID)
}
}
case id := <-pm.removePeerChannel:
if pm.removePeer(id) {
if meta, found := pm.designatedPeers[id]; found {
pm.rm.AddJob(meta)
}
}
case <-addrTicker.C:
pm.checkAndCollectPeerListFromAll()
pm.logPeerMetrics()
case peerMetas := <-pm.fillPoolChannel:
pm.tryFillPool(&peerMetas)
case <-pm.finishChannel:
addrTicker.Stop()
pm.rm.Stop()
// TODO need to keep loop till all remote peer objects are removed, otherwise panic or channel deadlock can come.
break MANLOOP
}
}
// cleanup peers
for peerID := range pm.remotePeers {
pm.removePeer(peerID)
}
}
func (pm *peerManager) logPeerMetrics() {
if pm.logger.IsDebugEnabled() {
pm.logger.Debug().Msg(pm.mm.Summary())
}
}
// addOutboundPeer try to connect and handshake to remote peer. it can be called after peermanager is inited.
// It return true if peer is added or already exist, or return false if failed to add peer.
func (pm *peerManager) addOutboundPeer(meta PeerMeta) bool {
addrString := fmt.Sprintf("/ip4/%s/tcp/%d", meta.IPAddress, meta.Port)
var peerAddr, err = ma.NewMultiaddr(addrString)
if err != nil {
pm.logger.Warn().Err(err).Str("addr", addrString).Msg("invalid NPAddPeer address")
return false
}
var peerID = meta.ID
pm.mutex.Lock()
inboundPeer, ok := pm.remotePeers[peerID]
if ok {
// peer is already exist (and maybe inbound peer)
pm.logger.Info().Str(LogPeerID, inboundPeer.meta.ID.Pretty()).Msg("Peer is already managed by peermanager")
if meta.Designated {
// If remote peer was connected first. designated flag is not set yet.
inboundPeer.meta.Designated = true
}
pm.mutex.Unlock()
return true
}
pm.mutex.Unlock()
pm.Peerstore().AddAddr(peerID, peerAddr, meta.TTL())
ctx := context.Background()
s, err := pm.NewStream(ctx, meta.ID, aergoP2PSub)
if err != nil {
pm.logger.Info().Err(err).Str("addr", addrString).Str(LogPeerID, meta.ID.Pretty()).Str(LogProtoID, string(aergoP2PSub)).Msg("Error while get stream")
return false
}
rd := metric.NewReader(s)
wt := metric.NewWriter(s)
h := newHandshaker(pm, pm.actorServ, pm.logger, peerID)
rw, remoteStatus, err := h.handshakeOutboundPeerTimeout(rd, wt, defaultHandshakeTTL)
if err != nil {
pm.logger.Debug().Err(err).Str(LogPeerID, meta.ID.Pretty()).Msg("Failed to handshake")
//pm.sendGoAway(rw, "Failed to handshake")
s.Close()
return false
}
pm.mutex.Lock()
inboundPeer, ok = pm.remotePeers[peerID]
if ok {
if ComparePeerID(pm.selfMeta.ID, meta.ID) <= 0 {
pm.logger.Info().Str(LogPeerID, inboundPeer.meta.ID.Pretty()).Msg("Inbound connection was already handshaked while handshaking outbound connection, and remote peer is higher priority so closing this outbound connection.")
pm.mutex.Unlock()
pm.sendGoAway(rw, "Already handshaked")
s.Close()
return true
} else {
pm.logger.Info().Str(LogPeerID, inboundPeer.meta.ID.Pretty()).Msg("Inbound connection was already handshaked while handshaking outbound connection, but local peer is higher priority so closing that inbound connection")
// disconnect lower valued connection
pm.deletePeer(meta.ID)
inboundPeer.stop()
}
}
// update peer info to remote sent infor
meta = FromPeerAddress(remoteStatus.Sender)
outboundPeer := newRemotePeer(meta, pm, pm.actorServ, pm.logger, pm.mf, pm.signer, rw)
// insert Handlers
pm.handlerFactory.insertHandlers(outboundPeer)
go outboundPeer.runPeer()
pm.insertPeer(peerID, outboundPeer)
pm.logger.Info().Str(LogPeerID, peerID.Pretty()).Str("addr", net.ParseIP(meta.IPAddress).String()+":"+strconv.Itoa(int(meta.Port))).Msg("Outbound peer is added to peerService")
outboundPeer.metric = pm.mm.Add(peerID, rd, wt)
pm.mutex.Unlock()
addrs := pm.Peerstore().Addrs(peerID)
addrStrs := make([]string, len(addrs))
for i, addr := range addrs {
addrStrs[i] = addr.String()
}
pm.logger.Debug().Strs("addrs", addrStrs).Str(LogPeerID, outboundPeer.meta.ID.Pretty()).Msg("addresses of peer")
// peer is ready
h.doInitialSync()
// notice to p2pmanager that handshaking is finished
pm.NotifyPeerHandshake(peerID)
return true
}
func (pm *peerManager) sendGoAway(rw MsgReadWriter, msg string) {
goMsg := &types.GoAwayNotice{Message: msg}
// TODO code smell. non safe casting.
mo := pm.mf.newMsgRequestOrder(false, GoAway, goMsg).(*pbRequestOrder)
container := mo.message
rw.WriteMsg(container)
}
func (pm *peerManager) checkInPeerstore(peerID peer.ID) bool {
found := false
for _, existingPeerID := range pm.Peerstore().Peers() {
if existingPeerID == peerID {
found = true
break
}
}
return found
}
func (pm *peerManager) AddNewPeer(peer PeerMeta) {
pm.addPeerChannel <- peer
}
func (pm *peerManager) RemovePeer(peerID peer.ID) {
pm.removePeerChannel <- peerID
}
func (pm *peerManager) NotifyPeerHandshake(peerID peer.ID) {
pm.checkAndCollectPeerList(peerID)
}
func (pm *peerManager) NotifyPeerAddressReceived(metas []PeerMeta) {
pm.fillPoolChannel <- metas
}
// removePeer remove and disconnect managed remote peer connection
// It return true if peer is exist and managed by peermanager
func (pm *peerManager) removePeer(peerID peer.ID) bool {
pm.mutex.Lock()
target, ok := pm.remotePeers[peerID]
if !ok {
pm.mutex.Unlock()
return false
}
pm.deletePeer(peerID)
// No internal module access this peer anymore, but remote message can be received.
target.stop()
pm.mutex.Unlock()
// also disconnect connection
for _, existingPeerID := range pm.Peerstore().Peers() {
if existingPeerID == peerID {
for _, listener := range pm.eventListeners {
listener.OnRemovePeer(peerID)
}
pm.Network().ClosePeer(peerID)
return true
}
}
return true
}
func (pm *peerManager) | Peerstore | identifier_name |
|
peermanager.go | , meta.ID.Pretty()).Msg("Failed to handshake")
//pm.sendGoAway(rw, "Failed to handshake")
s.Close()
return false
}
pm.mutex.Lock()
inboundPeer, ok = pm.remotePeers[peerID]
if ok {
if ComparePeerID(pm.selfMeta.ID, meta.ID) <= 0 {
pm.logger.Info().Str(LogPeerID, inboundPeer.meta.ID.Pretty()).Msg("Inbound connection was already handshaked while handshaking outbound connection, and remote peer is higher priority so closing this outbound connection.")
pm.mutex.Unlock()
pm.sendGoAway(rw, "Already handshaked")
s.Close()
return true
} else {
pm.logger.Info().Str(LogPeerID, inboundPeer.meta.ID.Pretty()).Msg("Inbound connection was already handshaked while handshaking outbound connection, but local peer is higher priority so closing that inbound connection")
// disconnect lower valued connection
pm.deletePeer(meta.ID)
inboundPeer.stop()
}
}
// update peer info to remote sent infor
meta = FromPeerAddress(remoteStatus.Sender)
outboundPeer := newRemotePeer(meta, pm, pm.actorServ, pm.logger, pm.mf, pm.signer, rw)
// insert Handlers
pm.handlerFactory.insertHandlers(outboundPeer)
go outboundPeer.runPeer()
pm.insertPeer(peerID, outboundPeer)
pm.logger.Info().Str(LogPeerID, peerID.Pretty()).Str("addr", net.ParseIP(meta.IPAddress).String()+":"+strconv.Itoa(int(meta.Port))).Msg("Outbound peer is added to peerService")
outboundPeer.metric = pm.mm.Add(peerID, rd, wt)
pm.mutex.Unlock()
addrs := pm.Peerstore().Addrs(peerID)
addrStrs := make([]string, len(addrs))
for i, addr := range addrs {
addrStrs[i] = addr.String()
}
pm.logger.Debug().Strs("addrs", addrStrs).Str(LogPeerID, outboundPeer.meta.ID.Pretty()).Msg("addresses of peer")
// peer is ready
h.doInitialSync()
// notice to p2pmanager that handshaking is finished
pm.NotifyPeerHandshake(peerID)
return true
}
func (pm *peerManager) sendGoAway(rw MsgReadWriter, msg string) {
goMsg := &types.GoAwayNotice{Message: msg}
// TODO code smell. non safe casting.
mo := pm.mf.newMsgRequestOrder(false, GoAway, goMsg).(*pbRequestOrder)
container := mo.message
rw.WriteMsg(container)
}
func (pm *peerManager) checkInPeerstore(peerID peer.ID) bool {
found := false
for _, existingPeerID := range pm.Peerstore().Peers() {
if existingPeerID == peerID {
found = true
break
}
}
return found
}
func (pm *peerManager) AddNewPeer(peer PeerMeta) {
pm.addPeerChannel <- peer
}
func (pm *peerManager) RemovePeer(peerID peer.ID) {
pm.removePeerChannel <- peerID
}
func (pm *peerManager) NotifyPeerHandshake(peerID peer.ID) {
pm.checkAndCollectPeerList(peerID)
}
func (pm *peerManager) NotifyPeerAddressReceived(metas []PeerMeta) {
pm.fillPoolChannel <- metas
}
// removePeer remove and disconnect managed remote peer connection
// It return true if peer is exist and managed by peermanager
func (pm *peerManager) removePeer(peerID peer.ID) bool {
pm.mutex.Lock()
target, ok := pm.remotePeers[peerID]
if !ok {
pm.mutex.Unlock()
return false
}
pm.deletePeer(peerID)
// No internal module access this peer anymore, but remote message can be received.
target.stop()
pm.mutex.Unlock()
// also disconnect connection
for _, existingPeerID := range pm.Peerstore().Peers() {
if existingPeerID == peerID {
for _, listener := range pm.eventListeners {
listener.OnRemovePeer(peerID)
}
pm.Network().ClosePeer(peerID)
return true
}
}
return true
}
func (pm *peerManager) Peerstore() pstore.Peerstore {
return pm.Host.Peerstore()
}
func (pm *peerManager) startListener() {
var err error
listens := make([]ma.Multiaddr, 0, 2)
// FIXME: should also support ip6 later
listen, err := ma.NewMultiaddr(fmt.Sprintf("/ip4/%s/tcp/%d", pm.bindAddress, pm.bindPort))
if err != nil {
panic("Can't estabilish listening address: " + err.Error())
}
listens = append(listens, listen)
peerStore := pstore.NewPeerstore(pstoremem.NewKeyBook(), pstoremem.NewAddrBook(), pstoremem.NewPeerMetadata())
newHost, err := libp2p.New(context.Background(), libp2p.Identity(pm.privateKey), libp2p.Peerstore(peerStore), libp2p.ListenAddrs(listens...))
if err != nil {
pm.logger.Fatal().Err(err).Str("addr", listen.String()).Msg("Couldn't listen from")
panic(err.Error())
}
pm.logger.Info().Str("pid", pm.SelfNodeID().Pretty()).Str("addr[0]", listens[0].String()).
Msg("Set self node's pid, and listening for connections")
pm.Host = newHost
pm.SetStreamHandler(aergoP2PSub, pm.onHandshake)
}
func (pm *peerManager) onHandshake(s inet.Stream) {
peerID := s.Conn().RemotePeer()
h := newHandshaker(pm, pm.actorServ, pm.logger, peerID)
rd := metric.NewReader(s)
wt := metric.NewWriter(s)
rw, statusMsg, err := h.handshakeInboundPeer(rd, wt)
if err != nil {
pm.logger.Info().Str(LogPeerID, peerID.Pretty()).Err(err).Msg("fail to handshake")
s.Close()
return
}
// TODO: check status
meta := FromPeerAddress(statusMsg.Sender)
// try Add peer
if inboundPeer, success := pm.tryAddInboundPeer(meta, rw); !success {
// failed to add
pm.sendGoAway(rw, "Concurrent handshake")
s.Close()
return
} else {
inboundPeer.metric = pm.mm.Add(peerID, rd, wt)
}
h.doInitialSync()
// notice to p2pmanager that handshaking is finished
pm.NotifyPeerHandshake(peerID)
}
func (pm *peerManager) tryAddInboundPeer(meta PeerMeta, rw MsgReadWriter) (*remotePeerImpl, bool) {
pm.mutex.Lock()
defer pm.mutex.Unlock()
peerID := meta.ID
outboundPeer, found := pm.remotePeers[peerID]
if found {
if ComparePeerID(pm.selfMeta.ID, meta.ID) <= 0 {
pm.logger.Info().Str(LogPeerID, peerID.Pretty()).Msg("Outbound connection was already handshaked while handshaking inbound connection, and remote peer is higher priority so closing that outbound connection.")
pm.sendGoAway(rw, "Already handshaked")
pm.deletePeer(meta.ID)
outboundPeer.stop()
} else {
pm.logger.Info().Str(LogPeerID, peerID.Pretty()).Msg("Outbound connection was already handshaked while handshaking inbound connection, but local peer is higher priority and closing this inbound connection.")
// disconnect lower valued connection
return nil, false
}
}
inboundPeer := newRemotePeer(meta, pm, pm.actorServ, pm.logger, pm.mf, pm.signer, rw)
pm.handlerFactory.insertHandlers(inboundPeer)
go inboundPeer.runPeer()
pm.insertPeer(peerID, inboundPeer)
peerAddr := meta.ToPeerAddress()
addrs := pm.Peerstore().Addrs(peerID)
addrStrs := make([]string, len(addrs))
for i, addr := range addrs {
addrStrs[i] = addr.String()
}
pm.logger.Debug().Strs("addrs", addrStrs).Str(LogPeerID, inboundPeer.meta.ID.Pretty()).Msg("addresses of peer")
pm.logger.Info().Str(LogPeerID, peerID.Pretty()).Str("addr", getIP(&peerAddr).String()+":"+strconv.Itoa(int(peerAddr.Port))).Msg("Inbound peer is added to peerService")
return inboundPeer, true
}
func (pm *peerManager) Start() error {
pm.run()
//pm.conf.NPAddPeers
return nil
}
func (pm *peerManager) Stop() error {
// TODO stop service
// close(pm.addPeerChannel)
// close(pm.removePeerChannel)
pm.finishChannel <- struct{}{}
return nil
}
func (pm *peerManager) GetName() string {
return "p2p service"
}
func (pm *peerManager) checkAndCollectPeerListFromAll() {
if pm.hasEnoughPeers() {
return
}
for _, remotePeer := range pm.remotePeers {
pm.actorServ.SendRequest(message.P2PSvc, &message.GetAddressesMsg{ToWhom: remotePeer.meta.ID, Size: 20, Offset: 0})
}
}
func (pm *peerManager) checkAndCollectPeerList(ID peer.ID) { | if pm.hasEnoughPeers() {
return | random_line_split |
|
Config.ts | I don't know his lives . // I don't know where he~his lives .
He get to cleaned his son . // He got his~his son~son to:O clean:O the~ room~ .
We wrote down the number . // We wrote the number down~down .
`
.trim()
.split(/\n\n+/gm)
.map(line => ex.apply({}, line.split('//').map(side => side.trim()) as [string, string]))
const order_changing_labels: Record<string, true> = {
'S-adv': true,
'S-finV': true,
'S-WO': true,
WO: true,
INV: true,
OINV: true,
}
export const label_args: Record<string, number> = {
/*age_string: 1,*/
}
export type TaxonomyGroup = {
group: string
entries: {
label: string
desc: string
}[]
}
export type Taxonomy = TaxonomyGroup[]
const extra = 'gen def pl foreign'.split(' ')
const temporary = 'OBS! Cit-FL Com!'.split(' ')
const digits = /^\d+$/
/** An ordered set of label categories. */
export enum LabelOrder {
BASE,
NUM,
EXTRA,
TEMP,
}
/** Maps a label to a category in LabelOrder. */
export function label_order(label: string): LabelOrder {
if (temporary.includes(label)) {
return LabelOrder.TEMP
} else if (extra.includes(label)) {
return LabelOrder.EXTRA
} else if (digits.test(label)) {
return LabelOrder.NUM
} else {
return LabelOrder.BASE
}
}
/** Sorting function for labels. */
// Sort first by taxonomy, then label type, and finally alphabetically.
export const label_sort: Comparator<string> = chain_cmps(
mkcmp(label_taxonomy),
mkcmp(label_order),
cmp_order
)
const anonymization: Taxonomy = [
{
group: 'Morphology',
entries: [
{label: 'gen', desc: 'genitive'},
{label: 'def', desc: 'definite'},
{label: 'pl', desc: 'plural'},
],
},
{
group: 'Names',
entries: [
{label: 'firstname_male', desc: ''},
{label: 'firstname_female', desc: ''},
{label: 'firstname_unknown', desc: ''},
{label: 'initials', desc: ''},
{label: 'middlename', desc: ''},
{label: 'surname', desc: ''},
],
},
{
group: 'Geographic data',
entries: [
{label: 'foreign', desc: ''},
{label: 'area', desc: ''},
{label: 'city', desc: 'city including villages'},
{label: 'country', desc: 'except Sweden'},
{label: 'geo', desc: 'forest, lake, mountain, etc'},
{label: 'place', desc: ''},
{label: 'region', desc: ''},
{label: 'street_nr', desc: 'street number'},
{label: 'zip_code', desc: ''},
],
},
{
group: 'Institutions',
entries: [
{label: 'school', desc: ''},
{label: 'work', desc: ''},
{label: 'other_institution', desc: ''},
],
},
{
group: 'Transportation',
entries: [
{label: 'transport_name', desc: 'bus, metro, tram, train, express'},
{label: 'transport_nr', desc: 'number, color'},
],
},
{
group: 'Age',
entries: [{label: 'age_digits', desc: ''}, {label: 'age_string', desc: ''}],
},
{
group: 'Dates',
entries: [
{label: 'date_digits', desc: 'numerical date represenation, delimiters are retained'},
{label: 'day', desc: ''},
{label: 'month_digit', desc: ''},
{label: 'month_word', desc: ''},
{label: 'year', desc: ''},
],
},
{
group: 'Misc',
entries: [
{label: 'account_nr', desc: ''},
{label: 'email', desc: ''},
{label: 'extra', desc: ''},
{label: 'license_nr', desc: ''},
{label: 'other_nr_seq', desc: 'a sequence of numbers'},
{label: 'phone_nr', desc: ''},
{label: 'personid_nr', desc: ''},
{label: 'url', desc: ''},
],
},
{
group: 'Mark',
entries: [
{label: 'edu', desc: 'education, courses'},
{label: 'fam', desc: 'family members'},
{label: 'prof', desc: 'profession'},
{label: 'sensitive', desc: ''},
],
},
{
group: 'Other',
entries: [
{label: 'Cit-FL', desc: 'Citation for a language'},
{label: 'Com!', desc: 'Comment'},
{label: 'OBS!', desc: 'Attention'},
],
},
]
export const normalization: Taxonomy = [
{
group: 'Other',
entries: [
{
label: 'Cit-FL',
desc: 'Citation for a language'
},
{
label: 'Com!',
desc: 'Comment'
},
{
label: 'OBS!',
desc: 'Attention'
},
{
label: 'X',
desc: 'Impossible to interpret the writer’s intention',
},
],
},
]
// Julia's updated taxonomy 19 April 2018
export const correctannot: Taxonomy = [
{
group: 'Orthographic',
entries: [
{
label: 'O',
desc: 'Spelling',
},
{
label: 'O-Cap',
desc: 'Upper/lower case',
},
{
label: 'O-Comp',
desc: 'Spaces and hyphens between words',
},
],
},
{
group: 'Lexical',
entries: [
{
label: 'L-Der',
desc: 'Word formation (derivation and compounding)',
},
{
label: 'L-FL',
desc: 'Non-Swedish word corrected to Swedish word',
},
{
label: 'L-Ref',
desc: 'Choice of anaphoric expression',
},
{
label: 'L-W',
desc:
'Wrong word or phrase, other',
},
],
},
{
group: 'Morphological',
entries: [
{
label: 'M-Adj/adv',
desc: 'Adjective form of word corrected to adverb form',
},
{
label: 'M-Case',
desc: 'Nominative vs genitive/accusative',
},
{label: 'M-Def', desc: 'Definiteness: articles; forms of nouns and adjectives'},
{label: 'M-F', desc: 'Grammatical category kept, form changed'},
{label: 'M-Gend', desc: 'Gender'},
{label: 'M-Num', desc: 'Number'},
{
label: 'M-Other',
desc:
'Other morphological corrections, including change between different comparational forms of adjectives',
},
{label: 'M-Verb', desc: 'Verb forms; use of ha, komma and skola auxiliaries'},
],
},
{
group: 'Punctuation',
entries: [
{
label: 'P-M',
desc: 'Punctuation missing (added)',
},
{
label: 'P-R',
desc: 'Punctuation redundant (removed)',
},
{
label: 'P-Sent',
desc: 'Sentence segmentation',
},
{
label: 'P-W',
desc: 'Wrong punctuation',
},
],
},
{
group: 'Syntactical',
entries: [
{
label: 'S-Adv',
desc: 'Adverbial placement',
},
{
label: 'S-Comp',
desc: 'Compound vs multi-word expression, and other restructuring of the same lexical morphemes within a phrase',
},
{
label: 'S-Clause',
desc: 'Change of basic clause structure: syntactic function of components, hierarchical clause structure',
},
{
label: 'S-Ext',
desc: 'Extensive and complex correction',
},
{
label: 'S-FinV',
desc: 'Finite verb placement',
},
{
label: 'S-M',
desc:
'Word missing (added)',
},
{
label: 'S-Msubj',
desc: 'Subject missing (added)',
},
{
label: 'S-Other',
desc:
'Other syntactical correction',
},
{
label: 'S-R',
desc: 'Word redundant (removed)',
},
{
label: 'S-Type',
desc: 'Change of phrase |
Their was a problem yesteray . // There was a problem yesterday .
| random_line_split |
|
Config.ts | , mountain, etc'},
{label: 'place', desc: ''},
{label: 'region', desc: ''},
{label: 'street_nr', desc: 'street number'},
{label: 'zip_code', desc: ''},
],
},
{
group: 'Institutions',
entries: [
{label: 'school', desc: ''},
{label: 'work', desc: ''},
{label: 'other_institution', desc: ''},
],
},
{
group: 'Transportation',
entries: [
{label: 'transport_name', desc: 'bus, metro, tram, train, express'},
{label: 'transport_nr', desc: 'number, color'},
],
},
{
group: 'Age',
entries: [{label: 'age_digits', desc: ''}, {label: 'age_string', desc: ''}],
},
{
group: 'Dates',
entries: [
{label: 'date_digits', desc: 'numerical date represenation, delimiters are retained'},
{label: 'day', desc: ''},
{label: 'month_digit', desc: ''},
{label: 'month_word', desc: ''},
{label: 'year', desc: ''},
],
},
{
group: 'Misc',
entries: [
{label: 'account_nr', desc: ''},
{label: 'email', desc: ''},
{label: 'extra', desc: ''},
{label: 'license_nr', desc: ''},
{label: 'other_nr_seq', desc: 'a sequence of numbers'},
{label: 'phone_nr', desc: ''},
{label: 'personid_nr', desc: ''},
{label: 'url', desc: ''},
],
},
{
group: 'Mark',
entries: [
{label: 'edu', desc: 'education, courses'},
{label: 'fam', desc: 'family members'},
{label: 'prof', desc: 'profession'},
{label: 'sensitive', desc: ''},
],
},
{
group: 'Other',
entries: [
{label: 'Cit-FL', desc: 'Citation for a language'},
{label: 'Com!', desc: 'Comment'},
{label: 'OBS!', desc: 'Attention'},
],
},
]
export const normalization: Taxonomy = [
{
group: 'Other',
entries: [
{
label: 'Cit-FL',
desc: 'Citation for a language'
},
{
label: 'Com!',
desc: 'Comment'
},
{
label: 'OBS!',
desc: 'Attention'
},
{
label: 'X',
desc: 'Impossible to interpret the writer’s intention',
},
],
},
]
// Julia's updated taxonomy 19 April 2018
export const correctannot: Taxonomy = [
{
group: 'Orthographic',
entries: [
{
label: 'O',
desc: 'Spelling',
},
{
label: 'O-Cap',
desc: 'Upper/lower case',
},
{
label: 'O-Comp',
desc: 'Spaces and hyphens between words',
},
],
},
{
group: 'Lexical',
entries: [
{
label: 'L-Der',
desc: 'Word formation (derivation and compounding)',
},
{
label: 'L-FL',
desc: 'Non-Swedish word corrected to Swedish word',
},
{
label: 'L-Ref',
desc: 'Choice of anaphoric expression',
},
{
label: 'L-W',
desc:
'Wrong word or phrase, other',
},
],
},
{
group: 'Morphological',
entries: [
{
label: 'M-Adj/adv',
desc: 'Adjective form of word corrected to adverb form',
},
{
label: 'M-Case',
desc: 'Nominative vs genitive/accusative',
},
{label: 'M-Def', desc: 'Definiteness: articles; forms of nouns and adjectives'},
{label: 'M-F', desc: 'Grammatical category kept, form changed'},
{label: 'M-Gend', desc: 'Gender'},
{label: 'M-Num', desc: 'Number'},
{
label: 'M-Other',
desc:
'Other morphological corrections, including change between different comparational forms of adjectives',
},
{label: 'M-Verb', desc: 'Verb forms; use of ha, komma and skola auxiliaries'},
],
},
{
group: 'Punctuation',
entries: [
{
label: 'P-M',
desc: 'Punctuation missing (added)',
},
{
label: 'P-R',
desc: 'Punctuation redundant (removed)',
},
{
label: 'P-Sent',
desc: 'Sentence segmentation',
},
{
label: 'P-W',
desc: 'Wrong punctuation',
},
],
},
{
group: 'Syntactical',
entries: [
{
label: 'S-Adv',
desc: 'Adverbial placement',
},
{
label: 'S-Comp',
desc: 'Compound vs multi-word expression, and other restructuring of the same lexical morphemes within a phrase',
},
{
label: 'S-Clause',
desc: 'Change of basic clause structure: syntactic function of components, hierarchical clause structure',
},
{
label: 'S-Ext',
desc: 'Extensive and complex correction',
},
{
label: 'S-FinV',
desc: 'Finite verb placement',
},
{
label: 'S-M',
desc:
'Word missing (added)',
},
{
label: 'S-Msubj',
desc: 'Subject missing (added)',
},
{
label: 'S-Other',
desc:
'Other syntactical correction',
},
{
label: 'S-R',
desc: 'Word redundant (removed)',
},
{
label: 'S-Type',
desc: 'Change of phrase type/part of speech',
},
{
label: 'S-WO',
desc: 'Word order, other',
},
],
},
{
group: 'Other',
entries: [
{
label: 'C',
desc: 'Consistency correction, necessitated by other correction',
},
{
label: 'Cit-FL',
desc: 'Non-Swedish word kept, i.e. not corrected',
},
{
label: 'Com!',
desc: 'Comments for the corpus user'
},
{
label: 'OBS!',
desc: 'Internal and temporary comments for the annotators'
},
{
label: 'Unid',
desc: 'Unidentified correction',
},
{
label: 'X',
desc: 'Unintelligible string',
},
],
},
]
function doc_url(title: string): string {
return 'https://spraakbanken.github.io/swell-project/' + title
}
const docs: Record<string, Record<string, string>> = {
anonymization: {
'pseudonymization guidelines': doc_url('Pseudonymization_guidelines'),
},
normalization: {
'normalization guidelines': doc_url('Normalization_guidelines'),
},
correctannot: {
'annotation guidelines': doc_url('Correction-annotation_guidelines'),
},
}
export const config = {
order_changing_labels,
examples,
image_ws_url,
pseuws_url,
taxonomy: {anonymization, normalization, correctannot},
docs,
}
/** What group does this label belong to?
(label_group('country') as TaxonomyGroup).group // => 'Geographic data'
label_group('quux') // => undefined
*/
export function label_group(label: string): TaxonomyGroup | undefined {
return config.taxonomy.anonymization.find(
group => !!group.entries.find(entry => entry.label == label)
)
}
export interface TaxonomyFind {
taxonomy: string
group: string
entry: {label: string; desc: string}
}
export function find_label(label: string): TaxonomyFind | undefined {
const order = label_order(label)
if (order === LabelOrder.NUM) {
return {taxonomy: 'anonymization', group: 'Number', entry: {label, desc: 'number'}}
}
if (order === LabelOrder.TEMP) {
return undefined
}
for (let taxonomy in config.taxonomy) {
for (let group of (config.taxonomy as {[mode: string]: Taxonomy})[taxonomy]) {
let entry = group.entries.find(entry => entry.label == label)
if (entry !== undefined) return {taxonomy, group: group.group, entry}
}
}
}
/** Get the taxonomy domain (editor mode) of a label. */
export function label_taxonomy(label: string): string | null {
return find_label(label) ? find_label(label)!.taxonomy : null
}
/** Does the named taxonomy include the given label? */
export function ta | xonomy_has_label(t | identifier_name |
|
Config.ts | desc: string
}[]
}
export type Taxonomy = TaxonomyGroup[]
const extra = 'gen def pl foreign'.split(' ')
const temporary = 'OBS! Cit-FL Com!'.split(' ')
const digits = /^\d+$/
/** An ordered set of label categories. */
export enum LabelOrder {
BASE,
NUM,
EXTRA,
TEMP,
}
/** Maps a label to a category in LabelOrder. */
export function label_order(label: string): LabelOrder {
if (temporary.includes(label)) {
return LabelOrder.TEMP
} else if (extra.includes(label)) {
return LabelOrder.EXTRA
} else if (digits.test(label)) {
return LabelOrder.NUM
} else {
return LabelOrder.BASE
}
}
/** Sorting function for labels. */
// Sort first by taxonomy, then label type, and finally alphabetically.
export const label_sort: Comparator<string> = chain_cmps(
mkcmp(label_taxonomy),
mkcmp(label_order),
cmp_order
)
const anonymization: Taxonomy = [
{
group: 'Morphology',
entries: [
{label: 'gen', desc: 'genitive'},
{label: 'def', desc: 'definite'},
{label: 'pl', desc: 'plural'},
],
},
{
group: 'Names',
entries: [
{label: 'firstname_male', desc: ''},
{label: 'firstname_female', desc: ''},
{label: 'firstname_unknown', desc: ''},
{label: 'initials', desc: ''},
{label: 'middlename', desc: ''},
{label: 'surname', desc: ''},
],
},
{
group: 'Geographic data',
entries: [
{label: 'foreign', desc: ''},
{label: 'area', desc: ''},
{label: 'city', desc: 'city including villages'},
{label: 'country', desc: 'except Sweden'},
{label: 'geo', desc: 'forest, lake, mountain, etc'},
{label: 'place', desc: ''},
{label: 'region', desc: ''},
{label: 'street_nr', desc: 'street number'},
{label: 'zip_code', desc: ''},
],
},
{
group: 'Institutions',
entries: [
{label: 'school', desc: ''},
{label: 'work', desc: ''},
{label: 'other_institution', desc: ''},
],
},
{
group: 'Transportation',
entries: [
{label: 'transport_name', desc: 'bus, metro, tram, train, express'},
{label: 'transport_nr', desc: 'number, color'},
],
},
{
group: 'Age',
entries: [{label: 'age_digits', desc: ''}, {label: 'age_string', desc: ''}],
},
{
group: 'Dates',
entries: [
{label: 'date_digits', desc: 'numerical date represenation, delimiters are retained'},
{label: 'day', desc: ''},
{label: 'month_digit', desc: ''},
{label: 'month_word', desc: ''},
{label: 'year', desc: ''},
],
},
{
group: 'Misc',
entries: [
{label: 'account_nr', desc: ''},
{label: 'email', desc: ''},
{label: 'extra', desc: ''},
{label: 'license_nr', desc: ''},
{label: 'other_nr_seq', desc: 'a sequence of numbers'},
{label: 'phone_nr', desc: ''},
{label: 'personid_nr', desc: ''},
{label: 'url', desc: ''},
],
},
{
group: 'Mark',
entries: [
{label: 'edu', desc: 'education, courses'},
{label: 'fam', desc: 'family members'},
{label: 'prof', desc: 'profession'},
{label: 'sensitive', desc: ''},
],
},
{
group: 'Other',
entries: [
{label: 'Cit-FL', desc: 'Citation for a language'},
{label: 'Com!', desc: 'Comment'},
{label: 'OBS!', desc: 'Attention'},
],
},
]
export const normalization: Taxonomy = [
{
group: 'Other',
entries: [
{
label: 'Cit-FL',
desc: 'Citation for a language'
},
{
label: 'Com!',
desc: 'Comment'
},
{
label: 'OBS!',
desc: 'Attention'
},
{
label: 'X',
desc: 'Impossible to interpret the writer’s intention',
},
],
},
]
// Julia's updated taxonomy 19 April 2018
export const correctannot: Taxonomy = [
{
group: 'Orthographic',
entries: [
{
label: 'O',
desc: 'Spelling',
},
{
label: 'O-Cap',
desc: 'Upper/lower case',
},
{
label: 'O-Comp',
desc: 'Spaces and hyphens between words',
},
],
},
{
group: 'Lexical',
entries: [
{
label: 'L-Der',
desc: 'Word formation (derivation and compounding)',
},
{
label: 'L-FL',
desc: 'Non-Swedish word corrected to Swedish word',
},
{
label: 'L-Ref',
desc: 'Choice of anaphoric expression',
},
{
label: 'L-W',
desc:
'Wrong word or phrase, other',
},
],
},
{
group: 'Morphological',
entries: [
{
label: 'M-Adj/adv',
desc: 'Adjective form of word corrected to adverb form',
},
{
label: 'M-Case',
desc: 'Nominative vs genitive/accusative',
},
{label: 'M-Def', desc: 'Definiteness: articles; forms of nouns and adjectives'},
{label: 'M-F', desc: 'Grammatical category kept, form changed'},
{label: 'M-Gend', desc: 'Gender'},
{label: 'M-Num', desc: 'Number'},
{
label: 'M-Other',
desc:
'Other morphological corrections, including change between different comparational forms of adjectives',
},
{label: 'M-Verb', desc: 'Verb forms; use of ha, komma and skola auxiliaries'},
],
},
{
group: 'Punctuation',
entries: [
{
label: 'P-M',
desc: 'Punctuation missing (added)',
},
{
label: 'P-R',
desc: 'Punctuation redundant (removed)',
},
{
label: 'P-Sent',
desc: 'Sentence segmentation',
},
{
label: 'P-W',
desc: 'Wrong punctuation',
},
],
},
{
group: 'Syntactical',
entries: [
{
label: 'S-Adv',
desc: 'Adverbial placement',
},
{
label: 'S-Comp',
desc: 'Compound vs multi-word expression, and other restructuring of the same lexical morphemes within a phrase',
},
{
label: 'S-Clause',
desc: 'Change of basic clause structure: syntactic function of components, hierarchical clause structure',
},
{
label: 'S-Ext',
desc: 'Extensive and complex correction',
},
{
label: 'S-FinV',
desc: 'Finite verb placement',
},
{
label: 'S-M',
desc:
'Word missing (added)',
},
{
label: 'S-Msubj',
desc: 'Subject missing (added)',
},
{
label: 'S-Other',
desc:
'Other syntactical correction',
},
{
label: 'S-R',
desc: 'Word redundant (removed)',
},
{
label: 'S-Type',
desc: 'Change of phrase type/part of speech',
},
{
label: 'S-WO',
desc: 'Word order, other',
},
],
},
{
group: 'Other',
entries: [
{
label: 'C',
desc: 'Consistency correction, necessitated by other correction',
},
{
label: 'Cit-FL',
desc: 'Non-Swedish word kept, i.e. not corrected',
},
{
label: 'Com!',
desc: 'Comments for the corpus user'
},
{
label: 'OBS!',
desc: 'Internal and temporary comments for the annotators'
},
{
label: 'Unid',
desc: 'Unidentified correction',
},
{
label: 'X',
desc: 'Unintelligible string',
},
],
},
]
function doc_url(title: string): string {
| return 'https://spraakbanken.github.io/swell-project/' + title
}
| identifier_body |
|
filesystem.rs | = args[0];
let distributed_filename = args[1];
// Figure out who I am giving this file to
let dest_ids = gen_file_owners(&distributed_filename)?;
// Gossip who has the file now
sender.send(
SendableOperation::for_successors(Box::new(NewFileOwnersOperation {
distributed_filename: distributed_filename.to_string(),
new_owners: dest_ids
.iter()
.map(|x| x.to_string())
.collect::<HashSet<_>>()
}))
)?;
// Send them the file
async_std::task::block_on(send_file_to_all(local_path.to_string(),
distributed_filename.to_string(),
&dest_ids))?;
Ok(())
}
pub fn ls(args: Vec<&str>) -> BoxedErrorResult<()> {
check_joined()?;
let invalid_args: BoxedErrorResult<()> = Err("Usage: ls [distributed_filename]".into());
match args.len() {
0 => {
// All
print_file_owners(None, true)?;
Ok(())
},
1 => {
// Just File
let distributed_filename = args[0];
print_file_owners(Some(distributed_filename), false)?;
Ok(())
},
_ => invalid_args
}
}
// TODO: You wrote this very late - maybe fix
fn print_file_owners(maybe_distributed_filename: Option<&str>, full: bool) -> BoxedErrorResult<()> {
let all_file_owners = globals::ALL_FILE_OWNERS.read();
match (maybe_distributed_filename, full) {
(Some(_), true) => {
Err("Cannot set distributed_filename and full to true when printing owners".into())
},
(Some(distributed_filename), false) => {
// Print the files owners
match all_file_owners.get(distributed_filename) {
Some(owners) => {
println!("{:?}", owners);
},
None => {
// A little unoptimal - change if above format changes
println!("{{}}");
}
}
Ok(())
},
(None, true) => {
// Print the whole map
println!("{:?}", *all_file_owners);
Ok(())
},
(None, false) => {
Err("Cannot print owners of nonexistant distributed_filename with full set to false".into())
}
}
}
async fn get_distributed_file(distributed_filename: String, local_path: String) -> BoxedErrorResult<()> {
// TODO: Find owners
let operation = SendableOperation::for_owners(&distributed_filename, Box::new(GetOperation {
distributed_filename: distributed_filename.clone(),
local_path: local_path
}));
let mut streams = operation
.write_all_tcp_async()
.await?;
// TODO: Redo whatever tf going on here
match streams.len() {
0 => Err(format!("No owners found for file {}", distributed_filename).into()),
_ => {
let (result, source) = streams[0]
.try_read_operation()
.await?;
result.execute(source)?;
Ok(())
}
}
}
async fn read_file_to_buf(local_path: &String) -> BoxedErrorResult<Vec<u8>> {
let mut data_buf: Vec<u8> = Vec::new();
let mut file = async_std::fs::File::open(&local_path).await?;
file.read_to_end(&mut data_buf).await?;
Ok(data_buf)
}
async fn send_file_to_all(local_path: String, distributed_filename: String, dest_ids: &Vec<String>) ->
BoxedErrorResult<()> {
let data_buf = read_file_to_buf(&local_path).await?;
let operation = SendableOperation::for_id_list(dest_ids.clone(), Box::new(SendFileOperation {
filename: distributed_filename,
data: data_buf,
is_distributed: true
}));
operation.write_all_tcp_async().await?;
Ok(())
}
pub async fn file_server<'a>(_sender: &'a OperationSender) -> BoxedErrorResult<()> {
let server = globals::SERVER_SOCKET.read();
let mut incoming = server.incoming();
while let Some(stream) = incoming.next().await {
let connection = stream?;
log(format!("Handling connection from {:?}", connection.peer_addr()));
spawn(handle_connection(connection));
}
Ok(())
}
async fn handle_connection(mut connection: async_std::net::TcpStream) -> BoxedErrorResult<()> {
let (operation, source) = connection.try_read_operation().await?;
// TODO: Think about what standard we want with these
let _generated_operations = operation.execute(source)?;
Ok(())
}
// Helpers
fn gen_file_owners(filename: &str) -> BoxedErrorResult<Vec<String>> {
let file_idx = filename.easyhash();
heartbeat::gen_neighbor_list_from(file_idx as i32, 1, constants::NUM_OWNERS, true)
}
// TODO: This function makes the entire system assume there are always at least two nodes in the system
// and the file must have an owner or else the operation will not work correctly. This is fine for now
// but it is worth improving sooner rather than later (make distinct Error types to differentiate, etc).
fn gen_new_file_owner(filename: &str) -> BoxedErrorResult<String> {
match globals::ALL_FILE_OWNERS.read().get(filename) {
Some(owners) => {
let potential_owners = gen_file_owners(filename)?;
for potential_owner in &potential_owners {
if !owners.contains(potential_owner) {
return Ok(potential_owner.clone());
}
}
Err(format!("No new owners available for file {}", filename).into())
},
None => Err(format!("No owner found for file {}", filename).into())
}
}
fn | (filename: &String) -> String {
format!("{}/{}", constants::DATA_DIR, filename)
}
// Returns messages to be gossiped
pub fn handle_failed_node(failed_id: &String) -> BoxedErrorResult<Vec<SendableOperation>> {
match heartbeat::is_master() {
true => {
let mut generated_operations: Vec<SendableOperation> = Vec::new();
let myself_source = Source::myself();
// Find files they owned
let mut lost_files: HashSet<String> = HashSet::new();
for (distributed_filename, owners) in globals::ALL_FILE_OWNERS.read().iter() {
if owners.contains(failed_id) {
lost_files.insert(distributed_filename.clone());
}
}
log("Found lost files".to_string());
// Send that they no longer own those files
// Separate operation so that this acts like a confirmation to fully forget about the node from
// the master. Can be used with a delay later if you want more error resistance.
let lost_file_operation = LostFilesOperation {
failed_owner: failed_id.clone(),
lost_files: lost_files.clone()
};
generated_operations.append(&mut lost_file_operation.execute(myself_source.clone())?);
log("Executed lost files operation locally".to_string());
// Gen new owners of the file and propagate
let mut new_owners: HashMap<String, HashSet<String>> = HashMap::new();
for lost_file in &lost_files {
let new_owner = gen_new_file_owner(&lost_file)?;
// TODO: Maybe optimize this into one fat packet - probably a new operation?
let new_owner_operation = NewFileOwnersOperation {
distributed_filename: lost_file.clone(),
new_owners: vec![new_owner].iter().map(|x| x.to_string()).collect()
};
generated_operations.append(&mut new_owner_operation.execute(myself_source.clone())?);
}
log("Executed all new_owner operations locally".to_string());
Ok(generated_operations)
},
false => {
Ok(vec![])
}
}
}
// Operations
#[derive(Serialize, Deserialize, Debug, Clone)]
pub struct GetOperation {
pub distributed_filename: String,
pub local_path: String
}
#[derive(Serialize, Deserialize, Debug, Clone)]
pub struct NewFileOwnersOperation {
pub distributed_filename: String,
pub new_owners: HashSet<String>
}
#[derive(Serialize, Deserialize, Clone)]
pub struct SendFileOperation {
pub filename: String,
pub data: Vec<u8>,
pub is_distributed: bool
}
#[derive(Serialize, Deserialize, Debug, Clone)]
pub struct LostFilesOperation {
pub failed_owner: String,
pub lost_files: HashSet<String>
}
// Trait Impls
impl OperationWriteExecute for GetOperation {
fn to_bytes(&self) -> BoxedErrorResult<Vec<u8>> {
Ok(create_buf(&self, str_to_vec("GET ")))
}
fn execute(&self, source: Source) -> BoxedErrorResult<Vec<SendableOperation>> {
let local_path = distributed_file_path(&self.distributed_filename);
let data_buf = async_std::task::block_on(read_file_to_buf(&local_path))?;
let operation = SendableOperation::for_single_tcp_stream(
TryInto::<async_std::net::TcpStream>::try_into(source)?,
Box::new(SendFileOperation {
filename: self.local_path.clone(),
data: data_buf,
is_distributed: false
}));
async_std::task::block_on(operation.write_all_tcp_async());
Ok(vec![])
}
fn to_string(&self) -> String { format!("{:?}", self) }
}
impl OperationWriteExecute for NewFileOwnersOperation {
fn to_bytes(&self) | distributed_file_path | identifier_name |
filesystem.rs | = args[0];
let distributed_filename = args[1];
// Figure out who I am giving this file to
let dest_ids = gen_file_owners(&distributed_filename)?;
// Gossip who has the file now
sender.send(
SendableOperation::for_successors(Box::new(NewFileOwnersOperation {
distributed_filename: distributed_filename.to_string(),
new_owners: dest_ids
.iter()
.map(|x| x.to_string())
.collect::<HashSet<_>>()
}))
)?;
// Send them the file
async_std::task::block_on(send_file_to_all(local_path.to_string(),
distributed_filename.to_string(),
&dest_ids))?;
Ok(())
}
pub fn ls(args: Vec<&str>) -> BoxedErrorResult<()> {
check_joined()?;
let invalid_args: BoxedErrorResult<()> = Err("Usage: ls [distributed_filename]".into());
match args.len() {
0 => {
// All
print_file_owners(None, true)?;
Ok(())
},
1 => {
// Just File
let distributed_filename = args[0];
print_file_owners(Some(distributed_filename), false)?;
Ok(())
},
_ => invalid_args
}
}
// TODO: You wrote this very late - maybe fix
fn print_file_owners(maybe_distributed_filename: Option<&str>, full: bool) -> BoxedErrorResult<()> {
let all_file_owners = globals::ALL_FILE_OWNERS.read();
match (maybe_distributed_filename, full) {
(Some(_), true) => {
Err("Cannot set distributed_filename and full to true when printing owners".into())
},
(Some(distributed_filename), false) => {
// Print the files owners
match all_file_owners.get(distributed_filename) {
Some(owners) => {
println!("{:?}", owners);
},
None => {
// A little unoptimal - change if above format changes
println!("{{}}");
}
}
Ok(())
},
(None, true) => {
// Print the whole map
println!("{:?}", *all_file_owners);
Ok(())
},
(None, false) => {
Err("Cannot print owners of nonexistant distributed_filename with full set to false".into())
}
}
}
async fn get_distributed_file(distributed_filename: String, local_path: String) -> BoxedErrorResult<()> {
// TODO: Find owners
let operation = SendableOperation::for_owners(&distributed_filename, Box::new(GetOperation {
distributed_filename: distributed_filename.clone(),
local_path: local_path
}));
let mut streams = operation
.write_all_tcp_async()
.await?;
// TODO: Redo whatever tf going on here
match streams.len() {
0 => Err(format!("No owners found for file {}", distributed_filename).into()),
_ => {
let (result, source) = streams[0]
.try_read_operation()
.await?;
result.execute(source)?;
Ok(())
}
}
}
async fn read_file_to_buf(local_path: &String) -> BoxedErrorResult<Vec<u8>> {
let mut data_buf: Vec<u8> = Vec::new();
let mut file = async_std::fs::File::open(&local_path).await?;
file.read_to_end(&mut data_buf).await?;
Ok(data_buf)
}
async fn send_file_to_all(local_path: String, distributed_filename: String, dest_ids: &Vec<String>) ->
BoxedErrorResult<()> {
let data_buf = read_file_to_buf(&local_path).await?;
let operation = SendableOperation::for_id_list(dest_ids.clone(), Box::new(SendFileOperation {
filename: distributed_filename,
data: data_buf,
is_distributed: true
}));
operation.write_all_tcp_async().await?;
Ok(())
}
pub async fn file_server<'a>(_sender: &'a OperationSender) -> BoxedErrorResult<()> {
let server = globals::SERVER_SOCKET.read();
let mut incoming = server.incoming();
while let Some(stream) = incoming.next().await {
let connection = stream?;
log(format!("Handling connection from {:?}", connection.peer_addr()));
spawn(handle_connection(connection));
}
Ok(())
}
async fn handle_connection(mut connection: async_std::net::TcpStream) -> BoxedErrorResult<()> {
let (operation, source) = connection.try_read_operation().await?;
// TODO: Think about what standard we want with these
let _generated_operations = operation.execute(source)?;
Ok(())
}
// Helpers
fn gen_file_owners(filename: &str) -> BoxedErrorResult<Vec<String>> {
let file_idx = filename.easyhash();
heartbeat::gen_neighbor_list_from(file_idx as i32, 1, constants::NUM_OWNERS, true)
}
// TODO: This function makes the entire system assume there are always at least two nodes in the system
// and the file must have an owner or else the operation will not work correctly. This is fine for now
// but it is worth improving sooner rather than later (make distinct Error types to differentiate, etc).
fn gen_new_file_owner(filename: &str) -> BoxedErrorResult<String> {
match globals::ALL_FILE_OWNERS.read().get(filename) {
Some(owners) => {
let potential_owners = gen_file_owners(filename)?;
for potential_owner in &potential_owners {
if !owners.contains(potential_owner) |
}
Err(format!("No new owners available for file {}", filename).into())
},
None => Err(format!("No owner found for file {}", filename).into())
}
}
fn distributed_file_path(filename: &String) -> String {
format!("{}/{}", constants::DATA_DIR, filename)
}
// Returns messages to be gossiped
pub fn handle_failed_node(failed_id: &String) -> BoxedErrorResult<Vec<SendableOperation>> {
match heartbeat::is_master() {
true => {
let mut generated_operations: Vec<SendableOperation> = Vec::new();
let myself_source = Source::myself();
// Find files they owned
let mut lost_files: HashSet<String> = HashSet::new();
for (distributed_filename, owners) in globals::ALL_FILE_OWNERS.read().iter() {
if owners.contains(failed_id) {
lost_files.insert(distributed_filename.clone());
}
}
log("Found lost files".to_string());
// Send that they no longer own those files
// Separate operation so that this acts like a confirmation to fully forget about the node from
// the master. Can be used with a delay later if you want more error resistance.
let lost_file_operation = LostFilesOperation {
failed_owner: failed_id.clone(),
lost_files: lost_files.clone()
};
generated_operations.append(&mut lost_file_operation.execute(myself_source.clone())?);
log("Executed lost files operation locally".to_string());
// Gen new owners of the file and propagate
let mut new_owners: HashMap<String, HashSet<String>> = HashMap::new();
for lost_file in &lost_files {
let new_owner = gen_new_file_owner(&lost_file)?;
// TODO: Maybe optimize this into one fat packet - probably a new operation?
let new_owner_operation = NewFileOwnersOperation {
distributed_filename: lost_file.clone(),
new_owners: vec![new_owner].iter().map(|x| x.to_string()).collect()
};
generated_operations.append(&mut new_owner_operation.execute(myself_source.clone())?);
}
log("Executed all new_owner operations locally".to_string());
Ok(generated_operations)
},
false => {
Ok(vec![])
}
}
}
// Operations
#[derive(Serialize, Deserialize, Debug, Clone)]
pub struct GetOperation {
pub distributed_filename: String,
pub local_path: String
}
#[derive(Serialize, Deserialize, Debug, Clone)]
pub struct NewFileOwnersOperation {
pub distributed_filename: String,
pub new_owners: HashSet<String>
}
#[derive(Serialize, Deserialize, Clone)]
pub struct SendFileOperation {
pub filename: String,
pub data: Vec<u8>,
pub is_distributed: bool
}
#[derive(Serialize, Deserialize, Debug, Clone)]
pub struct LostFilesOperation {
pub failed_owner: String,
pub lost_files: HashSet<String>
}
// Trait Impls
impl OperationWriteExecute for GetOperation {
fn to_bytes(&self) -> BoxedErrorResult<Vec<u8>> {
Ok(create_buf(&self, str_to_vec("GET ")))
}
fn execute(&self, source: Source) -> BoxedErrorResult<Vec<SendableOperation>> {
let local_path = distributed_file_path(&self.distributed_filename);
let data_buf = async_std::task::block_on(read_file_to_buf(&local_path))?;
let operation = SendableOperation::for_single_tcp_stream(
TryInto::<async_std::net::TcpStream>::try_into(source)?,
Box::new(SendFileOperation {
filename: self.local_path.clone(),
data: data_buf,
is_distributed: false
}));
async_std::task::block_on(operation.write_all_tcp_async());
Ok(vec![])
}
fn to_string(&self) -> String { format!("{:?}", self) }
}
impl OperationWriteExecute for NewFileOwnersOperation {
fn to_bytes(&self | {
return Ok(potential_owner.clone());
} | conditional_block |
filesystem.rs | = args[0];
let distributed_filename = args[1];
// Figure out who I am giving this file to
let dest_ids = gen_file_owners(&distributed_filename)?;
// Gossip who has the file now
sender.send(
SendableOperation::for_successors(Box::new(NewFileOwnersOperation {
distributed_filename: distributed_filename.to_string(),
new_owners: dest_ids
.iter()
.map(|x| x.to_string())
.collect::<HashSet<_>>()
}))
)?;
// Send them the file
async_std::task::block_on(send_file_to_all(local_path.to_string(),
distributed_filename.to_string(),
&dest_ids))?;
Ok(())
}
pub fn ls(args: Vec<&str>) -> BoxedErrorResult<()> {
check_joined()?;
let invalid_args: BoxedErrorResult<()> = Err("Usage: ls [distributed_filename]".into());
match args.len() {
0 => {
// All
print_file_owners(None, true)?;
Ok(())
},
1 => {
// Just File
let distributed_filename = args[0];
print_file_owners(Some(distributed_filename), false)?;
Ok(())
},
_ => invalid_args
}
}
// TODO: You wrote this very late - maybe fix
fn print_file_owners(maybe_distributed_filename: Option<&str>, full: bool) -> BoxedErrorResult<()> {
let all_file_owners = globals::ALL_FILE_OWNERS.read();
match (maybe_distributed_filename, full) {
(Some(_), true) => {
Err("Cannot set distributed_filename and full to true when printing owners".into())
},
(Some(distributed_filename), false) => {
// Print the files owners
match all_file_owners.get(distributed_filename) {
Some(owners) => {
println!("{:?}", owners);
},
None => {
// A little unoptimal - change if above format changes
println!("{{}}");
}
}
Ok(())
},
(None, true) => {
// Print the whole map
println!("{:?}", *all_file_owners);
Ok(())
},
(None, false) => {
Err("Cannot print owners of nonexistant distributed_filename with full set to false".into())
}
}
}
async fn get_distributed_file(distributed_filename: String, local_path: String) -> BoxedErrorResult<()> {
// TODO: Find owners
let operation = SendableOperation::for_owners(&distributed_filename, Box::new(GetOperation {
distributed_filename: distributed_filename.clone(),
local_path: local_path
}));
let mut streams = operation
.write_all_tcp_async()
.await?;
// TODO: Redo whatever tf going on here
match streams.len() {
0 => Err(format!("No owners found for file {}", distributed_filename).into()),
_ => {
let (result, source) = streams[0]
.try_read_operation()
.await?;
result.execute(source)?;
Ok(())
}
}
}
async fn read_file_to_buf(local_path: &String) -> BoxedErrorResult<Vec<u8>> {
let mut data_buf: Vec<u8> = Vec::new();
let mut file = async_std::fs::File::open(&local_path).await?;
file.read_to_end(&mut data_buf).await?;
Ok(data_buf)
}
async fn send_file_to_all(local_path: String, distributed_filename: String, dest_ids: &Vec<String>) ->
BoxedErrorResult<()> {
let data_buf = read_file_to_buf(&local_path).await?;
let operation = SendableOperation::for_id_list(dest_ids.clone(), Box::new(SendFileOperation {
filename: distributed_filename,
data: data_buf,
is_distributed: true
}));
operation.write_all_tcp_async().await?;
Ok(())
}
pub async fn file_server<'a>(_sender: &'a OperationSender) -> BoxedErrorResult<()> {
let server = globals::SERVER_SOCKET.read();
let mut incoming = server.incoming();
while let Some(stream) = incoming.next().await {
let connection = stream?;
log(format!("Handling connection from {:?}", connection.peer_addr()));
spawn(handle_connection(connection));
}
Ok(())
}
async fn handle_connection(mut connection: async_std::net::TcpStream) -> BoxedErrorResult<()> {
let (operation, source) = connection.try_read_operation().await?;
// TODO: Think about what standard we want with these
let _generated_operations = operation.execute(source)?;
Ok(())
}
// Helpers
fn gen_file_owners(filename: &str) -> BoxedErrorResult<Vec<String>> {
let file_idx = filename.easyhash();
heartbeat::gen_neighbor_list_from(file_idx as i32, 1, constants::NUM_OWNERS, true)
} |
// TODO: This function makes the entire system assume there are always at least two nodes in the system
// and the file must have an owner or else the operation will not work correctly. This is fine for now
// but it is worth improving sooner rather than later (make distinct Error types to differentiate, etc).
fn gen_new_file_owner(filename: &str) -> BoxedErrorResult<String> {
match globals::ALL_FILE_OWNERS.read().get(filename) {
Some(owners) => {
let potential_owners = gen_file_owners(filename)?;
for potential_owner in &potential_owners {
if !owners.contains(potential_owner) {
return Ok(potential_owner.clone());
}
}
Err(format!("No new owners available for file {}", filename).into())
},
None => Err(format!("No owner found for file {}", filename).into())
}
}
fn distributed_file_path(filename: &String) -> String {
format!("{}/{}", constants::DATA_DIR, filename)
}
// Returns messages to be gossiped
pub fn handle_failed_node(failed_id: &String) -> BoxedErrorResult<Vec<SendableOperation>> {
match heartbeat::is_master() {
true => {
let mut generated_operations: Vec<SendableOperation> = Vec::new();
let myself_source = Source::myself();
// Find files they owned
let mut lost_files: HashSet<String> = HashSet::new();
for (distributed_filename, owners) in globals::ALL_FILE_OWNERS.read().iter() {
if owners.contains(failed_id) {
lost_files.insert(distributed_filename.clone());
}
}
log("Found lost files".to_string());
// Send that they no longer own those files
// Separate operation so that this acts like a confirmation to fully forget about the node from
// the master. Can be used with a delay later if you want more error resistance.
let lost_file_operation = LostFilesOperation {
failed_owner: failed_id.clone(),
lost_files: lost_files.clone()
};
generated_operations.append(&mut lost_file_operation.execute(myself_source.clone())?);
log("Executed lost files operation locally".to_string());
// Gen new owners of the file and propagate
let mut new_owners: HashMap<String, HashSet<String>> = HashMap::new();
for lost_file in &lost_files {
let new_owner = gen_new_file_owner(&lost_file)?;
// TODO: Maybe optimize this into one fat packet - probably a new operation?
let new_owner_operation = NewFileOwnersOperation {
distributed_filename: lost_file.clone(),
new_owners: vec![new_owner].iter().map(|x| x.to_string()).collect()
};
generated_operations.append(&mut new_owner_operation.execute(myself_source.clone())?);
}
log("Executed all new_owner operations locally".to_string());
Ok(generated_operations)
},
false => {
Ok(vec![])
}
}
}
// Operations
#[derive(Serialize, Deserialize, Debug, Clone)]
pub struct GetOperation {
pub distributed_filename: String,
pub local_path: String
}
#[derive(Serialize, Deserialize, Debug, Clone)]
pub struct NewFileOwnersOperation {
pub distributed_filename: String,
pub new_owners: HashSet<String>
}
#[derive(Serialize, Deserialize, Clone)]
pub struct SendFileOperation {
pub filename: String,
pub data: Vec<u8>,
pub is_distributed: bool
}
#[derive(Serialize, Deserialize, Debug, Clone)]
pub struct LostFilesOperation {
pub failed_owner: String,
pub lost_files: HashSet<String>
}
// Trait Impls
impl OperationWriteExecute for GetOperation {
fn to_bytes(&self) -> BoxedErrorResult<Vec<u8>> {
Ok(create_buf(&self, str_to_vec("GET ")))
}
fn execute(&self, source: Source) -> BoxedErrorResult<Vec<SendableOperation>> {
let local_path = distributed_file_path(&self.distributed_filename);
let data_buf = async_std::task::block_on(read_file_to_buf(&local_path))?;
let operation = SendableOperation::for_single_tcp_stream(
TryInto::<async_std::net::TcpStream>::try_into(source)?,
Box::new(SendFileOperation {
filename: self.local_path.clone(),
data: data_buf,
is_distributed: false
}));
async_std::task::block_on(operation.write_all_tcp_async());
Ok(vec![])
}
fn to_string(&self) -> String { format!("{:?}", self) }
}
impl OperationWriteExecute for NewFileOwnersOperation {
fn to_bytes(&self) | random_line_split |
|
filesystem.rs | = args[0];
let distributed_filename = args[1];
// Figure out who I am giving this file to
let dest_ids = gen_file_owners(&distributed_filename)?;
// Gossip who has the file now
sender.send(
SendableOperation::for_successors(Box::new(NewFileOwnersOperation {
distributed_filename: distributed_filename.to_string(),
new_owners: dest_ids
.iter()
.map(|x| x.to_string())
.collect::<HashSet<_>>()
}))
)?;
// Send them the file
async_std::task::block_on(send_file_to_all(local_path.to_string(),
distributed_filename.to_string(),
&dest_ids))?;
Ok(())
}
pub fn ls(args: Vec<&str>) -> BoxedErrorResult<()> {
check_joined()?;
let invalid_args: BoxedErrorResult<()> = Err("Usage: ls [distributed_filename]".into());
match args.len() {
0 => {
// All
print_file_owners(None, true)?;
Ok(())
},
1 => {
// Just File
let distributed_filename = args[0];
print_file_owners(Some(distributed_filename), false)?;
Ok(())
},
_ => invalid_args
}
}
// TODO: You wrote this very late - maybe fix
fn print_file_owners(maybe_distributed_filename: Option<&str>, full: bool) -> BoxedErrorResult<()> {
let all_file_owners = globals::ALL_FILE_OWNERS.read();
match (maybe_distributed_filename, full) {
(Some(_), true) => {
Err("Cannot set distributed_filename and full to true when printing owners".into())
},
(Some(distributed_filename), false) => {
// Print the files owners
match all_file_owners.get(distributed_filename) {
Some(owners) => {
println!("{:?}", owners);
},
None => {
// A little unoptimal - change if above format changes
println!("{{}}");
}
}
Ok(())
},
(None, true) => {
// Print the whole map
println!("{:?}", *all_file_owners);
Ok(())
},
(None, false) => {
Err("Cannot print owners of nonexistant distributed_filename with full set to false".into())
}
}
}
async fn get_distributed_file(distributed_filename: String, local_path: String) -> BoxedErrorResult<()> {
// TODO: Find owners
let operation = SendableOperation::for_owners(&distributed_filename, Box::new(GetOperation {
distributed_filename: distributed_filename.clone(),
local_path: local_path
}));
let mut streams = operation
.write_all_tcp_async()
.await?;
// TODO: Redo whatever tf going on here
match streams.len() {
0 => Err(format!("No owners found for file {}", distributed_filename).into()),
_ => {
let (result, source) = streams[0]
.try_read_operation()
.await?;
result.execute(source)?;
Ok(())
}
}
}
async fn read_file_to_buf(local_path: &String) -> BoxedErrorResult<Vec<u8>> {
let mut data_buf: Vec<u8> = Vec::new();
let mut file = async_std::fs::File::open(&local_path).await?;
file.read_to_end(&mut data_buf).await?;
Ok(data_buf)
}
async fn send_file_to_all(local_path: String, distributed_filename: String, dest_ids: &Vec<String>) ->
BoxedErrorResult<()> {
let data_buf = read_file_to_buf(&local_path).await?;
let operation = SendableOperation::for_id_list(dest_ids.clone(), Box::new(SendFileOperation {
filename: distributed_filename,
data: data_buf,
is_distributed: true
}));
operation.write_all_tcp_async().await?;
Ok(())
}
pub async fn file_server<'a>(_sender: &'a OperationSender) -> BoxedErrorResult<()> |
async fn handle_connection(mut connection: async_std::net::TcpStream) -> BoxedErrorResult<()> {
let (operation, source) = connection.try_read_operation().await?;
// TODO: Think about what standard we want with these
let _generated_operations = operation.execute(source)?;
Ok(())
}
// Helpers
fn gen_file_owners(filename: &str) -> BoxedErrorResult<Vec<String>> {
let file_idx = filename.easyhash();
heartbeat::gen_neighbor_list_from(file_idx as i32, 1, constants::NUM_OWNERS, true)
}
// TODO: This function makes the entire system assume there are always at least two nodes in the system
// and the file must have an owner or else the operation will not work correctly. This is fine for now
// but it is worth improving sooner rather than later (make distinct Error types to differentiate, etc).
fn gen_new_file_owner(filename: &str) -> BoxedErrorResult<String> {
match globals::ALL_FILE_OWNERS.read().get(filename) {
Some(owners) => {
let potential_owners = gen_file_owners(filename)?;
for potential_owner in &potential_owners {
if !owners.contains(potential_owner) {
return Ok(potential_owner.clone());
}
}
Err(format!("No new owners available for file {}", filename).into())
},
None => Err(format!("No owner found for file {}", filename).into())
}
}
fn distributed_file_path(filename: &String) -> String {
format!("{}/{}", constants::DATA_DIR, filename)
}
// Returns messages to be gossiped
pub fn handle_failed_node(failed_id: &String) -> BoxedErrorResult<Vec<SendableOperation>> {
match heartbeat::is_master() {
true => {
let mut generated_operations: Vec<SendableOperation> = Vec::new();
let myself_source = Source::myself();
// Find files they owned
let mut lost_files: HashSet<String> = HashSet::new();
for (distributed_filename, owners) in globals::ALL_FILE_OWNERS.read().iter() {
if owners.contains(failed_id) {
lost_files.insert(distributed_filename.clone());
}
}
log("Found lost files".to_string());
// Send that they no longer own those files
// Separate operation so that this acts like a confirmation to fully forget about the node from
// the master. Can be used with a delay later if you want more error resistance.
let lost_file_operation = LostFilesOperation {
failed_owner: failed_id.clone(),
lost_files: lost_files.clone()
};
generated_operations.append(&mut lost_file_operation.execute(myself_source.clone())?);
log("Executed lost files operation locally".to_string());
// Gen new owners of the file and propagate
let mut new_owners: HashMap<String, HashSet<String>> = HashMap::new();
for lost_file in &lost_files {
let new_owner = gen_new_file_owner(&lost_file)?;
// TODO: Maybe optimize this into one fat packet - probably a new operation?
let new_owner_operation = NewFileOwnersOperation {
distributed_filename: lost_file.clone(),
new_owners: vec![new_owner].iter().map(|x| x.to_string()).collect()
};
generated_operations.append(&mut new_owner_operation.execute(myself_source.clone())?);
}
log("Executed all new_owner operations locally".to_string());
Ok(generated_operations)
},
false => {
Ok(vec![])
}
}
}
// Operations
#[derive(Serialize, Deserialize, Debug, Clone)]
pub struct GetOperation {
pub distributed_filename: String,
pub local_path: String
}
#[derive(Serialize, Deserialize, Debug, Clone)]
pub struct NewFileOwnersOperation {
pub distributed_filename: String,
pub new_owners: HashSet<String>
}
#[derive(Serialize, Deserialize, Clone)]
pub struct SendFileOperation {
pub filename: String,
pub data: Vec<u8>,
pub is_distributed: bool
}
#[derive(Serialize, Deserialize, Debug, Clone)]
pub struct LostFilesOperation {
pub failed_owner: String,
pub lost_files: HashSet<String>
}
// Trait Impls
impl OperationWriteExecute for GetOperation {
fn to_bytes(&self) -> BoxedErrorResult<Vec<u8>> {
Ok(create_buf(&self, str_to_vec("GET ")))
}
fn execute(&self, source: Source) -> BoxedErrorResult<Vec<SendableOperation>> {
let local_path = distributed_file_path(&self.distributed_filename);
let data_buf = async_std::task::block_on(read_file_to_buf(&local_path))?;
let operation = SendableOperation::for_single_tcp_stream(
TryInto::<async_std::net::TcpStream>::try_into(source)?,
Box::new(SendFileOperation {
filename: self.local_path.clone(),
data: data_buf,
is_distributed: false
}));
async_std::task::block_on(operation.write_all_tcp_async());
Ok(vec![])
}
fn to_string(&self) -> String { format!("{:?}", self) }
}
impl OperationWriteExecute for NewFileOwnersOperation {
fn to_bytes(&self | {
let server = globals::SERVER_SOCKET.read();
let mut incoming = server.incoming();
while let Some(stream) = incoming.next().await {
let connection = stream?;
log(format!("Handling connection from {:?}", connection.peer_addr()));
spawn(handle_connection(connection));
}
Ok(())
} | identifier_body |
main.rs | let guid = &item.guid;
items.entry(&guid).or_insert_with(|| Item {
title: item.title.to_owned(),
link: item.link.to_owned(),
content: item.content.to_owned(),
pub_date: item.pub_date.to_owned(),
guid: guid.to_owned(),
unread: true,
});
}
self.items = items.drain().map(|(_, v)| v).collect();
self.items.sort_by_key(|item| item.pub_date.clone());
}
}
#[derive(Serialize)]
struct FeedInfo {
pub url: String,
pub title: String,
pub last_updated: DateTime<Utc>,
}
struct DownloadFeed(String);
#[derive(Deserialize)]
struct AddFeed { url: String }
#[derive(Deserialize)]
struct RemoveFeed { url: String }
#[derive(Deserialize, Debug)]
struct GetFeed { url: String }
struct ListFeeds;
#[derive(Deserialize, Debug)]
struct MarkRead { url: String, guid: String }
#[derive(Message)]
struct UpdateFeed { url: String, feed: Feed }
impl Message for DownloadFeed {
type Result = Result<Feed, String>;
}
impl Message for AddFeed {
type Result = Result<(), String>;
}
impl Message for RemoveFeed {
type Result = Result<(), String>;
}
impl Message for GetFeed {
type Result = Result<Feed, String>;
}
impl Message for ListFeeds {
type Result = Result<Vec<FeedInfo>, String>;
}
impl Message for MarkRead {
type Result = Result<bool, String>;
}
struct FeedStorage {
feeds: HashMap<String, Feed>,
downloader: actix::Addr<Downloader>,
}
impl Actor for FeedStorage {
type Context = actix::SyncContext<Self>;
}
impl Handler<DownloadFeed> for FeedStorage {
type Result = <DownloadFeed as Message>::Result;
fn handle(&mut self, msg: DownloadFeed, _: &mut Self::Context) -> Self::Result {
self.downloader.send(msg).wait().or_err("Download failed")?
}
}
impl Handler<AddFeed> for FeedStorage {
type Result = <AddFeed as Message>::Result;
fn handle(&mut self, msg: AddFeed, _: &mut Self::Context) -> Self::Result {
match self.feeds.entry(msg.url.clone()) {
std::collections::hash_map::Entry::Occupied(_) => Err("Feed already exists".into()),
std::collections::hash_map::Entry::Vacant(e) => {
debug!("will download {}", &msg.url);
self.downloader.send(DownloadFeed(msg.url))
.wait()
.or_err("Failed to download")?
.map(|feed| {
debug!("downloaded");
e.insert(feed);
})
}
}
}
}
impl Handler<RemoveFeed> for FeedStorage {
type Result = <RemoveFeed as Message>::Result;
fn handle(&mut self, msg: RemoveFeed, _: &mut Self::Context) -> Self::Result {
self.feeds.remove(&msg.url);
Ok(())
}
}
impl Handler<GetFeed> for FeedStorage {
type Result = <GetFeed as Message>::Result;
fn handle(&mut self, msg: GetFeed, _: &mut Self::Context) -> Self::Result {
match self.feeds.get(&msg.url) {
None => Err("Feed not found".into()),
Some(feed) => Ok(feed.clone()),
}
}
}
impl Handler<ListFeeds> for FeedStorage {
type Result = <ListFeeds as Message>::Result;
fn handle(&mut self, _: ListFeeds, _: &mut Self::Context) -> Self::Result {
Ok(self.feeds.iter().map(|(k, v)| FeedInfo{url: k.clone(), title: v.title.clone(), last_updated: v.last_updated.clone()}).collect())
}
}
impl Handler<MarkRead> for FeedStorage {
type Result = <MarkRead as Message>::Result;
fn handle(&mut self, msg: MarkRead, _: &mut Self::Context) -> Self::Result {
let mut updated = false;
if let Some(feed) = self.feeds.get_mut(&msg.url) {
for item in feed.items.iter_mut().filter(|k| &k.guid == &msg.guid).take(1) {
item.unread = false;
updated = true;
}
}
Ok(updated)
}
}
impl Handler<UpdateFeed> for FeedStorage {
type Result = <UpdateFeed as Message>::Result;
fn handle(&mut self, msg: UpdateFeed, _: &mut Self::Context) -> Self::Result {
if let Some(feed) = self.feeds.get_mut(&msg.url) | ;
}
}
struct Downloader;
impl Actor for Downloader {
type Context = actix::Context<Self>;
}
impl Handler<DownloadFeed> for Downloader {
type Result = <DownloadFeed as Message>::Result;
fn handle(&mut self, msg: DownloadFeed, _: &mut Self::Context) -> Self::Result {
let channel = rss::Channel::from_url(&msg.0).or_err("Channel not downloaded")?;
let mut items = vec![];
for item in channel.items().iter() {
let guid = item.guid().or_err("broken channel")?.value();
items.push(Item {
title: item.title().map(|s| s.to_string()),
link: item.link().map(|s| s.to_string()),
content: item.content().or(item.description()).map(|s| s.to_string()),
pub_date: item.pub_date().and_then(|date| DateTime::parse_from_rfc2822(date).ok().map(|d| d.with_timezone(&Utc))),
guid: guid.to_string(),
unread: true,
});
}
Ok(Feed{
title: channel.title().to_owned(),
last_updated: match channel.last_build_date() {
None => items
.iter()
.map(|item| &item.pub_date)
.max()
.map(|date| date.to_owned())
.unwrap_or(Some(Utc.timestamp(0, 0)))
.unwrap_or(Utc.timestamp(0, 0)),
Some(s) => DateTime::parse_from_rfc2822(s).map(|d| d.with_timezone(&Utc)).unwrap_or(Utc.timestamp(0, 0))
},
items: items
})
}
}
struct Updater {
storage: actix::Addr<FeedStorage>,
downloader: actix::Addr<Downloader>,
handle: Option<actix::SpawnHandle>,
arbiter: actix::Arbiter,
}
impl Actor for Updater {
type Context = actix::Context<Self>;
fn started(&mut self, ctx: &mut <Self as Actor>::Context) {
let storage = self.storage.clone();
let downloader = self.downloader.clone();
let arbiter = self.arbiter.clone();
self.handle = Some(ctx.run_interval(std::time::Duration::new(60, 0), move |_, _| {
let storage = storage.clone();
let downloader = downloader.clone();
let arbiter = arbiter.clone();
arbiter.exec_fn(move || {
if let Ok(Ok(infos)) = storage.send(ListFeeds).wait() {
debug!("got {} feeds, updating", infos.len());
for info in infos {
if let Ok(Ok(new_feed)) = downloader.send(DownloadFeed(info.url.clone())).wait() {
if let Ok(()) = storage.send(UpdateFeed{url: info.url.clone(), feed: new_feed}).wait() {
debug!("successfully updated {}", info.url);
}
}
}
}
});
}));
}
}
fn process_response<T: Serialize, E: Serialize, E2, F: FnOnce(E) -> actix_web::Error>(response: Result<Result<T, E>, E2>, f: F) -> Result<HttpResponse, actix_web::Error> {
match response {
Ok(Ok(data)) => Ok(HttpResponse::Ok().json(data)),
Ok(Err(e)) => Err(f(e)),
_ => Err(actix_web::error::ErrorInternalServerError("Application overload"))
}
}
#[derive(Clone)]
struct State { storage: actix::Addr<FeedStorage> }
fn add_feed(url_info: web::Form<AddFeed>, data: web::Data<State>) -> impl Future<Item = HttpResponse, Error = actix_web::Error> {
web::block(move || data.storage.send(url_info.into_inner()).wait())
.then(|res| process_response(res, actix_web::error::ErrorInternalServerError))
}
fn remove_feed(url_info: web::Form<RemoveFeed>, data: web::Data<State>) -> impl Future<Item = HttpResponse, Error = actix_web::Error> {
web::block(move || data.storage.send(url_info.into_inner()).wait())
.then(|res| process_response(res, actix_web::error::ErrorInternalServerError))
}
fn get_feed(url_info: web::Query<GetFeed>, data: web::Data<State>) -> impl Future<Item = HttpResponse, Error = actix_web::Error> {
web::block(move || data.storage.send(url_info.into_inner()).wait())
.then(|res| process_response(res, actix_web::error::ErrorNotFound))
}
fn list_feeds(data: web::Data<State>) -> impl Future<Item = HttpResponse, Error = actix_web::Error> {
web::block(move || data.storage.send(ListFeeds).wait())
.then(|res| process_response(res, actix_web:: | {
feed.merge(msg.feed)
} | conditional_block |
main.rs | Entry::Occupied(_) => Err("Feed already exists".into()),
std::collections::hash_map::Entry::Vacant(e) => {
debug!("will download {}", &msg.url);
self.downloader.send(DownloadFeed(msg.url))
.wait()
.or_err("Failed to download")?
.map(|feed| {
debug!("downloaded");
e.insert(feed);
})
}
}
}
}
impl Handler<RemoveFeed> for FeedStorage {
type Result = <RemoveFeed as Message>::Result;
fn handle(&mut self, msg: RemoveFeed, _: &mut Self::Context) -> Self::Result {
self.feeds.remove(&msg.url);
Ok(())
}
}
impl Handler<GetFeed> for FeedStorage {
type Result = <GetFeed as Message>::Result;
fn handle(&mut self, msg: GetFeed, _: &mut Self::Context) -> Self::Result {
match self.feeds.get(&msg.url) {
None => Err("Feed not found".into()),
Some(feed) => Ok(feed.clone()),
}
}
}
impl Handler<ListFeeds> for FeedStorage {
type Result = <ListFeeds as Message>::Result;
fn handle(&mut self, _: ListFeeds, _: &mut Self::Context) -> Self::Result {
Ok(self.feeds.iter().map(|(k, v)| FeedInfo{url: k.clone(), title: v.title.clone(), last_updated: v.last_updated.clone()}).collect())
}
}
impl Handler<MarkRead> for FeedStorage {
type Result = <MarkRead as Message>::Result;
fn handle(&mut self, msg: MarkRead, _: &mut Self::Context) -> Self::Result {
let mut updated = false;
if let Some(feed) = self.feeds.get_mut(&msg.url) {
for item in feed.items.iter_mut().filter(|k| &k.guid == &msg.guid).take(1) {
item.unread = false;
updated = true;
}
}
Ok(updated)
}
}
impl Handler<UpdateFeed> for FeedStorage {
type Result = <UpdateFeed as Message>::Result;
fn handle(&mut self, msg: UpdateFeed, _: &mut Self::Context) -> Self::Result {
if let Some(feed) = self.feeds.get_mut(&msg.url) {
feed.merge(msg.feed)
};
}
}
struct Downloader;
impl Actor for Downloader {
type Context = actix::Context<Self>;
}
impl Handler<DownloadFeed> for Downloader {
type Result = <DownloadFeed as Message>::Result;
fn handle(&mut self, msg: DownloadFeed, _: &mut Self::Context) -> Self::Result {
let channel = rss::Channel::from_url(&msg.0).or_err("Channel not downloaded")?;
let mut items = vec![];
for item in channel.items().iter() {
let guid = item.guid().or_err("broken channel")?.value();
items.push(Item {
title: item.title().map(|s| s.to_string()),
link: item.link().map(|s| s.to_string()),
content: item.content().or(item.description()).map(|s| s.to_string()),
pub_date: item.pub_date().and_then(|date| DateTime::parse_from_rfc2822(date).ok().map(|d| d.with_timezone(&Utc))),
guid: guid.to_string(),
unread: true,
});
}
Ok(Feed{
title: channel.title().to_owned(),
last_updated: match channel.last_build_date() {
None => items
.iter()
.map(|item| &item.pub_date)
.max()
.map(|date| date.to_owned())
.unwrap_or(Some(Utc.timestamp(0, 0)))
.unwrap_or(Utc.timestamp(0, 0)),
Some(s) => DateTime::parse_from_rfc2822(s).map(|d| d.with_timezone(&Utc)).unwrap_or(Utc.timestamp(0, 0))
},
items: items
})
}
}
struct Updater {
storage: actix::Addr<FeedStorage>,
downloader: actix::Addr<Downloader>,
handle: Option<actix::SpawnHandle>,
arbiter: actix::Arbiter,
}
impl Actor for Updater {
type Context = actix::Context<Self>;
fn started(&mut self, ctx: &mut <Self as Actor>::Context) {
let storage = self.storage.clone();
let downloader = self.downloader.clone();
let arbiter = self.arbiter.clone();
self.handle = Some(ctx.run_interval(std::time::Duration::new(60, 0), move |_, _| {
let storage = storage.clone();
let downloader = downloader.clone();
let arbiter = arbiter.clone();
arbiter.exec_fn(move || {
if let Ok(Ok(infos)) = storage.send(ListFeeds).wait() {
debug!("got {} feeds, updating", infos.len());
for info in infos {
if let Ok(Ok(new_feed)) = downloader.send(DownloadFeed(info.url.clone())).wait() {
if let Ok(()) = storage.send(UpdateFeed{url: info.url.clone(), feed: new_feed}).wait() {
debug!("successfully updated {}", info.url);
}
}
}
}
});
}));
}
}
fn process_response<T: Serialize, E: Serialize, E2, F: FnOnce(E) -> actix_web::Error>(response: Result<Result<T, E>, E2>, f: F) -> Result<HttpResponse, actix_web::Error> {
match response {
Ok(Ok(data)) => Ok(HttpResponse::Ok().json(data)),
Ok(Err(e)) => Err(f(e)),
_ => Err(actix_web::error::ErrorInternalServerError("Application overload"))
}
}
#[derive(Clone)]
struct State { storage: actix::Addr<FeedStorage> }
fn add_feed(url_info: web::Form<AddFeed>, data: web::Data<State>) -> impl Future<Item = HttpResponse, Error = actix_web::Error> {
web::block(move || data.storage.send(url_info.into_inner()).wait())
.then(|res| process_response(res, actix_web::error::ErrorInternalServerError))
}
fn remove_feed(url_info: web::Form<RemoveFeed>, data: web::Data<State>) -> impl Future<Item = HttpResponse, Error = actix_web::Error> {
web::block(move || data.storage.send(url_info.into_inner()).wait())
.then(|res| process_response(res, actix_web::error::ErrorInternalServerError))
}
fn get_feed(url_info: web::Query<GetFeed>, data: web::Data<State>) -> impl Future<Item = HttpResponse, Error = actix_web::Error> {
web::block(move || data.storage.send(url_info.into_inner()).wait())
.then(|res| process_response(res, actix_web::error::ErrorNotFound))
}
fn list_feeds(data: web::Data<State>) -> impl Future<Item = HttpResponse, Error = actix_web::Error> {
web::block(move || data.storage.send(ListFeeds).wait())
.then(|res| process_response(res, actix_web::error::ErrorInternalServerError))
}
fn mark_read(url_info: web::Form<MarkRead>, data: web::Data<State>) -> impl Future<Item = HttpResponse, Error = actix_web::Error> {
web::block(move || data.storage.send(url_info.into_inner()).wait())
.then(|res| process_response(res, actix_web::error::ErrorInternalServerError))
}
fn actix_main() -> Result<(), std::io::Error> {
let downloader_addr = Downloader.start();
let feed_storage_addr = {
let addr = downloader_addr.clone();
actix::SyncArbiter::start(1, move || FeedStorage{
feeds: HashMap::new(),
downloader: addr.clone(),
})
};
let state = State{storage: feed_storage_addr.clone()};
let updater = Updater{storage: feed_storage_addr, downloader: downloader_addr.clone(), handle: None, arbiter: actix::Arbiter::new()};
updater.start();
let mut server = actix_web::HttpServer::new(move || {
App::new()
.data(state.clone())
.wrap(
Cors::new()
.allowed_methods(vec!["GET", "POST", "OPTIONS"])
.allowed_headers(vec![
http::header::ACCEPT,
http::header::CONTENT_TYPE,
http::header::HeaderName::from_static("x-requested-with")
])
.max_age(3600)
)
.wrap(actix_web::middleware::Logger::default())
.route("/add", web::post().to_async(add_feed))
.route("/remove", web::post().to_async(remove_feed))
.route("/read", web::post().to_async(mark_read))
.route("/list", web::get().to_async(list_feeds))
.route("/get", web::get().to_async(get_feed))
});
let mut listenfd = listenfd::ListenFd::from_env();
server = if let Some(l) = listenfd.take_tcp_listener(0)? { | server.listen(l)?
} else {
server.bind("[::1]:8000")?
};
println!("Started HTTP server on {:?}", server.addrs_with_scheme().iter().map(|(a, s)| format!("{}://{}/", s, a)).collect::<Vec<_>>()); | random_line_split |
|
main.rs | {
pub title: Option<String>,
pub link: Option<String>,
pub content: Option<String>,
pub pub_date: Option<DateTime<Utc>>,
pub guid: String,
pub unread: bool,
}
#[derive(Clone, Debug, Serialize)]
struct Feed {
pub title: String,
pub last_updated: DateTime<Utc>,
pub items: Vec<Item>,
}
impl Feed {
pub fn merge(&mut self, other: Feed) {
self.title = other.title;
self.last_updated = other.last_updated;
let mut items: HashMap<&str, Item> = self.items.iter().map(|item| (item.guid.as_str(), item.clone())).collect();
for item in other.items.iter() {
let guid = &item.guid;
items.entry(&guid).or_insert_with(|| Item {
title: item.title.to_owned(),
link: item.link.to_owned(),
content: item.content.to_owned(),
pub_date: item.pub_date.to_owned(),
guid: guid.to_owned(),
unread: true,
});
}
self.items = items.drain().map(|(_, v)| v).collect();
self.items.sort_by_key(|item| item.pub_date.clone());
}
}
#[derive(Serialize)]
struct FeedInfo {
pub url: String,
pub title: String,
pub last_updated: DateTime<Utc>,
}
struct DownloadFeed(String);
#[derive(Deserialize)]
struct AddFeed { url: String }
#[derive(Deserialize)]
struct RemoveFeed { url: String }
#[derive(Deserialize, Debug)]
struct GetFeed { url: String }
struct ListFeeds;
#[derive(Deserialize, Debug)]
struct MarkRead { url: String, guid: String }
#[derive(Message)]
struct UpdateFeed { url: String, feed: Feed }
impl Message for DownloadFeed {
type Result = Result<Feed, String>;
}
impl Message for AddFeed {
type Result = Result<(), String>;
}
impl Message for RemoveFeed {
type Result = Result<(), String>;
}
impl Message for GetFeed {
type Result = Result<Feed, String>;
}
impl Message for ListFeeds {
type Result = Result<Vec<FeedInfo>, String>;
}
impl Message for MarkRead {
type Result = Result<bool, String>;
}
struct FeedStorage {
feeds: HashMap<String, Feed>,
downloader: actix::Addr<Downloader>,
}
impl Actor for FeedStorage {
type Context = actix::SyncContext<Self>;
}
impl Handler<DownloadFeed> for FeedStorage {
type Result = <DownloadFeed as Message>::Result;
fn handle(&mut self, msg: DownloadFeed, _: &mut Self::Context) -> Self::Result {
self.downloader.send(msg).wait().or_err("Download failed")?
}
}
impl Handler<AddFeed> for FeedStorage {
type Result = <AddFeed as Message>::Result;
fn handle(&mut self, msg: AddFeed, _: &mut Self::Context) -> Self::Result {
match self.feeds.entry(msg.url.clone()) {
std::collections::hash_map::Entry::Occupied(_) => Err("Feed already exists".into()),
std::collections::hash_map::Entry::Vacant(e) => {
debug!("will download {}", &msg.url);
self.downloader.send(DownloadFeed(msg.url))
.wait()
.or_err("Failed to download")?
.map(|feed| {
debug!("downloaded");
e.insert(feed);
})
}
}
}
}
impl Handler<RemoveFeed> for FeedStorage {
type Result = <RemoveFeed as Message>::Result;
fn handle(&mut self, msg: RemoveFeed, _: &mut Self::Context) -> Self::Result {
self.feeds.remove(&msg.url);
Ok(())
}
}
impl Handler<GetFeed> for FeedStorage {
type Result = <GetFeed as Message>::Result;
fn handle(&mut self, msg: GetFeed, _: &mut Self::Context) -> Self::Result {
match self.feeds.get(&msg.url) {
None => Err("Feed not found".into()),
Some(feed) => Ok(feed.clone()),
}
}
}
impl Handler<ListFeeds> for FeedStorage {
type Result = <ListFeeds as Message>::Result;
fn handle(&mut self, _: ListFeeds, _: &mut Self::Context) -> Self::Result {
Ok(self.feeds.iter().map(|(k, v)| FeedInfo{url: k.clone(), title: v.title.clone(), last_updated: v.last_updated.clone()}).collect())
}
}
impl Handler<MarkRead> for FeedStorage {
type Result = <MarkRead as Message>::Result;
fn handle(&mut self, msg: MarkRead, _: &mut Self::Context) -> Self::Result {
let mut updated = false;
if let Some(feed) = self.feeds.get_mut(&msg.url) {
for item in feed.items.iter_mut().filter(|k| &k.guid == &msg.guid).take(1) {
item.unread = false;
updated = true;
}
}
Ok(updated)
}
}
impl Handler<UpdateFeed> for FeedStorage {
type Result = <UpdateFeed as Message>::Result;
fn handle(&mut self, msg: UpdateFeed, _: &mut Self::Context) -> Self::Result {
if let Some(feed) = self.feeds.get_mut(&msg.url) {
feed.merge(msg.feed)
};
}
}
struct Downloader;
impl Actor for Downloader {
type Context = actix::Context<Self>;
}
impl Handler<DownloadFeed> for Downloader {
type Result = <DownloadFeed as Message>::Result;
fn handle(&mut self, msg: DownloadFeed, _: &mut Self::Context) -> Self::Result {
let channel = rss::Channel::from_url(&msg.0).or_err("Channel not downloaded")?;
let mut items = vec![];
for item in channel.items().iter() {
let guid = item.guid().or_err("broken channel")?.value();
items.push(Item {
title: item.title().map(|s| s.to_string()),
link: item.link().map(|s| s.to_string()),
content: item.content().or(item.description()).map(|s| s.to_string()),
pub_date: item.pub_date().and_then(|date| DateTime::parse_from_rfc2822(date).ok().map(|d| d.with_timezone(&Utc))),
guid: guid.to_string(),
unread: true,
});
}
Ok(Feed{
title: channel.title().to_owned(),
last_updated: match channel.last_build_date() {
None => items
.iter()
.map(|item| &item.pub_date)
.max()
.map(|date| date.to_owned())
.unwrap_or(Some(Utc.timestamp(0, 0)))
.unwrap_or(Utc.timestamp(0, 0)),
Some(s) => DateTime::parse_from_rfc2822(s).map(|d| d.with_timezone(&Utc)).unwrap_or(Utc.timestamp(0, 0))
},
items: items
})
}
}
struct Updater {
storage: actix::Addr<FeedStorage>,
downloader: actix::Addr<Downloader>,
handle: Option<actix::SpawnHandle>,
arbiter: actix::Arbiter,
}
impl Actor for Updater {
type Context = actix::Context<Self>;
fn started(&mut self, ctx: &mut <Self as Actor>::Context) {
let storage = self.storage.clone();
let downloader = self.downloader.clone();
let arbiter = self.arbiter.clone();
self.handle = Some(ctx.run_interval(std::time::Duration::new(60, 0), move |_, _| {
let storage = storage.clone();
let downloader = downloader.clone();
let arbiter = arbiter.clone();
arbiter.exec_fn(move || {
if let Ok(Ok(infos)) = storage.send(ListFeeds).wait() {
debug!("got {} feeds, updating", infos.len());
for info in infos {
if let Ok(Ok(new_feed)) = downloader.send(DownloadFeed(info.url.clone())).wait() {
if let Ok(()) = storage.send(UpdateFeed{url: info.url.clone(), feed: new_feed}).wait() {
debug!("successfully updated {}", info.url);
}
}
}
}
});
}));
}
}
fn process_response<T: Serialize, E: Serialize, E2, F: FnOnce(E) -> actix_web::Error>(response: Result<Result<T, E>, E2>, f: F) -> Result<HttpResponse, actix_web::Error> {
match response {
Ok(Ok(data)) => Ok(HttpResponse::Ok().json(data)),
Ok(Err(e)) => Err(f(e)),
_ => Err(actix_web::error::ErrorInternalServerError("Application overload"))
}
}
#[derive(Clone)]
struct State { storage: actix::Addr<FeedStorage> }
fn add_feed(url_info: web::Form<AddFeed>, data: web::Data<State>) -> impl Future<Item = HttpResponse, Error = actix_web::Error> {
web::block(move || data.storage.send(url_info.into_inner()).wait())
.then(|res| process_response(res, actix_web::error::ErrorInternalServerError))
}
fn remove_feed(url_info: web::Form<RemoveFeed>, data: web::Data<State>) -> impl Future<Item = HttpResponse, Error = actix_web::Error> {
web::block(move || data.storage.send(url_info.into_inner()). | Item | identifier_name |
|
sssmc39_scheme.rs | cover_secret(&self.member_shares, self.member_threshold)
}
}
/// Split a master secret into mnemonic shares
/// group_threshold: The number of groups required to reconstruct the master secret
/// groups: A list of (member_threshold, member_count) pairs for each group, where member_count
/// is the number of shares to generate for the group and member_threshold is the number of
/// members required to reconstruct the group secret.
/// master_secret: The master secret to split.
/// passphrase: The passphrase used to encrypt the master secret.
/// iteration_exponent: The iteration exponent.
/// return: List of mnemonics.
pub fn generate_mnemonics(
group_threshold: u8,
groups: &[(u8, u8)],
master_secret: &[u8],
passphrase: &str,
iteration_exponent: u8,
) -> Result<Vec<GroupShare>, Error> {
// Generate a 'proto share' so to speak, with identifer generated and group data filled
let mut proto_share = Share::new()?;
proto_share.group_threshold = group_threshold;
proto_share.group_count = groups.len() as u8;
if master_secret.len() * 8 < proto_share.config.min_strength_bits as usize {
return Err(ErrorKind::Value(format!(
"The length of the master secret ({} bytes) must be at least {} bytes.",
master_secret.len(),
(f64::from(proto_share.config.min_strength_bits) / 8f64).ceil(),
)))?;
}
if master_secret.len() % 2 != 0 {
return Err(ErrorKind::Value(
"The length of the master secret in bytes must be an even number".to_string(),
))?;
}
if group_threshold as usize > groups.len() {
return Err(ErrorKind::Value(format!(
"The requested group threshold ({}) must not exceed the number of groups ({}).",
group_threshold,
groups.len()
)))?;
}
let encoder = util::encrypt::MasterSecretEnc::new()?;
let encrypted_master_secret = encoder.encrypt(
master_secret,
passphrase,
iteration_exponent,
proto_share.identifier,
);
let sp = Splitter::new(None);
let group_shares = sp.split_secret(
&proto_share,
group_threshold,
groups.len() as u8,
&encrypted_master_secret,
)?;
let mut retval: Vec<GroupShare> = vec![];
let gs_len = group_shares.len();
for (i, elem) in group_shares.into_iter().enumerate() {
proto_share.group_index = i as u8;
proto_share.group_threshold = group_threshold;
proto_share.group_count = gs_len as u8;
let (member_threshold, member_count) = groups[i];
let member_shares = sp.split_secret(
&proto_share,
member_threshold,
member_count,
&elem.share_value,
)?;
retval.push(GroupShare {
group_id: proto_share.identifier,
iteration_exponent,
group_index: i as u8,
group_threshold,
group_count: gs_len as u8,
member_threshold,
member_shares,
});
}
Ok(retval)
}
pub fn generate_mnemonics_random(
group_threshold: u8,
groups: &[(u8, u8)],
strength_bits: u16,
passphrase: &str,
iteration_exponent: u8,
) -> Result<Vec<GroupShare>, Error> {
let proto_share = Share::new()?;
if strength_bits < proto_share.config.min_strength_bits {
return Err(ErrorKind::Value(format!(
"The requested strength of the master secret({} bits) must be at least {} bits.",
strength_bits, proto_share.config.min_strength_bits,
)))?;
}
if strength_bits % 16 != 0 {
return Err(ErrorKind::Value(format!(
"The requested strength of the master secret({} bits) must be a multiple of 16 bits.",
strength_bits,
)))?;
}
generate_mnemonics(
group_threshold,
groups,
&util::fill_vec_rand(strength_bits as usize / 8),
passphrase,
iteration_exponent,
)
}
/// Combines mnemonic shares to obtain the master secret which was previously split using
/// Shamir's secret sharing scheme.
/// mnemonics: List of mnemonics.
/// passphrase: The passphrase used to encrypt the master secret.
/// return: The master secret.
pub fn combine_mnemonics(mnemonics: &[Vec<String>], passphrase: &str) -> Result<Vec<u8>, Error> {
let group_shares = decode_mnemonics(mnemonics)?;
let mut shares = vec![];
for mut gs in group_shares {
shares.push(gs.decode_shares()?);
}
let sp = Splitter::new(None);
// restore proper member index for groups
let shares = shares
.into_iter()
.map(|mut s| {
s.member_index = s.group_index;
s
})
.collect::<Vec<_>>();
let ems = sp.recover_secret(&shares, shares[0].group_threshold)?;
let encoder = util::encrypt::MasterSecretEnc::new()?;
let dms = encoder.decrypt( | );
Ok(dms)
}
/// Decodes all Mnemonics to a list of shares and performs error checking
fn decode_mnemonics(mnemonics: &[Vec<String>]) -> Result<Vec<GroupShare>, Error> {
let mut shares = vec![];
if mnemonics.is_empty() {
return Err(ErrorKind::Mnemonic(
"List of mnemonics is empty.".to_string(),
))?;
}
let check_len = mnemonics[0].len();
for m in mnemonics {
if m.len() != check_len {
return Err(ErrorKind::Mnemonic(
"Invalid set of mnemonics. All mnemonics must have the same length.".to_string(),
))?;
}
shares.push(Share::from_mnemonic(m)?);
}
let check_share = shares[0].clone();
for s in shares.iter() {
if s.identifier != check_share.identifier
|| s.iteration_exponent != check_share.iteration_exponent
{
return Err(ErrorKind::Mnemonic(format!(
"Invalid set of mnemonics. All mnemonics must begin with the same {} words. \
(Identifier and iteration exponent must be the same).",
s.config.id_exp_length_words,
)))?;
}
if s.group_threshold != check_share.group_threshold {
return Err(ErrorKind::Mnemonic(
"Invalid set of mnemonics. All mnemonics must have the same group threshold"
.to_string(),
))?;
}
if s.group_count != check_share.group_count {
return Err(ErrorKind::Mnemonic(
"Invalid set of mnemonics. All mnemonics must have the same group count"
.to_string(),
))?;
}
}
let mut group_index_map = BTreeMap::new();
for s in shares {
if !group_index_map.contains_key(&s.group_index) {
let group_share = GroupShare {
group_id: s.identifier,
group_index: s.group_index,
group_threshold: s.group_threshold,
iteration_exponent: s.iteration_exponent,
group_count: s.group_count,
member_shares: vec![s.clone()],
member_threshold: s.member_threshold,
};
group_index_map.insert(group_share.group_index, group_share);
} else {
let e = group_index_map.get_mut(&s.group_index).unwrap();
e.member_shares.push(s);
}
}
if group_index_map.len() < check_share.group_threshold as usize {
return Err(ErrorKind::Mnemonic(format!(
"Insufficient number of mnemonic groups ({}). The required number \
of groups is {}.",
group_index_map.len(),
check_share.group_threshold,
)))?;
}
let groups: Vec<GroupShare> = group_index_map
.into_iter()
.map(|g| g.1)
// remove groups where number of shares is below the member threshold
.filter(|g| g.member_shares.len() >= g.member_threshold as usize)
.collect();
if groups.len() < check_share.group_threshold as usize {
return Err(ErrorKind::Mnemonic(
"Insufficient number of groups with member counts that meet member threshold."
.to_string(),
))?;
}
// TODO: Should probably return info making problem mnemonics easier to identify
for g in groups.iter() {
if g.member_shares.len() < g.member_threshold as usize {
return Err(ErrorKind::Mnemonic(format!(
"Insufficient number of mnemonics (Group {}). At least {} mnemonics \
are required.",
g.group_index, g.member_threshold,
)))?;
}
let test_share = g.member_shares[0].clone();
for ms in g.member_shares.iter() {
if test_share.member_threshold != ms.member_threshold {
return Err(ErrorKind::Mnemonic(
"Mismatching member thresholds".to_string(),
))?;
}
}
}
Ok(groups)
}
#[cfg(test)]
mod | &ems.share_value,
passphrase,
ems.iteration_exponent,
ems.identifier, | random_line_split |
sssmc39_scheme.rs | % 2 != 0 {
return Err(ErrorKind::Value(
"The length of the master secret in bytes must be an even number".to_string(),
))?;
}
if group_threshold as usize > groups.len() {
return Err(ErrorKind::Value(format!(
"The requested group threshold ({}) must not exceed the number of groups ({}).",
group_threshold,
groups.len()
)))?;
}
let encoder = util::encrypt::MasterSecretEnc::new()?;
let encrypted_master_secret = encoder.encrypt(
master_secret,
passphrase,
iteration_exponent,
proto_share.identifier,
);
let sp = Splitter::new(None);
let group_shares = sp.split_secret(
&proto_share,
group_threshold,
groups.len() as u8,
&encrypted_master_secret,
)?;
let mut retval: Vec<GroupShare> = vec![];
let gs_len = group_shares.len();
for (i, elem) in group_shares.into_iter().enumerate() {
proto_share.group_index = i as u8;
proto_share.group_threshold = group_threshold;
proto_share.group_count = gs_len as u8;
let (member_threshold, member_count) = groups[i];
let member_shares = sp.split_secret(
&proto_share,
member_threshold,
member_count,
&elem.share_value,
)?;
retval.push(GroupShare {
group_id: proto_share.identifier,
iteration_exponent,
group_index: i as u8,
group_threshold,
group_count: gs_len as u8,
member_threshold,
member_shares,
});
}
Ok(retval)
}
pub fn generate_mnemonics_random(
group_threshold: u8,
groups: &[(u8, u8)],
strength_bits: u16,
passphrase: &str,
iteration_exponent: u8,
) -> Result<Vec<GroupShare>, Error> {
let proto_share = Share::new()?;
if strength_bits < proto_share.config.min_strength_bits {
return Err(ErrorKind::Value(format!(
"The requested strength of the master secret({} bits) must be at least {} bits.",
strength_bits, proto_share.config.min_strength_bits,
)))?;
}
if strength_bits % 16 != 0 {
return Err(ErrorKind::Value(format!(
"The requested strength of the master secret({} bits) must be a multiple of 16 bits.",
strength_bits,
)))?;
}
generate_mnemonics(
group_threshold,
groups,
&util::fill_vec_rand(strength_bits as usize / 8),
passphrase,
iteration_exponent,
)
}
/// Combines mnemonic shares to obtain the master secret which was previously split using
/// Shamir's secret sharing scheme.
/// mnemonics: List of mnemonics.
/// passphrase: The passphrase used to encrypt the master secret.
/// return: The master secret.
pub fn combine_mnemonics(mnemonics: &[Vec<String>], passphrase: &str) -> Result<Vec<u8>, Error> {
let group_shares = decode_mnemonics(mnemonics)?;
let mut shares = vec![];
for mut gs in group_shares {
shares.push(gs.decode_shares()?);
}
let sp = Splitter::new(None);
// restore proper member index for groups
let shares = shares
.into_iter()
.map(|mut s| {
s.member_index = s.group_index;
s
})
.collect::<Vec<_>>();
let ems = sp.recover_secret(&shares, shares[0].group_threshold)?;
let encoder = util::encrypt::MasterSecretEnc::new()?;
let dms = encoder.decrypt(
&ems.share_value,
passphrase,
ems.iteration_exponent,
ems.identifier,
);
Ok(dms)
}
/// Decodes all Mnemonics to a list of shares and performs error checking
fn decode_mnemonics(mnemonics: &[Vec<String>]) -> Result<Vec<GroupShare>, Error> {
let mut shares = vec![];
if mnemonics.is_empty() {
return Err(ErrorKind::Mnemonic(
"List of mnemonics is empty.".to_string(),
))?;
}
let check_len = mnemonics[0].len();
for m in mnemonics {
if m.len() != check_len {
return Err(ErrorKind::Mnemonic(
"Invalid set of mnemonics. All mnemonics must have the same length.".to_string(),
))?;
}
shares.push(Share::from_mnemonic(m)?);
}
let check_share = shares[0].clone();
for s in shares.iter() {
if s.identifier != check_share.identifier
|| s.iteration_exponent != check_share.iteration_exponent
{
return Err(ErrorKind::Mnemonic(format!(
"Invalid set of mnemonics. All mnemonics must begin with the same {} words. \
(Identifier and iteration exponent must be the same).",
s.config.id_exp_length_words,
)))?;
}
if s.group_threshold != check_share.group_threshold {
return Err(ErrorKind::Mnemonic(
"Invalid set of mnemonics. All mnemonics must have the same group threshold"
.to_string(),
))?;
}
if s.group_count != check_share.group_count {
return Err(ErrorKind::Mnemonic(
"Invalid set of mnemonics. All mnemonics must have the same group count"
.to_string(),
))?;
}
}
let mut group_index_map = BTreeMap::new();
for s in shares {
if !group_index_map.contains_key(&s.group_index) {
let group_share = GroupShare {
group_id: s.identifier,
group_index: s.group_index,
group_threshold: s.group_threshold,
iteration_exponent: s.iteration_exponent,
group_count: s.group_count,
member_shares: vec![s.clone()],
member_threshold: s.member_threshold,
};
group_index_map.insert(group_share.group_index, group_share);
} else {
let e = group_index_map.get_mut(&s.group_index).unwrap();
e.member_shares.push(s);
}
}
if group_index_map.len() < check_share.group_threshold as usize {
return Err(ErrorKind::Mnemonic(format!(
"Insufficient number of mnemonic groups ({}). The required number \
of groups is {}.",
group_index_map.len(),
check_share.group_threshold,
)))?;
}
let groups: Vec<GroupShare> = group_index_map
.into_iter()
.map(|g| g.1)
// remove groups where number of shares is below the member threshold
.filter(|g| g.member_shares.len() >= g.member_threshold as usize)
.collect();
if groups.len() < check_share.group_threshold as usize {
return Err(ErrorKind::Mnemonic(
"Insufficient number of groups with member counts that meet member threshold."
.to_string(),
))?;
}
// TODO: Should probably return info making problem mnemonics easier to identify
for g in groups.iter() {
if g.member_shares.len() < g.member_threshold as usize {
return Err(ErrorKind::Mnemonic(format!(
"Insufficient number of mnemonics (Group {}). At least {} mnemonics \
are required.",
g.group_index, g.member_threshold,
)))?;
}
let test_share = g.member_shares[0].clone();
for ms in g.member_shares.iter() {
if test_share.member_threshold != ms.member_threshold {
return Err(ErrorKind::Mnemonic(
"Mismatching member thresholds".to_string(),
))?;
}
}
}
Ok(groups)
}
#[cfg(test)]
mod tests {
use super::*;
fn flatten_mnemonics(nms: &[GroupShare]) -> Result<Vec<Vec<String>>, Error> {
let mut ret = vec![];
for m in nms {
for s in m.member_shares.iter() {
ret.push(s.to_mnemonic()?);
}
}
Ok(ret)
}
#[test]
fn generate_mnemonics_test() -> Result<(), Error> | {
let master_secret = b"\x0c\x94\x90\xbcn\xd6\xbc\xbf\xac>\xbe}\xeeV\xf2P".to_vec();
// single 3 of 5 test, splat out all mnemonics
println!("Single 3 of 5 Encoded: {:?}", master_secret);
let mns = generate_mnemonics(1, &[(3, 5)], &master_secret, "", 0)?;
for s in &mns {
println!("{}", s);
}
let result = combine_mnemonics(&flatten_mnemonics(&mns)?, "")?;
println!("Single 3 of 5 Decoded: {:?}", result);
assert_eq!(result, master_secret);
// Test a few distinct groups
let mns = generate_mnemonics(
2,
&[(3, 5), (2, 5), (3, 3), (13, 16)],
&master_secret,
"",
0, | identifier_body |
|
sssmc39_scheme.rs | _secret(&self.member_shares, self.member_threshold)
}
}
/// Split a master secret into mnemonic shares
/// group_threshold: The number of groups required to reconstruct the master secret
/// groups: A list of (member_threshold, member_count) pairs for each group, where member_count
/// is the number of shares to generate for the group and member_threshold is the number of
/// members required to reconstruct the group secret.
/// master_secret: The master secret to split.
/// passphrase: The passphrase used to encrypt the master secret.
/// iteration_exponent: The iteration exponent.
/// return: List of mnemonics.
pub fn generate_mnemonics(
group_threshold: u8,
groups: &[(u8, u8)],
master_secret: &[u8],
passphrase: &str,
iteration_exponent: u8,
) -> Result<Vec<GroupShare>, Error> {
// Generate a 'proto share' so to speak, with identifer generated and group data filled
let mut proto_share = Share::new()?;
proto_share.group_threshold = group_threshold;
proto_share.group_count = groups.len() as u8;
if master_secret.len() * 8 < proto_share.config.min_strength_bits as usize {
return Err(ErrorKind::Value(format!(
"The length of the master secret ({} bytes) must be at least {} bytes.",
master_secret.len(),
(f64::from(proto_share.config.min_strength_bits) / 8f64).ceil(),
)))?;
}
if master_secret.len() % 2 != 0 {
return Err(ErrorKind::Value(
"The length of the master secret in bytes must be an even number".to_string(),
))?;
}
if group_threshold as usize > groups.len() {
return Err(ErrorKind::Value(format!(
"The requested group threshold ({}) must not exceed the number of groups ({}).",
group_threshold,
groups.len()
)))?;
}
let encoder = util::encrypt::MasterSecretEnc::new()?;
let encrypted_master_secret = encoder.encrypt(
master_secret,
passphrase,
iteration_exponent,
proto_share.identifier,
);
let sp = Splitter::new(None);
let group_shares = sp.split_secret(
&proto_share,
group_threshold,
groups.len() as u8,
&encrypted_master_secret,
)?;
let mut retval: Vec<GroupShare> = vec![];
let gs_len = group_shares.len();
for (i, elem) in group_shares.into_iter().enumerate() {
proto_share.group_index = i as u8;
proto_share.group_threshold = group_threshold;
proto_share.group_count = gs_len as u8;
let (member_threshold, member_count) = groups[i];
let member_shares = sp.split_secret(
&proto_share,
member_threshold,
member_count,
&elem.share_value,
)?;
retval.push(GroupShare {
group_id: proto_share.identifier,
iteration_exponent,
group_index: i as u8,
group_threshold,
group_count: gs_len as u8,
member_threshold,
member_shares,
});
}
Ok(retval)
}
pub fn generate_mnemonics_random(
group_threshold: u8,
groups: &[(u8, u8)],
strength_bits: u16,
passphrase: &str,
iteration_exponent: u8,
) -> Result<Vec<GroupShare>, Error> {
let proto_share = Share::new()?;
if strength_bits < proto_share.config.min_strength_bits {
return Err(ErrorKind::Value(format!(
"The requested strength of the master secret({} bits) must be at least {} bits.",
strength_bits, proto_share.config.min_strength_bits,
)))?;
}
if strength_bits % 16 != 0 {
return Err(ErrorKind::Value(format!(
"The requested strength of the master secret({} bits) must be a multiple of 16 bits.",
strength_bits,
)))?;
}
generate_mnemonics(
group_threshold,
groups,
&util::fill_vec_rand(strength_bits as usize / 8),
passphrase,
iteration_exponent,
)
}
/// Combines mnemonic shares to obtain the master secret which was previously split using
/// Shamir's secret sharing scheme.
/// mnemonics: List of mnemonics.
/// passphrase: The passphrase used to encrypt the master secret.
/// return: The master secret.
pub fn combine_mnemonics(mnemonics: &[Vec<String>], passphrase: &str) -> Result<Vec<u8>, Error> {
let group_shares = decode_mnemonics(mnemonics)?;
let mut shares = vec![];
for mut gs in group_shares {
shares.push(gs.decode_shares()?);
}
let sp = Splitter::new(None);
// restore proper member index for groups
let shares = shares
.into_iter()
.map(|mut s| {
s.member_index = s.group_index;
s
})
.collect::<Vec<_>>();
let ems = sp.recover_secret(&shares, shares[0].group_threshold)?;
let encoder = util::encrypt::MasterSecretEnc::new()?;
let dms = encoder.decrypt(
&ems.share_value,
passphrase,
ems.iteration_exponent,
ems.identifier,
);
Ok(dms)
}
/// Decodes all Mnemonics to a list of shares and performs error checking
fn decode_mnemonics(mnemonics: &[Vec<String>]) -> Result<Vec<GroupShare>, Error> {
let mut shares = vec![];
if mnemonics.is_empty() |
let check_len = mnemonics[0].len();
for m in mnemonics {
if m.len() != check_len {
return Err(ErrorKind::Mnemonic(
"Invalid set of mnemonics. All mnemonics must have the same length.".to_string(),
))?;
}
shares.push(Share::from_mnemonic(m)?);
}
let check_share = shares[0].clone();
for s in shares.iter() {
if s.identifier != check_share.identifier
|| s.iteration_exponent != check_share.iteration_exponent
{
return Err(ErrorKind::Mnemonic(format!(
"Invalid set of mnemonics. All mnemonics must begin with the same {} words. \
(Identifier and iteration exponent must be the same).",
s.config.id_exp_length_words,
)))?;
}
if s.group_threshold != check_share.group_threshold {
return Err(ErrorKind::Mnemonic(
"Invalid set of mnemonics. All mnemonics must have the same group threshold"
.to_string(),
))?;
}
if s.group_count != check_share.group_count {
return Err(ErrorKind::Mnemonic(
"Invalid set of mnemonics. All mnemonics must have the same group count"
.to_string(),
))?;
}
}
let mut group_index_map = BTreeMap::new();
for s in shares {
if !group_index_map.contains_key(&s.group_index) {
let group_share = GroupShare {
group_id: s.identifier,
group_index: s.group_index,
group_threshold: s.group_threshold,
iteration_exponent: s.iteration_exponent,
group_count: s.group_count,
member_shares: vec![s.clone()],
member_threshold: s.member_threshold,
};
group_index_map.insert(group_share.group_index, group_share);
} else {
let e = group_index_map.get_mut(&s.group_index).unwrap();
e.member_shares.push(s);
}
}
if group_index_map.len() < check_share.group_threshold as usize {
return Err(ErrorKind::Mnemonic(format!(
"Insufficient number of mnemonic groups ({}). The required number \
of groups is {}.",
group_index_map.len(),
check_share.group_threshold,
)))?;
}
let groups: Vec<GroupShare> = group_index_map
.into_iter()
.map(|g| g.1)
// remove groups where number of shares is below the member threshold
.filter(|g| g.member_shares.len() >= g.member_threshold as usize)
.collect();
if groups.len() < check_share.group_threshold as usize {
return Err(ErrorKind::Mnemonic(
"Insufficient number of groups with member counts that meet member threshold."
.to_string(),
))?;
}
// TODO: Should probably return info making problem mnemonics easier to identify
for g in groups.iter() {
if g.member_shares.len() < g.member_threshold as usize {
return Err(ErrorKind::Mnemonic(format!(
"Insufficient number of mnemonics (Group {}). At least {} mnemonics \
are required.",
g.group_index, g.member_threshold,
)))?;
}
let test_share = g.member_shares[0].clone();
for ms in g.member_shares.iter() {
if test_share.member_threshold != ms.member_threshold {
return Err(ErrorKind::Mnemonic(
"Mismatching member thresholds".to_string(),
))?;
}
}
}
Ok(groups)
}
#[cfg(test)]
| {
return Err(ErrorKind::Mnemonic(
"List of mnemonics is empty.".to_string(),
))?;
} | conditional_block |
sssmc39_scheme.rs | _secret(&self.member_shares, self.member_threshold)
}
}
/// Split a master secret into mnemonic shares
/// group_threshold: The number of groups required to reconstruct the master secret
/// groups: A list of (member_threshold, member_count) pairs for each group, where member_count
/// is the number of shares to generate for the group and member_threshold is the number of
/// members required to reconstruct the group secret.
/// master_secret: The master secret to split.
/// passphrase: The passphrase used to encrypt the master secret.
/// iteration_exponent: The iteration exponent.
/// return: List of mnemonics.
pub fn generate_mnemonics(
group_threshold: u8,
groups: &[(u8, u8)],
master_secret: &[u8],
passphrase: &str,
iteration_exponent: u8,
) -> Result<Vec<GroupShare>, Error> {
// Generate a 'proto share' so to speak, with identifer generated and group data filled
let mut proto_share = Share::new()?;
proto_share.group_threshold = group_threshold;
proto_share.group_count = groups.len() as u8;
if master_secret.len() * 8 < proto_share.config.min_strength_bits as usize {
return Err(ErrorKind::Value(format!(
"The length of the master secret ({} bytes) must be at least {} bytes.",
master_secret.len(),
(f64::from(proto_share.config.min_strength_bits) / 8f64).ceil(),
)))?;
}
if master_secret.len() % 2 != 0 {
return Err(ErrorKind::Value(
"The length of the master secret in bytes must be an even number".to_string(),
))?;
}
if group_threshold as usize > groups.len() {
return Err(ErrorKind::Value(format!(
"The requested group threshold ({}) must not exceed the number of groups ({}).",
group_threshold,
groups.len()
)))?;
}
let encoder = util::encrypt::MasterSecretEnc::new()?;
let encrypted_master_secret = encoder.encrypt(
master_secret,
passphrase,
iteration_exponent,
proto_share.identifier,
);
let sp = Splitter::new(None);
let group_shares = sp.split_secret(
&proto_share,
group_threshold,
groups.len() as u8,
&encrypted_master_secret,
)?;
let mut retval: Vec<GroupShare> = vec![];
let gs_len = group_shares.len();
for (i, elem) in group_shares.into_iter().enumerate() {
proto_share.group_index = i as u8;
proto_share.group_threshold = group_threshold;
proto_share.group_count = gs_len as u8;
let (member_threshold, member_count) = groups[i];
let member_shares = sp.split_secret(
&proto_share,
member_threshold,
member_count,
&elem.share_value,
)?;
retval.push(GroupShare {
group_id: proto_share.identifier,
iteration_exponent,
group_index: i as u8,
group_threshold,
group_count: gs_len as u8,
member_threshold,
member_shares,
});
}
Ok(retval)
}
pub fn generate_mnemonics_random(
group_threshold: u8,
groups: &[(u8, u8)],
strength_bits: u16,
passphrase: &str,
iteration_exponent: u8,
) -> Result<Vec<GroupShare>, Error> {
let proto_share = Share::new()?;
if strength_bits < proto_share.config.min_strength_bits {
return Err(ErrorKind::Value(format!(
"The requested strength of the master secret({} bits) must be at least {} bits.",
strength_bits, proto_share.config.min_strength_bits,
)))?;
}
if strength_bits % 16 != 0 {
return Err(ErrorKind::Value(format!(
"The requested strength of the master secret({} bits) must be a multiple of 16 bits.",
strength_bits,
)))?;
}
generate_mnemonics(
group_threshold,
groups,
&util::fill_vec_rand(strength_bits as usize / 8),
passphrase,
iteration_exponent,
)
}
/// Combines mnemonic shares to obtain the master secret which was previously split using
/// Shamir's secret sharing scheme.
/// mnemonics: List of mnemonics.
/// passphrase: The passphrase used to encrypt the master secret.
/// return: The master secret.
pub fn combine_mnemonics(mnemonics: &[Vec<String>], passphrase: &str) -> Result<Vec<u8>, Error> {
let group_shares = decode_mnemonics(mnemonics)?;
let mut shares = vec![];
for mut gs in group_shares {
shares.push(gs.decode_shares()?);
}
let sp = Splitter::new(None);
// restore proper member index for groups
let shares = shares
.into_iter()
.map(|mut s| {
s.member_index = s.group_index;
s
})
.collect::<Vec<_>>();
let ems = sp.recover_secret(&shares, shares[0].group_threshold)?;
let encoder = util::encrypt::MasterSecretEnc::new()?;
let dms = encoder.decrypt(
&ems.share_value,
passphrase,
ems.iteration_exponent,
ems.identifier,
);
Ok(dms)
}
/// Decodes all Mnemonics to a list of shares and performs error checking
fn | (mnemonics: &[Vec<String>]) -> Result<Vec<GroupShare>, Error> {
let mut shares = vec![];
if mnemonics.is_empty() {
return Err(ErrorKind::Mnemonic(
"List of mnemonics is empty.".to_string(),
))?;
}
let check_len = mnemonics[0].len();
for m in mnemonics {
if m.len() != check_len {
return Err(ErrorKind::Mnemonic(
"Invalid set of mnemonics. All mnemonics must have the same length.".to_string(),
))?;
}
shares.push(Share::from_mnemonic(m)?);
}
let check_share = shares[0].clone();
for s in shares.iter() {
if s.identifier != check_share.identifier
|| s.iteration_exponent != check_share.iteration_exponent
{
return Err(ErrorKind::Mnemonic(format!(
"Invalid set of mnemonics. All mnemonics must begin with the same {} words. \
(Identifier and iteration exponent must be the same).",
s.config.id_exp_length_words,
)))?;
}
if s.group_threshold != check_share.group_threshold {
return Err(ErrorKind::Mnemonic(
"Invalid set of mnemonics. All mnemonics must have the same group threshold"
.to_string(),
))?;
}
if s.group_count != check_share.group_count {
return Err(ErrorKind::Mnemonic(
"Invalid set of mnemonics. All mnemonics must have the same group count"
.to_string(),
))?;
}
}
let mut group_index_map = BTreeMap::new();
for s in shares {
if !group_index_map.contains_key(&s.group_index) {
let group_share = GroupShare {
group_id: s.identifier,
group_index: s.group_index,
group_threshold: s.group_threshold,
iteration_exponent: s.iteration_exponent,
group_count: s.group_count,
member_shares: vec![s.clone()],
member_threshold: s.member_threshold,
};
group_index_map.insert(group_share.group_index, group_share);
} else {
let e = group_index_map.get_mut(&s.group_index).unwrap();
e.member_shares.push(s);
}
}
if group_index_map.len() < check_share.group_threshold as usize {
return Err(ErrorKind::Mnemonic(format!(
"Insufficient number of mnemonic groups ({}). The required number \
of groups is {}.",
group_index_map.len(),
check_share.group_threshold,
)))?;
}
let groups: Vec<GroupShare> = group_index_map
.into_iter()
.map(|g| g.1)
// remove groups where number of shares is below the member threshold
.filter(|g| g.member_shares.len() >= g.member_threshold as usize)
.collect();
if groups.len() < check_share.group_threshold as usize {
return Err(ErrorKind::Mnemonic(
"Insufficient number of groups with member counts that meet member threshold."
.to_string(),
))?;
}
// TODO: Should probably return info making problem mnemonics easier to identify
for g in groups.iter() {
if g.member_shares.len() < g.member_threshold as usize {
return Err(ErrorKind::Mnemonic(format!(
"Insufficient number of mnemonics (Group {}). At least {} mnemonics \
are required.",
g.group_index, g.member_threshold,
)))?;
}
let test_share = g.member_shares[0].clone();
for ms in g.member_shares.iter() {
if test_share.member_threshold != ms.member_threshold {
return Err(ErrorKind::Mnemonic(
"Mismatching member thresholds".to_string(),
))?;
}
}
}
Ok(groups)
}
#[cfg(test)]
mod | decode_mnemonics | identifier_name |
project.py | ", type(d["pick_pose"]), "place_pose", type(d["place_pose"])
data_dict = {"object_list": dict_list}
with open(yaml_filename, 'w') as outfile:
yaml.dump(data_dict, outfile, default_flow_style=False)
def statistical_outlier_removal(cloud):
# Much like the previous filters, we start by creating a filter object:
outlier_filter = cloud.make_statistical_outlier_filter()
# Set the number of neighboring points to analyze for any given point
outlier_filter.set_mean_k(50)
# Set threshold scale factor
x = 0.9
# Any point with a mean distance larger than global (mean distance+x*std_dev) will be considered outlier
outlier_filter.set_std_dev_mul_thresh(x)
# Finally call the filter function for magic
cloud_filtered = outlier_filter.filter()
return cloud_filtered
def voxel_downsample(cloud): | cloud (PointCloud_PointXYZRGB): A point cloud
Returns:
PointCloud_PointXYZRGB: A downsampled point cloud
"""
# Create a VoxelGrid filter object for our input point cloud
vox = cloud.make_voxel_grid_filter()
# Choose a voxel (also known as leaf) size
LEAF_SIZE = 0.005
# Set the voxel (or leaf) size
vox.set_leaf_size(LEAF_SIZE, LEAF_SIZE, LEAF_SIZE)
# Call the filter function to obtain the resultant downsampled point cloud
cloud_filtered = vox.filter()
return cloud_filtered
def apply_passthrough_filter(cloud, axis, axis_min, axis_max):
""" Apply a passthrough filter to a cloud
Args:
cloud (PointCloud_PointXYZRGB): A point cloud
Returns:
PointCloud_PointXYZRGB: A filtered point cloud
"""
# Create a PassThrough filter object.
passthrough = cloud.make_passthrough_filter()
# Assign axis and range to the passthrough filter object.
filter_axis = axis
passthrough.set_filter_field_name(filter_axis)
#axis_min = 0.6
#axis_max = 1.1
passthrough.set_filter_limits(axis_min, axis_max)
# Finally use the filter function to obtain the resultant point cloud.
cloud_filtered = passthrough.filter()
return cloud_filtered
def ransac(cloud, sacmodel):
""" Segments a cloud using a sac model
Args:
cloud (PointCloud_PointXYZRGB): A point cloud
sacmodel (pcl.SACMODEL): A model points will be fit to
Returns:
A set of inliers and coefficients
"""
# Create the segmentation object
seg = cloud.make_segmenter()
# Set the model you wish to fit
seg.set_model_type(sacmodel)
seg.set_method_type(pcl.SAC_RANSAC)
# Max distance for a point to be considered fitting the model
# Experiment with different values for max_distance
# for segmenting the table
max_distance = 0.01
seg.set_distance_threshold(max_distance)
# Call the segment function to obtain set of inlier indices and model coefficients
inliers, coefficients = seg.segment()
return inliers, coefficients
def euclidean_clustering(cloud):
white_cloud = XYZRGB_to_XYZ(cloud)
tree = white_cloud.make_kdtree()
# Create a cluster extraction object
ec = white_cloud.make_EuclideanClusterExtraction()
# Set tolerances for distance threshold
# as well as minimum and maximum cluster size (in points)
ec.set_ClusterTolerance(0.01)
ec.set_MinClusterSize(10)
ec.set_MaxClusterSize(25000)
# Search the k-d tree for clusters
ec.set_SearchMethod(tree)
# Extract indices for each of the discovered clusters
cluster_indices = ec.Extract()
return cluster_indices, white_cloud
def color_clusters(cluster_indices, white_cloud):
#Assign a color corresponding to each segmented object in scene
cluster_color = get_color_list(len(cluster_indices))
color_cluster_point_list = []
for j, indices in enumerate(cluster_indices):
for i, indice in enumerate(indices):
color_cluster_point_list.append([white_cloud[indice][0],
white_cloud[indice][1],
white_cloud[indice][2],
rgb_to_float(cluster_color[j])])
#Create new cloud containing all clusters, each with unique color
cluster_cloud = pcl.PointCloud_PointXYZRGB()
cluster_cloud.from_list(color_cluster_point_list)
return cluster_cloud
# Callback function for your Point Cloud Subscriber
def pcl_callback(pcl_msg):
# Exercise-2 TODOs:
# Convert ROS msg to PCL data
pcl_data = ros_to_pcl(pcl_msg)
#filename = 'pcl_data.pcd'
#pcl.save(pcl_data, filename)
# Statistical Outlier Filtering
cloud_filtered = statistical_outlier_removal(pcl_data)
#filename = 'statistical_outlier_removal.pcd'
#pcl.save(cloud_filtered, filename)
# Voxel Grid Downsampling
cloud_filtered = voxel_downsample(cloud_filtered)
#filename = 'voxel_downsampled.pcd'
#pcl.save(cloud_filtered, filename)
# PassThrough Filter along z
axis_min = 0.6
axis_max = 1.1
cloud_filtered = apply_passthrough_filter(cloud_filtered, 'z', axis_min, axis_max)
#filename = 'pass_through_filtered.pcd'
#pcl.save(cloud_filtered, filename)
# PassThrough Filter along y
axis_min = -0.5
axis_max = 0.5
cloud_filtered = apply_passthrough_filter(cloud_filtered, 'y', axis_min, axis_max)
filename = 'pass_through_filtered_y.pcd'
pcl.save(cloud_filtered, filename)
# RANSAC Plane Segmentation
inliers, coefficients = ransac(cloud_filtered, pcl.SACMODEL_PLANE)
# Extract inliers and outliers
extracted_inliers = cloud_filtered.extract(inliers, negative=False)
#filename = 'extracted_inliers.pcd'
#pcl.save(extracted_inliers, filename)
extracted_outliers = cloud_filtered.extract(inliers, negative=True)
#filename = 'extracted_outliers.pcd'
#pcl.save(extracted_outliers, filename)
cloud_table = extracted_inliers
cloud_objects = extracted_outliers
# Euclidean Clustering
cluster_indices, white_cloud = euclidean_clustering(cloud_objects)
# Create Cluster-Mask Point Cloud to visualize each cluster separately
cluster_cloud = color_clusters(cluster_indices, white_cloud)
filename = 'colored_cluster_cloud.pcd'
pcl.save(cluster_cloud, filename)
# Convert PCL data to ROS messages
ros_cloud_table = pcl_to_ros(cloud_table)
ros_cloud_objects = pcl_to_ros(cloud_objects)
ros_cluster_cloud = pcl_to_ros(cluster_cloud)
# Publish ROS messages
pcl_objects_pub.publish(ros_cloud_objects)
pcl_table_pub.publish(ros_cloud_table)
pcl_cluster_pub.publish(ros_cluster_cloud)
# Exercise-3 TODOs:
# Classify the clusters! (loop through each detected cluster one at a time)
detected_objects_labels = []
detected_objects = []
for index, pts_list in enumerate(cluster_indices):
# Grab the points for the cluster
pcl_cluster = cloud_objects.extract(pts_list)
# Compute the associated feature vector
ros_cluster = pcl_to_ros(pcl_cluster)
sample_cloud = ros_cluster
chists = compute_color_histograms(sample_cloud, using_hsv=True)
normals = get_normals(sample_cloud)
nhists = compute_normal_histograms(normals)
feature = np.concatenate((chists, nhists))
# Make the prediction
prediction = clf.predict(scaler.transform(feature.reshape(1,-1)))
label = encoder.inverse_transform(prediction)[0]
detected_objects_labels.append(label)
# Publish a label into RViz
label_pos = list(white_cloud[pts_list[0]])
label_pos[2] += .4
object_markers_pub.publish(make_label(label,label_pos, index))
# Add the detected object to the list of detected objects.
do = DetectedObject()
do.label = label
do.cloud = ros_cluster
detected_objects.append(do)
rospy.loginfo('Detected {} objects: {}'.format(len(detected_objects_labels), detected_objects_labels))
# Publish the list of detected objects
detected_objects_pub.publish(detected_objects)
# Suggested location for where to invoke your pr2_mover() function within pcl_callback()
# Could add some logic to determine whether or not your object detections are robust
# before calling pr2_mover()
detected_objects_list = detected_objects
try:
pr2_mover(detected_objects_list)
except rospy.ROSInterruptException:
pass
# function to load parameters and request PickPlace service
def pr2_mover(object_list):
# TODO: Initialize variables
dict_list = []
# TODO: Get/Read parameters
object_list_param = rospy.get_param('/object_list')
dropbox_list_param = rospy.get_param('/dropbox')
# TODO: Parse parameters into individual variables
test_scene_num = Int32()
test_scene_num.data = 3
""" labels = []
centroids = [] # to | """ Voxel Grid filter
Args: | random_line_split |
project.py | _pose", type(d["pick_pose"]), "place_pose", type(d["place_pose"])
data_dict = {"object_list": dict_list}
with open(yaml_filename, 'w') as outfile:
yaml.dump(data_dict, outfile, default_flow_style=False)
def statistical_outlier_removal(cloud):
# Much like the previous filters, we start by creating a filter object:
outlier_filter = cloud.make_statistical_outlier_filter()
# Set the number of neighboring points to analyze for any given point
outlier_filter.set_mean_k(50)
# Set threshold scale factor
x = 0.9
# Any point with a mean distance larger than global (mean distance+x*std_dev) will be considered outlier
outlier_filter.set_std_dev_mul_thresh(x)
# Finally call the filter function for magic
cloud_filtered = outlier_filter.filter()
return cloud_filtered
def voxel_downsample(cloud):
""" Voxel Grid filter
Args:
cloud (PointCloud_PointXYZRGB): A point cloud
Returns:
PointCloud_PointXYZRGB: A downsampled point cloud
"""
# Create a VoxelGrid filter object for our input point cloud
vox = cloud.make_voxel_grid_filter()
# Choose a voxel (also known as leaf) size
LEAF_SIZE = 0.005
# Set the voxel (or leaf) size
vox.set_leaf_size(LEAF_SIZE, LEAF_SIZE, LEAF_SIZE)
# Call the filter function to obtain the resultant downsampled point cloud
cloud_filtered = vox.filter()
return cloud_filtered
def apply_passthrough_filter(cloud, axis, axis_min, axis_max):
""" Apply a passthrough filter to a cloud
Args:
cloud (PointCloud_PointXYZRGB): A point cloud
Returns:
PointCloud_PointXYZRGB: A filtered point cloud
"""
# Create a PassThrough filter object.
passthrough = cloud.make_passthrough_filter()
# Assign axis and range to the passthrough filter object.
filter_axis = axis
passthrough.set_filter_field_name(filter_axis)
#axis_min = 0.6
#axis_max = 1.1
passthrough.set_filter_limits(axis_min, axis_max)
# Finally use the filter function to obtain the resultant point cloud.
cloud_filtered = passthrough.filter()
return cloud_filtered
def ransac(cloud, sacmodel):
| seg.set_distance_threshold(max_distance)
# Call the segment function to obtain set of inlier indices and model coefficients
inliers, coefficients = seg.segment()
return inliers, coefficients
def euclidean_clustering(cloud):
white_cloud = XYZRGB_to_XYZ(cloud)
tree = white_cloud.make_kdtree()
# Create a cluster extraction object
ec = white_cloud.make_EuclideanClusterExtraction()
# Set tolerances for distance threshold
# as well as minimum and maximum cluster size (in points)
ec.set_ClusterTolerance(0.01)
ec.set_MinClusterSize(10)
ec.set_MaxClusterSize(25000)
# Search the k-d tree for clusters
ec.set_SearchMethod(tree)
# Extract indices for each of the discovered clusters
cluster_indices = ec.Extract()
return cluster_indices, white_cloud
def color_clusters(cluster_indices, white_cloud):
#Assign a color corresponding to each segmented object in scene
cluster_color = get_color_list(len(cluster_indices))
color_cluster_point_list = []
for j, indices in enumerate(cluster_indices):
for i, indice in enumerate(indices):
color_cluster_point_list.append([white_cloud[indice][0],
white_cloud[indice][1],
white_cloud[indice][2],
rgb_to_float(cluster_color[j])])
#Create new cloud containing all clusters, each with unique color
cluster_cloud = pcl.PointCloud_PointXYZRGB()
cluster_cloud.from_list(color_cluster_point_list)
return cluster_cloud
# Callback function for your Point Cloud Subscriber
def pcl_callback(pcl_msg):
# Exercise-2 TODOs:
# Convert ROS msg to PCL data
pcl_data = ros_to_pcl(pcl_msg)
#filename = 'pcl_data.pcd'
#pcl.save(pcl_data, filename)
# Statistical Outlier Filtering
cloud_filtered = statistical_outlier_removal(pcl_data)
#filename = 'statistical_outlier_removal.pcd'
#pcl.save(cloud_filtered, filename)
# Voxel Grid Downsampling
cloud_filtered = voxel_downsample(cloud_filtered)
#filename = 'voxel_downsampled.pcd'
#pcl.save(cloud_filtered, filename)
# PassThrough Filter along z
axis_min = 0.6
axis_max = 1.1
cloud_filtered = apply_passthrough_filter(cloud_filtered, 'z', axis_min, axis_max)
#filename = 'pass_through_filtered.pcd'
#pcl.save(cloud_filtered, filename)
# PassThrough Filter along y
axis_min = -0.5
axis_max = 0.5
cloud_filtered = apply_passthrough_filter(cloud_filtered, 'y', axis_min, axis_max)
filename = 'pass_through_filtered_y.pcd'
pcl.save(cloud_filtered, filename)
# RANSAC Plane Segmentation
inliers, coefficients = ransac(cloud_filtered, pcl.SACMODEL_PLANE)
# Extract inliers and outliers
extracted_inliers = cloud_filtered.extract(inliers, negative=False)
#filename = 'extracted_inliers.pcd'
#pcl.save(extracted_inliers, filename)
extracted_outliers = cloud_filtered.extract(inliers, negative=True)
#filename = 'extracted_outliers.pcd'
#pcl.save(extracted_outliers, filename)
cloud_table = extracted_inliers
cloud_objects = extracted_outliers
# Euclidean Clustering
cluster_indices, white_cloud = euclidean_clustering(cloud_objects)
# Create Cluster-Mask Point Cloud to visualize each cluster separately
cluster_cloud = color_clusters(cluster_indices, white_cloud)
filename = 'colored_cluster_cloud.pcd'
pcl.save(cluster_cloud, filename)
# Convert PCL data to ROS messages
ros_cloud_table = pcl_to_ros(cloud_table)
ros_cloud_objects = pcl_to_ros(cloud_objects)
ros_cluster_cloud = pcl_to_ros(cluster_cloud)
# Publish ROS messages
pcl_objects_pub.publish(ros_cloud_objects)
pcl_table_pub.publish(ros_cloud_table)
pcl_cluster_pub.publish(ros_cluster_cloud)
# Exercise-3 TODOs:
# Classify the clusters! (loop through each detected cluster one at a time)
detected_objects_labels = []
detected_objects = []
for index, pts_list in enumerate(cluster_indices):
# Grab the points for the cluster
pcl_cluster = cloud_objects.extract(pts_list)
# Compute the associated feature vector
ros_cluster = pcl_to_ros(pcl_cluster)
sample_cloud = ros_cluster
chists = compute_color_histograms(sample_cloud, using_hsv=True)
normals = get_normals(sample_cloud)
nhists = compute_normal_histograms(normals)
feature = np.concatenate((chists, nhists))
# Make the prediction
prediction = clf.predict(scaler.transform(feature.reshape(1,-1)))
label = encoder.inverse_transform(prediction)[0]
detected_objects_labels.append(label)
# Publish a label into RViz
label_pos = list(white_cloud[pts_list[0]])
label_pos[2] += .4
object_markers_pub.publish(make_label(label,label_pos, index))
# Add the detected object to the list of detected objects.
do = DetectedObject()
do.label = label
do.cloud = ros_cluster
detected_objects.append(do)
rospy.loginfo('Detected {} objects: {}'.format(len(detected_objects_labels), detected_objects_labels))
# Publish the list of detected objects
detected_objects_pub.publish(detected_objects)
# Suggested location for where to invoke your pr2_mover() function within pcl_callback()
# Could add some logic to determine whether or not your object detections are robust
# before calling pr2_mover()
detected_objects_list = detected_objects
try:
pr2_mover(detected_objects_list)
except rospy.ROSInterruptException:
pass
# function to load parameters and request PickPlace service
def pr2_mover(object_list):
# TODO: Initialize variables
dict_list = []
# TODO: Get/Read parameters
object_list_param = rospy.get_param('/object_list')
dropbox_list_param = rospy.get_param('/dropbox')
# TODO: Parse parameters into individual variables
test_scene_num = Int32()
test_scene_num.data = 3
""" labels = []
centroids = [] # to | """ Segments a cloud using a sac model
Args:
cloud (PointCloud_PointXYZRGB): A point cloud
sacmodel (pcl.SACMODEL): A model points will be fit to
Returns:
A set of inliers and coefficients
"""
# Create the segmentation object
seg = cloud.make_segmenter()
# Set the model you wish to fit
seg.set_model_type(sacmodel)
seg.set_method_type(pcl.SAC_RANSAC)
# Max distance for a point to be considered fitting the model
# Experiment with different values for max_distance
# for segmenting the table
max_distance = 0.01 | identifier_body |
project.py | _cluster_point_list.append([white_cloud[indice][0],
white_cloud[indice][1],
white_cloud[indice][2],
rgb_to_float(cluster_color[j])])
#Create new cloud containing all clusters, each with unique color
cluster_cloud = pcl.PointCloud_PointXYZRGB()
cluster_cloud.from_list(color_cluster_point_list)
return cluster_cloud
# Callback function for your Point Cloud Subscriber
def pcl_callback(pcl_msg):
# Exercise-2 TODOs:
# Convert ROS msg to PCL data
pcl_data = ros_to_pcl(pcl_msg)
#filename = 'pcl_data.pcd'
#pcl.save(pcl_data, filename)
# Statistical Outlier Filtering
cloud_filtered = statistical_outlier_removal(pcl_data)
#filename = 'statistical_outlier_removal.pcd'
#pcl.save(cloud_filtered, filename)
# Voxel Grid Downsampling
cloud_filtered = voxel_downsample(cloud_filtered)
#filename = 'voxel_downsampled.pcd'
#pcl.save(cloud_filtered, filename)
# PassThrough Filter along z
axis_min = 0.6
axis_max = 1.1
cloud_filtered = apply_passthrough_filter(cloud_filtered, 'z', axis_min, axis_max)
#filename = 'pass_through_filtered.pcd'
#pcl.save(cloud_filtered, filename)
# PassThrough Filter along y
axis_min = -0.5
axis_max = 0.5
cloud_filtered = apply_passthrough_filter(cloud_filtered, 'y', axis_min, axis_max)
filename = 'pass_through_filtered_y.pcd'
pcl.save(cloud_filtered, filename)
# RANSAC Plane Segmentation
inliers, coefficients = ransac(cloud_filtered, pcl.SACMODEL_PLANE)
# Extract inliers and outliers
extracted_inliers = cloud_filtered.extract(inliers, negative=False)
#filename = 'extracted_inliers.pcd'
#pcl.save(extracted_inliers, filename)
extracted_outliers = cloud_filtered.extract(inliers, negative=True)
#filename = 'extracted_outliers.pcd'
#pcl.save(extracted_outliers, filename)
cloud_table = extracted_inliers
cloud_objects = extracted_outliers
# Euclidean Clustering
cluster_indices, white_cloud = euclidean_clustering(cloud_objects)
# Create Cluster-Mask Point Cloud to visualize each cluster separately
cluster_cloud = color_clusters(cluster_indices, white_cloud)
filename = 'colored_cluster_cloud.pcd'
pcl.save(cluster_cloud, filename)
# Convert PCL data to ROS messages
ros_cloud_table = pcl_to_ros(cloud_table)
ros_cloud_objects = pcl_to_ros(cloud_objects)
ros_cluster_cloud = pcl_to_ros(cluster_cloud)
# Publish ROS messages
pcl_objects_pub.publish(ros_cloud_objects)
pcl_table_pub.publish(ros_cloud_table)
pcl_cluster_pub.publish(ros_cluster_cloud)
# Exercise-3 TODOs:
# Classify the clusters! (loop through each detected cluster one at a time)
detected_objects_labels = []
detected_objects = []
for index, pts_list in enumerate(cluster_indices):
# Grab the points for the cluster
pcl_cluster = cloud_objects.extract(pts_list)
# Compute the associated feature vector
ros_cluster = pcl_to_ros(pcl_cluster)
sample_cloud = ros_cluster
chists = compute_color_histograms(sample_cloud, using_hsv=True)
normals = get_normals(sample_cloud)
nhists = compute_normal_histograms(normals)
feature = np.concatenate((chists, nhists))
# Make the prediction
prediction = clf.predict(scaler.transform(feature.reshape(1,-1)))
label = encoder.inverse_transform(prediction)[0]
detected_objects_labels.append(label)
# Publish a label into RViz
label_pos = list(white_cloud[pts_list[0]])
label_pos[2] += .4
object_markers_pub.publish(make_label(label,label_pos, index))
# Add the detected object to the list of detected objects.
do = DetectedObject()
do.label = label
do.cloud = ros_cluster
detected_objects.append(do)
rospy.loginfo('Detected {} objects: {}'.format(len(detected_objects_labels), detected_objects_labels))
# Publish the list of detected objects
detected_objects_pub.publish(detected_objects)
# Suggested location for where to invoke your pr2_mover() function within pcl_callback()
# Could add some logic to determine whether or not your object detections are robust
# before calling pr2_mover()
detected_objects_list = detected_objects
try:
pr2_mover(detected_objects_list)
except rospy.ROSInterruptException:
pass
# function to load parameters and request PickPlace service
def pr2_mover(object_list):
# TODO: Initialize variables
dict_list = []
# TODO: Get/Read parameters
object_list_param = rospy.get_param('/object_list')
dropbox_list_param = rospy.get_param('/dropbox')
# TODO: Parse parameters into individual variables
test_scene_num = Int32()
test_scene_num.data = 3
""" labels = []
centroids = [] # to be list of tuples (x, y, z)
for object in object_list:
labels.append(object.label)
points_arr = ros_to_pcl(object.cloud).to_array()
centroids.append(np.mean(points_arr, axis=0)[:3])
"""
# TODO: Rotate PR2 in place to capture side tables for the collision map
# TODO: Loop through the pick list
for object in object_list:
# TODO: Get the PointCloud for a given object and obtain it's centroid
points_arr = ros_to_pcl(object.cloud).to_array()
c = np.mean(points_arr, axis=0)[:3]
centroid = map(np.asscalar, c)
print "type(centroid)", type(centroid), "[0]", type(centroid[0]), "[1]", type(centroid[1]), "[2]", type(centroid[2])
#centroids.append(np.mean(points_arr, axis=0)[:3])
object_name = String()
object_name.data = str(object.label)
# TODO: Create 'place_pose' for the object
group = None
for o in object_list_param:
if object.label == o['name']:
group = o['group']
print "for ", o['name'], "group found",group
object_arm_name = String()
place_pose = Pose()
for box in dropbox_list_param:
if group == box['group']:
# Assign the arm to be used for pick_place
object_arm_name.data = box['name']
place_pose.position.x = box['position'][0]
place_pose.position.y = box['position'][1]
place_pose.position.z = box['position'][2]
print "type(place_pose.x,y,z): (", type(place_pose.position.x), ",", type(place_pose.position.y), ",", type(place_pose.position.z), ")"
pick_pose = Pose()
pick_pose.position.x = centroid[0]
pick_pose.position.y = centroid[1]
pick_pose.position.z = centroid[2]
# TODO: Create a list of dictionaries (made with make_yaml_dict()) for later output to yaml format
yaml_dict = make_yaml_dict(test_scene_num, object_arm_name, object_name, pick_pose, place_pose)
dict_list.append(yaml_dict)
# Wait for 'pick_place_routine' service to come up
rospy.wait_for_service('pick_place_routine')
try:
pick_place_routine = rospy.ServiceProxy('pick_place_routine', PickPlace)
# Insert your message variables to be sent as a service request
resp = pick_place_routine(test_scene_num, object_name, object_arm_name, pick_pose, place_pose)
print ("Response: ",resp.success)
except rospy.ServiceException, e:
print "Service call failed: %s"%e
# Output your request parameters into output yaml file
yaml_filename = "output_" + str(test_scene_num.data) + ".yaml"
send_to_yaml(yaml_filename, dict_list)
if __name__ == '__main__':
# ROS node initialization
rospy.init_node('clustering', anonymous=True)
# Create Subscribers
#pcl_sub = rospy.Subscriber("/sensor_stick/point_cloud", pc2.PointCloud2, pcl_callback, queue_size=1)
pcl_cam_sub = rospy.Subscriber("/pr2/world/points", pc2.PointCloud2, pcl_callback, queue_size=1)
# Create Publishers
pcl_objects_pub = rospy.Publisher("/pcl_objects", PointCloud2, queue_size=1)
pcl_table_pub = rospy.Publisher("/pcl_table", PointCloud2, queue_size=1)
pcl_cluster_pub = rospy.Publisher("/pcl_cluster", PointCloud2, queue_size=1)
object_markers_pub = rospy.Publisher("/object_markers", Marker, queue_size=1)
detected_objects_pub = rospy.Publisher("/detected_objects", DetectedObjectsArray, queue_size=1)
# Load Model From disk
model = pickle.load(open('model.sav', 'rb'))
clf = model['classifier']
encoder = LabelEncoder()
encoder.classes_ = model['classes']
scaler = model['scaler']
# Initialize color_list
get_color_list.color_list = []
# Spin while node is not shutdown
while not rospy.is_shutdown():
| rospy.spin() | conditional_block |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.