file_name
large_stringlengths 4
140
| prefix
large_stringlengths 0
12.1k
| suffix
large_stringlengths 0
12k
| middle
large_stringlengths 0
7.51k
| fim_type
large_stringclasses 4
values |
---|---|---|---|---|
lib.rs | fn from_str(raw: &str) -> Result<Mime, ()> {
if raw == "*/*" {
return Ok(mime!(Star/Star));
}
let ascii = raw.to_ascii_lowercase(); // lifetimes :(
let len = ascii.len();
let mut iter = ascii.chars().enumerate();
let mut params = vec![];
// toplevel
let mut start;
let top;
loop {
match iter.next() {
Some((0, c)) if is_restricted_name_first_char(c) => (),
Some((i, c)) if i > 0 && is_restricted_name_char(c) => (),
Some((i, '/')) if i > 0 => match FromStr::from_str(&ascii[..i]) {
Ok(t) => {
top = t;
start = i + 1;
break;
}
Err(_) => return Err(())
},
_ => return Err(()) // EOF and no toplevel is no Mime
};
}
// sublevel
let sub;
let mut sub_star = false;
loop {
match iter.next() {
Some((i, '*')) if i == start => {
sub_star = true;
},
Some((i, c)) if i == start && is_restricted_name_first_char(c) => (),
Some((i, c)) if !sub_star && i > start && is_restricted_name_char(c) => (),
Some((i, ';')) if i > start => match FromStr::from_str(&ascii[start..i]) {
Ok(s) => {
sub = s;
start = i + 1;
break;
}
Err(_) => return Err(())
},
None => match FromStr::from_str(&ascii[start..]) {
Ok(s) => return Ok(Mime(top, s, params)),
Err(_) => return Err(())
},
_ => return Err(())
};
}
// params
debug!("starting params, len={}", len);
loop {
match param_from_str(raw, &ascii, &mut iter, start) {
Some((p, end)) => {
params.push(p);
start = end;
if start >= len {
break;
}
}
None => break
}
}
Ok(Mime(top, sub, params))
}
}
#[cfg(feature = "serde")]
impl serde::ser::Serialize for Mime {
fn serialize<S>(&self, serializer: &mut S) -> Result<(), S::Error>
where S: serde::ser::Serializer
{
serializer.serialize_str(&*format!("{}",self))
}
}
#[cfg(feature = "serde")]
impl serde::de::Deserialize for Mime {
fn deserialize<D>(deserializer: &mut D) -> Result<Self, D::Error>
where D: serde::de::Deserializer
{
let string: String = try!(serde::Deserialize::deserialize(deserializer));
let mime: Mime = match FromStr::from_str(&*string) {
Ok(mime) => mime,
Err(_) => return Err(serde::de::Error::custom("Invalid serialized mime")),
};
Ok(mime)
}
}
fn param_from_str(raw: &str, ascii: &str, iter: &mut Enumerate<Chars>, mut start: usize) -> Option<(Param, usize)> {
let attr;
debug!("param_from_str, start={}", start);
loop {
match iter.next() {
Some((i, ' ')) if i == start => start = i + 1,
Some((i, c)) if i == start && is_restricted_name_first_char(c) => (),
Some((i, c)) if i > start && is_restricted_name_char(c) => (),
Some((i, '=')) if i > start => match FromStr::from_str(&ascii[start..i]) {
Ok(a) => {
attr = a;
start = i + 1;
break;
},
Err(_) => return None
},
_ => return None
}
}
let value;
// values must be restrict-name-char or "anything goes"
let mut is_quoted = false;
{
let substr = |a,b| { if attr==Attr::Charset { &ascii[a..b] } else { &raw[a..b] } };
let endstr = |a| { if attr==Attr::Charset { &ascii[a..] } else { &raw[a..] } };
loop {
match iter.next() {
Some((i, '"')) if i == start => {
debug!("quoted");
is_quoted = true;
start = i + 1;
},
Some((i, c)) if i == start && is_restricted_name_first_char(c) => (),
Some((i, '"')) if i > start && is_quoted => match FromStr::from_str(substr(start,i)) {
Ok(v) => {
value = v;
start = i + 1;
break;
},
Err(_) => return None
},
Some((i, c)) if i > start && is_quoted || is_restricted_name_char(c) => (),
Some((i, ';')) if i > start => match FromStr::from_str(substr(start,i)) {
Ok(v) => {
value = v;
start = i + 1;
break;
},
Err(_) => return None
},
None => match FromStr::from_str(endstr(start)) {
Ok(v) => {
value = v;
start = raw.len();
break;
},
Err(_) => return None
},
_ => return None
}
}
}
Some(((attr, value), start))
}
// From [RFC6838](http://tools.ietf.org/html/rfc6838#section-4.2):
//
// > All registered media types MUST be assigned top-level type and
// > subtype names. The combination of these names serves to uniquely
// > identify the media type, and the subtype name facet (or the absence
// > of one) identifies the registration tree. Both top-level type and
// > subtype names are case-insensitive.
// >
// > Type and subtype names MUST conform to the following ABNF:
// >
// > type-name = restricted-name
// > subtype-name = restricted-name
// >
// > restricted-name = restricted-name-first *126restricted-name-chars
// > restricted-name-first = ALPHA / DIGIT
// > restricted-name-chars = ALPHA / DIGIT / "!" / "#" /
// > "$" / "&" / "-" / "^" / "_"
// > restricted-name-chars =/ "." ; Characters before first dot always
// > ; specify a facet name
// > restricted-name-chars =/ "+" ; Characters after last plus always
// > ; specify a structured syntax suffix
//
fn is_restricted_name_first_char(c: char) -> bool {
match c {
'a'...'z' |
'0'...'9' => true,
_ => false
}
}
fn is_restricted_name_char(c: char) -> bool {
if is_restricted_name_first_char(c) {
true
} else {
match c {
'!' |
'#' |
'$' |
'&' |
'-' |
'^' |
'.' |
'+' |
'_' => true,
_ => false
}
}
}
#[cfg(test)]
mod tests {
use std::str::FromStr;
#[cfg(feature = "nightly")]
use test::Bencher;
use super::{Mime, Value, Attr};
#[test]
fn test_mime_show() {
let mime = mime!(Text/Plain);
assert_eq!(mime.to_string(), "text/plain".to_string());
let mime = mime!(Text/Plain; Charset=Utf8);
assert_eq!(mime.to_string(), "text/plain; charset=utf-8".to_string());
}
#[test]
fn test_mime_from_str() {
assert_eq!(Mime::from_str("text/plain").unwrap(), mime!(Text/Plain));
assert_eq!(Mime::from_str("TEXT/PLAIN").unwrap(), mime!(Text/Plain));
assert_eq!(Mime::from_str("text/plain; charset=utf-8").unwrap(), mime!(Text/Plain; Charset=Utf8));
assert_eq!(Mime::from_str("text/plain;charset=\"utf-8\"").unwrap(), mime!(Text/Plain; Charset=Utf8));
assert_eq!(Mime::from_str("text/plain; charset=utf-8; foo=bar").unwrap(),
mime!(Text/Plain; Charset=Utf8, ("foo")=("bar")));
assert_eq!("*/*".parse::<Mime>().unwrap(), mime!(Star/Star));
assert_eq!("image/*".parse::<Mime>().unwrap(), mime!(Image/Star));
assert_eq!("text/*; charset=utf-8".parse::<Mime>().unwrap(), mime!(Text/Star; Charset=Utf8));
assert!("*/png".parse::<Mime>().is_err());
assert!("*image/png".parse::<Mime>().is_err());
assert!("text/*plain".parse::<Mime>().is_err());
}
| #[test]
fn test_case_sensitive_values() {
assert_eq!(Mime::from_str("multipart/form-data; boundary=ABCDEFG").unwrap(), | random_line_split |
|
lib.rs | {
($enoom:ident::_) => (
$crate::$enoom::Star
);
($enoom:ident::($inner:expr)) => (
$crate::$enoom::Ext($inner.to_string())
);
($enoom:ident::$var:ident) => (
$crate::$enoom::$var
)
}
macro_rules! enoom {
(pub enum $en:ident; $ext:ident; $($ty:ident, $text:expr;)*) => (
#[derive(Clone, Debug, Eq, Hash, Ord, PartialOrd)]
pub enum $en {
$($ty),*,
$ext(String)
}
impl $en {
pub fn as_str(&self) -> &str {
match *self {
$($en::$ty => $text),*,
$en::$ext(ref s) => &s
}
}
}
impl ::std::ops::Deref for $en {
type Target = str;
fn deref(&self) -> &str {
self.as_str()
}
}
impl PartialEq for $en {
#[inline]
fn eq(&self, other: &$en) -> bool {
match (self, other) {
$( (&$en::$ty, &$en::$ty) => true ),*,
(&$en::$ext(ref a), &$en::$ext(ref b)) => a == b,
(_, _) => self.as_str() == other.as_str(),
}
}
}
impl PartialEq<String> for $en {
fn eq(&self, other: &String) -> bool {
self.as_str() == other
}
}
impl PartialEq<str> for $en {
fn eq(&self, other: &str) -> bool {
self.as_str() == other
}
}
impl<'a> PartialEq<&'a str> for $en {
fn eq(&self, other: &&'a str) -> bool {
self.as_str() == *other
}
}
impl PartialEq<$en> for String {
fn eq(&self, other: &$en) -> bool {
self == other.as_str()
}
}
impl PartialEq<$en> for str {
fn eq(&self, other: &$en) -> bool {
self == other.as_str()
}
}
impl<'a> PartialEq<$en> for &'a str {
fn eq(&self, other: &$en) -> bool {
*self == other.as_str()
}
}
impl fmt::Display for $en {
#[inline]
fn fmt(&self, fmt: &mut fmt::Formatter) -> fmt::Result {
fmt.write_str(match *self {
$($en::$ty => $text),*,
$en::$ext(ref s) => s
})
}
}
impl FromStr for $en {
type Err = ();
fn from_str(s: &str) -> Result<$en, ()> {
Ok(match s {
$(_s if _s == $text => $en::$ty),*,
s => $en::$ext(s.to_string())
})
}
}
#[cfg(feature = "heapsize")]
impl heapsize::HeapSizeOf for $en {
fn heap_size_of_children(&self) -> usize {
match *self {
$en::$ext(ref ext) => ext.heap_size_of_children(),
_ => 0,
}
}
}
)
}
enoom! {
pub enum TopLevel;
Ext;
Star, "*";
Text, "text";
Image, "image";
Audio, "audio";
Video, "video";
Application, "application";
Multipart, "multipart";
Message, "message";
Model, "model";
}
enoom! {
pub enum SubLevel;
Ext;
Star, "*";
// common text/*
Plain, "plain";
Html, "html";
Xml, "xml";
Javascript, "javascript";
Css, "css";
EventStream, "event-stream";
// common application/*
Json, "json";
WwwFormUrlEncoded, "x-www-form-urlencoded";
Msgpack, "msgpack";
OctetStream, "octet-stream";
// multipart/*
FormData, "form-data";
// common image/*
Png, "png";
Gif, "gif";
Bmp, "bmp";
Jpeg, "jpeg";
// audio/*
Mpeg, "mpeg";
Mp4, "mp4";
Ogg, "ogg";
}
enoom! {
pub enum Attr;
Ext;
Charset, "charset";
Boundary, "boundary";
Q, "q";
}
enoom! {
pub enum Value;
Ext;
Utf8, "utf-8";
}
pub type Param = (Attr, Value);
impl<T: AsRef<[Param]>> fmt::Display for Mime<T> {
#[inline]
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
// It's much faster to write a single string, as opposed to push
// several parts through f.write_str(). So, check for the most common
// mime types, and fast track them.
if let TopLevel::Text = self.0 {
if let SubLevel::Plain = self.1 {
let attrs = self.2.as_ref();
if attrs.len() == 0 {
return f.write_str("text/plain");
} else if &[(Attr::Charset, Value::Utf8)] == attrs {
return f.write_str("text/plain; charset=utf-8");
}
}
} else if let TopLevel::Application = self.0 | else if let TopLevel::Star = self.0 {
if let SubLevel::Star = self.1 {
let attrs = self.2.as_ref();
if attrs.len() == 0 {
return f.write_str("*/*");
}
}
}
// slower general purpose fmt
try!(fmt::Display::fmt(&self.0, f));
try!(f.write_str("/"));
try!(fmt::Display::fmt(&self.1, f));
for param in self.2.as_ref() {
try!(f.write_str("; "));
try!(fmt::Display::fmt(¶m.0, f));
try!(f.write_str("="));
try!(fmt::Display::fmt(¶m.1, f));
}
Ok(())
}
}
impl<P: AsRef<[Param]>> Mime<P> {
pub fn get_param<A: PartialEq<Attr>>(&self, attr: A) -> Option<&Value> {
self.2.as_ref().iter().find(|&&(ref name, _)| attr == *name).map(|&(_, ref value)| value)
}
}
impl FromStr for Mime {
type Err = ();
fn from_str(raw: &str) -> Result<Mime, ()> {
if raw == "*/*" {
return Ok(mime!(Star/Star));
}
let ascii = raw.to_ascii_lowercase(); // lifetimes :(
let len = ascii.len();
let mut iter = ascii.chars().enumerate();
let mut params = vec![];
// toplevel
let mut start;
let top;
loop {
match iter.next() {
Some((0, c)) if is_restricted_name_first_char(c) => (),
Some((i, c)) if i > 0 && is_restricted_name_char(c) => (),
Some((i, '/')) if i > 0 => match FromStr::from_str(&ascii[..i]) {
Ok(t) => {
top = t;
start = i + 1;
break;
}
Err(_) => return Err(())
},
_ => return Err(()) // EOF and no toplevel is no Mime
};
}
// sublevel
let sub;
let mut sub_star = false;
loop {
match iter.next() {
Some((i, '*')) if i == start => {
sub_star = true;
},
Some((i, c)) if i == start && is_restricted_name_first_char(c) => (),
Some((i, c)) if !sub_star && i > start && is_restricted_name_char(c) => (),
Some((i, ';')) if i > start => match FromStr::from_str(&ascii[start..i]) {
Ok(s) => {
sub = s;
start = i + 1;
break;
}
Err(_) => return Err(())
},
None => match FromStr::from_str(&ascii[start..]) {
Ok(s) => return Ok(Mime(top, s, params)),
Err(_) => return Err(())
},
_ => return Err(())
};
}
// params
debug!("starting params, len={}", len);
loop {
match param_from_str(raw, &ascii, &mut iter, start) {
Some((p, end)) => {
params.push(p);
start = end;
if start >= len {
break;
}
}
None => break
}
}
Ok(Mime(top, sub, params))
}
}
#[cfg(feature | {
if let SubLevel::Json = self.1 {
let attrs = self.2.as_ref();
if attrs.len() == 0 {
return f.write_str("application/json");
}
}
} | conditional_block |
lib.rs | 8";
}
pub type Param = (Attr, Value);
impl<T: AsRef<[Param]>> fmt::Display for Mime<T> {
#[inline]
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
// It's much faster to write a single string, as opposed to push
// several parts through f.write_str(). So, check for the most common
// mime types, and fast track them.
if let TopLevel::Text = self.0 {
if let SubLevel::Plain = self.1 {
let attrs = self.2.as_ref();
if attrs.len() == 0 {
return f.write_str("text/plain");
} else if &[(Attr::Charset, Value::Utf8)] == attrs {
return f.write_str("text/plain; charset=utf-8");
}
}
} else if let TopLevel::Application = self.0 {
if let SubLevel::Json = self.1 {
let attrs = self.2.as_ref();
if attrs.len() == 0 {
return f.write_str("application/json");
}
}
} else if let TopLevel::Star = self.0 {
if let SubLevel::Star = self.1 {
let attrs = self.2.as_ref();
if attrs.len() == 0 {
return f.write_str("*/*");
}
}
}
// slower general purpose fmt
try!(fmt::Display::fmt(&self.0, f));
try!(f.write_str("/"));
try!(fmt::Display::fmt(&self.1, f));
for param in self.2.as_ref() {
try!(f.write_str("; "));
try!(fmt::Display::fmt(¶m.0, f));
try!(f.write_str("="));
try!(fmt::Display::fmt(¶m.1, f));
}
Ok(())
}
}
impl<P: AsRef<[Param]>> Mime<P> {
pub fn get_param<A: PartialEq<Attr>>(&self, attr: A) -> Option<&Value> {
self.2.as_ref().iter().find(|&&(ref name, _)| attr == *name).map(|&(_, ref value)| value)
}
}
impl FromStr for Mime {
type Err = ();
fn from_str(raw: &str) -> Result<Mime, ()> {
if raw == "*/*" {
return Ok(mime!(Star/Star));
}
let ascii = raw.to_ascii_lowercase(); // lifetimes :(
let len = ascii.len();
let mut iter = ascii.chars().enumerate();
let mut params = vec![];
// toplevel
let mut start;
let top;
loop {
match iter.next() {
Some((0, c)) if is_restricted_name_first_char(c) => (),
Some((i, c)) if i > 0 && is_restricted_name_char(c) => (),
Some((i, '/')) if i > 0 => match FromStr::from_str(&ascii[..i]) {
Ok(t) => {
top = t;
start = i + 1;
break;
}
Err(_) => return Err(())
},
_ => return Err(()) // EOF and no toplevel is no Mime
};
}
// sublevel
let sub;
let mut sub_star = false;
loop {
match iter.next() {
Some((i, '*')) if i == start => {
sub_star = true;
},
Some((i, c)) if i == start && is_restricted_name_first_char(c) => (),
Some((i, c)) if !sub_star && i > start && is_restricted_name_char(c) => (),
Some((i, ';')) if i > start => match FromStr::from_str(&ascii[start..i]) {
Ok(s) => {
sub = s;
start = i + 1;
break;
}
Err(_) => return Err(())
},
None => match FromStr::from_str(&ascii[start..]) {
Ok(s) => return Ok(Mime(top, s, params)),
Err(_) => return Err(())
},
_ => return Err(())
};
}
// params
debug!("starting params, len={}", len);
loop {
match param_from_str(raw, &ascii, &mut iter, start) {
Some((p, end)) => {
params.push(p);
start = end;
if start >= len {
break;
}
}
None => break
}
}
Ok(Mime(top, sub, params))
}
}
#[cfg(feature = "serde")]
impl serde::ser::Serialize for Mime {
fn serialize<S>(&self, serializer: &mut S) -> Result<(), S::Error>
where S: serde::ser::Serializer
{
serializer.serialize_str(&*format!("{}",self))
}
}
#[cfg(feature = "serde")]
impl serde::de::Deserialize for Mime {
fn deserialize<D>(deserializer: &mut D) -> Result<Self, D::Error>
where D: serde::de::Deserializer
{
let string: String = try!(serde::Deserialize::deserialize(deserializer));
let mime: Mime = match FromStr::from_str(&*string) {
Ok(mime) => mime,
Err(_) => return Err(serde::de::Error::custom("Invalid serialized mime")),
};
Ok(mime)
}
}
fn param_from_str(raw: &str, ascii: &str, iter: &mut Enumerate<Chars>, mut start: usize) -> Option<(Param, usize)> {
let attr;
debug!("param_from_str, start={}", start);
loop {
match iter.next() {
Some((i, ' ')) if i == start => start = i + 1,
Some((i, c)) if i == start && is_restricted_name_first_char(c) => (),
Some((i, c)) if i > start && is_restricted_name_char(c) => (),
Some((i, '=')) if i > start => match FromStr::from_str(&ascii[start..i]) {
Ok(a) => {
attr = a;
start = i + 1;
break;
},
Err(_) => return None
},
_ => return None
}
}
let value;
// values must be restrict-name-char or "anything goes"
let mut is_quoted = false;
{
let substr = |a,b| { if attr==Attr::Charset { &ascii[a..b] } else { &raw[a..b] } };
let endstr = |a| { if attr==Attr::Charset { &ascii[a..] } else { &raw[a..] } };
loop {
match iter.next() {
Some((i, '"')) if i == start => {
debug!("quoted");
is_quoted = true;
start = i + 1;
},
Some((i, c)) if i == start && is_restricted_name_first_char(c) => (),
Some((i, '"')) if i > start && is_quoted => match FromStr::from_str(substr(start,i)) {
Ok(v) => {
value = v;
start = i + 1;
break;
},
Err(_) => return None
},
Some((i, c)) if i > start && is_quoted || is_restricted_name_char(c) => (),
Some((i, ';')) if i > start => match FromStr::from_str(substr(start,i)) {
Ok(v) => {
value = v;
start = i + 1;
break;
},
Err(_) => return None
},
None => match FromStr::from_str(endstr(start)) {
Ok(v) => {
value = v;
start = raw.len();
break;
},
Err(_) => return None
},
_ => return None
}
}
}
Some(((attr, value), start))
}
// From [RFC6838](http://tools.ietf.org/html/rfc6838#section-4.2):
//
// > All registered media types MUST be assigned top-level type and
// > subtype names. The combination of these names serves to uniquely
// > identify the media type, and the subtype name facet (or the absence
// > of one) identifies the registration tree. Both top-level type and
// > subtype names are case-insensitive.
// >
// > Type and subtype names MUST conform to the following ABNF:
// >
// > type-name = restricted-name
// > subtype-name = restricted-name
// >
// > restricted-name = restricted-name-first *126restricted-name-chars
// > restricted-name-first = ALPHA / DIGIT
// > restricted-name-chars = ALPHA / DIGIT / "!" / "#" /
// > "$" / "&" / "-" / "^" / "_"
// > restricted-name-chars =/ "." ; Characters before first dot always
// > ; specify a facet name
// > restricted-name-chars =/ "+" ; Characters after last plus always
// > ; specify a structured syntax suffix
//
fn is_restricted_name_first_char(c: char) -> bool {
match c {
'a'...'z' |
'0'...'9' => true,
_ => false
}
}
fn | is_restricted_name_char | identifier_name |
|
lib.rs | {
($enoom:ident::_) => (
$crate::$enoom::Star
);
($enoom:ident::($inner:expr)) => (
$crate::$enoom::Ext($inner.to_string())
);
($enoom:ident::$var:ident) => (
$crate::$enoom::$var
)
}
macro_rules! enoom {
(pub enum $en:ident; $ext:ident; $($ty:ident, $text:expr;)*) => (
#[derive(Clone, Debug, Eq, Hash, Ord, PartialOrd)]
pub enum $en {
$($ty),*,
$ext(String)
}
impl $en {
pub fn as_str(&self) -> &str {
match *self {
$($en::$ty => $text),*,
$en::$ext(ref s) => &s
}
}
}
impl ::std::ops::Deref for $en {
type Target = str;
fn deref(&self) -> &str {
self.as_str()
}
}
impl PartialEq for $en {
#[inline]
fn eq(&self, other: &$en) -> bool {
match (self, other) {
$( (&$en::$ty, &$en::$ty) => true ),*,
(&$en::$ext(ref a), &$en::$ext(ref b)) => a == b,
(_, _) => self.as_str() == other.as_str(),
}
}
}
impl PartialEq<String> for $en {
fn eq(&self, other: &String) -> bool {
self.as_str() == other
}
}
impl PartialEq<str> for $en {
fn eq(&self, other: &str) -> bool {
self.as_str() == other
}
}
impl<'a> PartialEq<&'a str> for $en {
fn eq(&self, other: &&'a str) -> bool {
self.as_str() == *other
}
}
impl PartialEq<$en> for String {
fn eq(&self, other: &$en) -> bool {
self == other.as_str()
}
}
impl PartialEq<$en> for str {
fn eq(&self, other: &$en) -> bool {
self == other.as_str()
}
}
impl<'a> PartialEq<$en> for &'a str {
fn eq(&self, other: &$en) -> bool {
*self == other.as_str()
}
}
impl fmt::Display for $en {
#[inline]
fn fmt(&self, fmt: &mut fmt::Formatter) -> fmt::Result {
fmt.write_str(match *self {
$($en::$ty => $text),*,
$en::$ext(ref s) => s
})
}
}
impl FromStr for $en {
type Err = ();
fn from_str(s: &str) -> Result<$en, ()> {
Ok(match s {
$(_s if _s == $text => $en::$ty),*,
s => $en::$ext(s.to_string())
})
}
}
#[cfg(feature = "heapsize")]
impl heapsize::HeapSizeOf for $en {
fn heap_size_of_children(&self) -> usize {
match *self {
$en::$ext(ref ext) => ext.heap_size_of_children(),
_ => 0,
}
}
}
)
}
enoom! {
pub enum TopLevel;
Ext;
Star, "*";
Text, "text";
Image, "image";
Audio, "audio";
Video, "video";
Application, "application";
Multipart, "multipart";
Message, "message";
Model, "model";
}
enoom! {
pub enum SubLevel;
Ext;
Star, "*";
// common text/*
Plain, "plain";
Html, "html";
Xml, "xml";
Javascript, "javascript";
Css, "css";
EventStream, "event-stream";
// common application/*
Json, "json";
WwwFormUrlEncoded, "x-www-form-urlencoded";
Msgpack, "msgpack";
OctetStream, "octet-stream";
// multipart/*
FormData, "form-data";
// common image/*
Png, "png";
Gif, "gif";
Bmp, "bmp";
Jpeg, "jpeg";
// audio/*
Mpeg, "mpeg";
Mp4, "mp4";
Ogg, "ogg";
}
enoom! {
pub enum Attr;
Ext;
Charset, "charset";
Boundary, "boundary";
Q, "q";
}
enoom! {
pub enum Value;
Ext;
Utf8, "utf-8";
}
pub type Param = (Attr, Value);
impl<T: AsRef<[Param]>> fmt::Display for Mime<T> {
#[inline]
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result | } else if let TopLevel::Star = self.0 {
if let SubLevel::Star = self.1 {
let attrs = self.2.as_ref();
if attrs.len() == 0 {
return f.write_str("*/*");
}
}
}
// slower general purpose fmt
try!(fmt::Display::fmt(&self.0, f));
try!(f.write_str("/"));
try!(fmt::Display::fmt(&self.1, f));
for param in self.2.as_ref() {
try!(f.write_str("; "));
try!(fmt::Display::fmt(¶m.0, f));
try!(f.write_str("="));
try!(fmt::Display::fmt(¶m.1, f));
}
Ok(())
}
}
impl<P: AsRef<[Param]>> Mime<P> {
pub fn get_param<A: PartialEq<Attr>>(&self, attr: A) -> Option<&Value> {
self.2.as_ref().iter().find(|&&(ref name, _)| attr == *name).map(|&(_, ref value)| value)
}
}
impl FromStr for Mime {
type Err = ();
fn from_str(raw: &str) -> Result<Mime, ()> {
if raw == "*/*" {
return Ok(mime!(Star/Star));
}
let ascii = raw.to_ascii_lowercase(); // lifetimes :(
let len = ascii.len();
let mut iter = ascii.chars().enumerate();
let mut params = vec![];
// toplevel
let mut start;
let top;
loop {
match iter.next() {
Some((0, c)) if is_restricted_name_first_char(c) => (),
Some((i, c)) if i > 0 && is_restricted_name_char(c) => (),
Some((i, '/')) if i > 0 => match FromStr::from_str(&ascii[..i]) {
Ok(t) => {
top = t;
start = i + 1;
break;
}
Err(_) => return Err(())
},
_ => return Err(()) // EOF and no toplevel is no Mime
};
}
// sublevel
let sub;
let mut sub_star = false;
loop {
match iter.next() {
Some((i, '*')) if i == start => {
sub_star = true;
},
Some((i, c)) if i == start && is_restricted_name_first_char(c) => (),
Some((i, c)) if !sub_star && i > start && is_restricted_name_char(c) => (),
Some((i, ';')) if i > start => match FromStr::from_str(&ascii[start..i]) {
Ok(s) => {
sub = s;
start = i + 1;
break;
}
Err(_) => return Err(())
},
None => match FromStr::from_str(&ascii[start..]) {
Ok(s) => return Ok(Mime(top, s, params)),
Err(_) => return Err(())
},
_ => return Err(())
};
}
// params
debug!("starting params, len={}", len);
loop {
match param_from_str(raw, &ascii, &mut iter, start) {
Some((p, end)) => {
params.push(p);
start = end;
if start >= len {
break;
}
}
None => break
}
}
Ok(Mime(top, sub, params))
}
}
#[cfg(feature | {
// It's much faster to write a single string, as opposed to push
// several parts through f.write_str(). So, check for the most common
// mime types, and fast track them.
if let TopLevel::Text = self.0 {
if let SubLevel::Plain = self.1 {
let attrs = self.2.as_ref();
if attrs.len() == 0 {
return f.write_str("text/plain");
} else if &[(Attr::Charset, Value::Utf8)] == attrs {
return f.write_str("text/plain; charset=utf-8");
}
}
} else if let TopLevel::Application = self.0 {
if let SubLevel::Json = self.1 {
let attrs = self.2.as_ref();
if attrs.len() == 0 {
return f.write_str("application/json");
}
} | identifier_body |
utils.py | the features to unit variance
if data_opts['scale_features'][m]:
print("Scaling features for view " + str(m) + " to unit variance...")
Y[m] = Y[m] / np.std(Y[m], axis=0, )
print("\nAfter data processing:")
for m in range(M): print("view %d has %d samples and %d features..." % (m, Y[m].shape[0], Y[m].shape[1]))
return Y
def dotd(A, B, out=None):
"""Diagonal of :math:`\mathrm A\mathrm B^\intercal`.
If ``A`` is :math:`n\times p` and ``B`` is :math:`p\times n`, it is done in :math:`O(pn)`.
Args:
A (array_like): Left matrix.
B (array_like): Right matrix.
out (:class:`numpy.ndarray`, optional): copy result to.
Returns:
:class:`numpy.ndarray`: Resulting diagonal.
"""
A = ma.asarray(A, float)
B = ma.asarray(B, float)
if A.ndim == 1 and B.ndim == 1:
if out is None:
return ma.dot(A, B)
return ma.dot(A, B, out)
if out is None:
out = ma.empty((A.shape[0], ), float)
out[:] = ma.sum(A * B.T, axis=1)
return out
def nans(shape, dtype=float):
""" Method to create an array filled with missing values """
a = np.empty(shape, dtype)
a.fill(np.nan)
return a
def corr(A,B):
""" Method to efficiently compute correlation coefficients between two matrices
PARMETERS
---------
A: np array
B: np array
"""
# Rowwise mean of input arrays & subtract from input arrays themeselves
A_mA = A - A.mean(1)[:,None]
B_mB = B - B.mean(1)[:,None]
# Sum of squares across rows
ssA = (A_mA**2).sum(1);
ssB = (B_mB**2).sum(1);
# Finally get corr coeff
return np.dot(A_mA,B_mB.T)/np.sqrt(np.dot(ssA[:,None],ssB[None]))
# NOT HERE
def logdet(X):
return np.log(np.linalg.det(X))
# UC = np.linalg.cholesky(X)
# return 2*sum(np.log(np.diag(UC)))
# NOT HERE
def ddot(d, mtx, left=True):
"""Multiply a full matrix by a diagonal matrix.
This function should always be faster than dot.
Input:
d -- 1D (N,) array (contains the diagonal elements)
mtx -- 2D (N,N) array
left: is the diagonal matrix on the left or on the right of the product?
Output:
ddot(d, mts, left=True) == dot(diag(d), mtx)
ddot(d, mts, left=False) == dot(mtx, diag(d))
"""
if left:
return (d*mtx.T).T
else:
return d*mtx
def saveParameters(model, hdf5, view_names=None):
""" Method to save the parameters of the model in an hdf5 file
PARAMETERS
----------
model: a BayesNet instance
hdf5:
view_names
"""
# Get nodes from the model
nodes = model.getNodes()
# Create groups
param_grp = hdf5.create_group("parameters")
# Iterate over nodes
for node in nodes:
# Collect node parameters
parameters = nodes[node].getParameters()
# Create node subgroup
node_subgrp = param_grp.create_group(node)
# Multi-view nodes
if type(parameters) == list:
# Loop through the views
for m in range(len(parameters)):
if view_names is not None:
tmp = view_names[m]
else:
tmp = "%d" % m
# Create subsubgroup for the view
view_subgrp = node_subgrp.create_group(tmp)
# Loop through the parameters of the view
if parameters[m] is not None:
# Variational nodes
if type(parameters[m]) == dict:
for param_name in parameters[m].keys():
if parameters[m][param_name] is not None:
view_subgrp.create_dataset(param_name, data=parameters[m][param_name].T)
# Non-variational nodes (no distributions)
elif type(parameters[m]) == np.ndarray:
view_subgrp.create_dataset("value", data=parameters[m].T)
# Single-view nodes
else:
for param_name in parameters.keys():
node_subgrp.create_dataset("%s" % (param_name), data=parameters[param_name].T)
pass
def saveExpectations(model, hdf5, view_names=None):
""" Method to save the expectations of the model in an hdf5 file
PARAMETERS
----------
model: a BayesNet instance
hdf5:
view_names:
"""
# Get nodes from the model
nodes = model.getNodes()
exp_grp = hdf5.create_group("expectations")
# Iterate over nodes
for node in nodes:
# Collect node expectations
expectations = nodes[node].getExpectations()
# Multi-view nodes
if type(expectations) == list:
# Create subgroup for the node
node_subgrp = exp_grp.create_group(node)
# Iterate over views
for m in range(len(expectations)):
if view_names is not None:
view = view_names[m]
else:
view = "%d" % m
# Collect expectations
exp = expectations[m]["E"]
if exp is not None:
if type(exp) == ma.core.MaskedArray:
tmp = ma.filled(exp, fill_value=np.nan)
node_subgrp.create_dataset(view, data=tmp.T)
else:
node_subgrp.create_dataset(view, data=exp.T)
# Single-view nodes
else:
exp_grp.create_dataset(node, data=expectations["E"].T)
def saveTrainingStats(model, hdf5):
""" Method to save the training statistics in an hdf5 file
PARAMETERS
----------
model: a BayesNet instance
hdf5:
"""
stats = model.getTrainingStats()
stats_grp = hdf5.create_group("training_stats")
stats_grp.create_dataset("activeK", data=stats["activeK"])
stats_grp.create_dataset("elbo", data=stats["elbo"])
stats_grp.create_dataset("elbo_terms", data=stats["elbo_terms"].T)
stats_grp['elbo_terms'].attrs['colnames'] = [a.encode('utf8') for a in stats["elbo_terms"].columns.values]
def saveTrainingOpts(opts, hdf5):
""" Method to save the training options in an hdf5 file
PARAMETERS
----------
opts:
hdf5:
"""
# Remove dictionaries from the options
for k,v in opts.copy().items():
if type(v)==dict:
for k1,v1 in v.items():
opts[str(k)+"_"+str(k1)] = v1
opts.pop(k)
# Create HDF5 data set
hdf5.create_dataset("training_opts", data=np.array(list(opts.values()), dtype=np.float))
hdf5['training_opts'].attrs['names'] = np.asarray(list(opts.keys())).astype('S')
def saveModelOpts(opts, hdf5):
""" Method to save the model options in an hdf5 file
PARAMETERS
----------
opts:
hdf5:
"""
opts_interest = ["learnIntercept","schedule","likelihood","sparsity"]
opts = dict((k, opts[k]) for k in opts_interest)
grp = hdf5.create_group('model_opts')
for k,v in opts.items():
grp.create_dataset(k, data=np.asarray(v).astype('S'))
grp[k].attrs['names'] = np.asarray(list(opts.keys())).astype('S')
def saveTrainingData(model, hdf5, view_names=None, sample_names=None, feature_names=None, likelihoods=None):
""" Method to save the training data in an hdf5 file
PARAMETERS
----------
model: a BayesNet instance
hdf5:
view_names
sample_names
feature_names
"""
data = model.getTrainingData()
data_grp = hdf5.create_group("data")
featuredata_grp = hdf5.create_group("features")
hdf5.create_dataset("samples", data=np.array(sample_names, dtype='S50'))
if likelihoods is not None:
data_grp.attrs['likelihood'] = np.array(likelihoods, dtype='S50')
for m in range(len(data)):
view = view_names[m] if view_names is not None else str(m)
data_grp.create_dataset(view, data=data[m].data.T)
if feature_names is not None:
# data_grp.attrs['features'] = np.array(feature_names[m], dtype='S')
featuredata_grp.create_dataset(view, data=np.array(feature_names[m], dtype='S50'))
def saveModel(model, outfile, train_opts, model_opts, view_names=None, sample_names=None, feature_names=None):
""" Method to save the model in an hdf5 file
PARAMETERS
----------
TO-FILL.... | """
| random_line_split |
|
utils.py | idxMask = np.random.choice(N, size=Nsamples2Mask, replace = False)
# idxMask = np.arange(Nsamples2Mask)
# print idxMask
tmp = data[m].copy()
tmp.ix[idxMask,:] = pd.np.nan
data[m] = tmp
return data
# Function to load the data
def loadData(data_opts, verbose=True):
""" Method to load the data
PARAMETERS
----------
data_opts: dic
verbose: boolean
"""
print ("\n")
print ("#"*18)
print ("## Loading data ##")
print ("#"*18)
print ("\n")
sleep(1)
M = len(data_opts['input_files'])
Y = [None]*M
for m in range(M):
# Read file
file = data_opts['input_files'][m]
Y[m] = pd.read_csv(file, delimiter=data_opts["delimiter"], header=data_opts["colnames"], index_col=data_opts["rownames"]).astype(pd.np.float32)
# Y[m] = pd.read_csv(file, delimiter=data_opts["delimiter"])
print("Loaded %s with %d samples and %d features..." % (file, Y[m].shape[0], Y[m].shape[1]))
# Checking missing values on features
# print max(np.isnan(Y[m]).mean(axis=1))
# exit()
# Checking missing values on samples
# print np.isnan(Y[m]).mean(axis=1)
# exit()
# Check that the dimensions match
if len(set([Y[m].shape[0] for m in range(M)])) != 1:
if all([Y[m].shape[1] for m in range(M)]):
print("\nColumns seem to be the shared axis, transposing the data...")
for m in range(M): Y[m] = Y[m].T
else:
print("\nDimensionalities do not match, aborting. Make sure that either columns or rows are shared!")
exit()
# TO-DO: CHECK IF ANY SAMPLE HAS MISSING VALUES IN ALL VIEWS
# Sanity checks on the data
print ("\n" +"#"*46)
print("## Doing sanity checks and parsing the data ##")
print ("#"*46 + "\n")
for m in range(M):
# Removing features with complete missing values
nas = np.isnan(Y[m]).mean(axis=0)
if np.any(nas==1.):
print("Warning: %d features(s) on view %d have missing values in all samples, removing them..." % ( (nas==1.).sum(), m) )
Y[m].drop(Y[m].columns[np.where(nas==1.)], axis=1, inplace=True)
# Warning if there are features with no variance
var = Y[m].std(axis=0)
if np.any(var==0.):
print("Warning: %d features(s) on view %d have zero variance, consider removing them..." % ( (var==0.).sum(),m) )
# Y[m].drop(Y[m].columns[np.where(var==0.)], axis=1, inplace=True)
# Center the features
if data_opts['center_features'][m]:
print("Centering features for view " + str(m) + "...")
Y[m] = (Y[m] - Y[m].mean(axis=0))
# Scale the views to unit variance
if data_opts['scale_views'][m]:
print("Scaling view " + str(m) + " to unit variance...")
Y[m] = Y[m] / np.nanstd(Y[m].as_matrix())
# Scale the features to unit variance
if data_opts['scale_features'][m]:
print("Scaling features for view " + str(m) + " to unit variance...")
Y[m] = Y[m] / np.std(Y[m], axis=0, )
print("\nAfter data processing:")
for m in range(M): print("view %d has %d samples and %d features..." % (m, Y[m].shape[0], Y[m].shape[1]))
return Y
def dotd(A, B, out=None):
"""Diagonal of :math:`\mathrm A\mathrm B^\intercal`.
If ``A`` is :math:`n\times p` and ``B`` is :math:`p\times n`, it is done in :math:`O(pn)`.
Args:
A (array_like): Left matrix.
B (array_like): Right matrix.
out (:class:`numpy.ndarray`, optional): copy result to.
Returns:
:class:`numpy.ndarray`: Resulting diagonal.
"""
A = ma.asarray(A, float)
B = ma.asarray(B, float)
if A.ndim == 1 and B.ndim == 1:
if out is None:
return ma.dot(A, B)
return ma.dot(A, B, out)
if out is None:
out = ma.empty((A.shape[0], ), float)
out[:] = ma.sum(A * B.T, axis=1)
return out
def nans(shape, dtype=float):
""" Method to create an array filled with missing values """
a = np.empty(shape, dtype)
a.fill(np.nan)
return a
def | (A,B):
""" Method to efficiently compute correlation coefficients between two matrices
PARMETERS
---------
A: np array
B: np array
"""
# Rowwise mean of input arrays & subtract from input arrays themeselves
A_mA = A - A.mean(1)[:,None]
B_mB = B - B.mean(1)[:,None]
# Sum of squares across rows
ssA = (A_mA**2).sum(1);
ssB = (B_mB**2).sum(1);
# Finally get corr coeff
return np.dot(A_mA,B_mB.T)/np.sqrt(np.dot(ssA[:,None],ssB[None]))
# NOT HERE
def logdet(X):
return np.log(np.linalg.det(X))
# UC = np.linalg.cholesky(X)
# return 2*sum(np.log(np.diag(UC)))
# NOT HERE
def ddot(d, mtx, left=True):
"""Multiply a full matrix by a diagonal matrix.
This function should always be faster than dot.
Input:
d -- 1D (N,) array (contains the diagonal elements)
mtx -- 2D (N,N) array
left: is the diagonal matrix on the left or on the right of the product?
Output:
ddot(d, mts, left=True) == dot(diag(d), mtx)
ddot(d, mts, left=False) == dot(mtx, diag(d))
"""
if left:
return (d*mtx.T).T
else:
return d*mtx
def saveParameters(model, hdf5, view_names=None):
""" Method to save the parameters of the model in an hdf5 file
PARAMETERS
----------
model: a BayesNet instance
hdf5:
view_names
"""
# Get nodes from the model
nodes = model.getNodes()
# Create groups
param_grp = hdf5.create_group("parameters")
# Iterate over nodes
for node in nodes:
# Collect node parameters
parameters = nodes[node].getParameters()
# Create node subgroup
node_subgrp = param_grp.create_group(node)
# Multi-view nodes
if type(parameters) == list:
# Loop through the views
for m in range(len(parameters)):
if view_names is not None:
tmp = view_names[m]
else:
tmp = "%d" % m
# Create subsubgroup for the view
view_subgrp = node_subgrp.create_group(tmp)
# Loop through the parameters of the view
if parameters[m] is not None:
# Variational nodes
if type(parameters[m]) == dict:
for param_name in parameters[m].keys():
if parameters[m][param_name] is not None:
view_subgrp.create_dataset(param_name, data=parameters[m][param_name].T)
# Non-variational nodes (no distributions)
elif type(parameters[m]) == np.ndarray:
view_subgrp.create_dataset("value", data=parameters[m].T)
# Single-view nodes
else:
for param_name in parameters.keys():
node_subgrp.create_dataset("%s" % (param_name), data=parameters[param_name].T)
pass
def saveExpectations(model, hdf5, view_names=None):
""" Method to save the expectations of the model in an hdf5 file
PARAMETERS
----------
model: a BayesNet instance
hdf5:
view_names:
"""
# Get nodes from the model
nodes = model.getNodes()
exp_grp = hdf5.create_group("expectations")
# Iterate over nodes
for node in nodes:
# Collect node expectations
expectations = nodes[node].getExpectations()
# Multi-view nodes
if type(expectations) == list:
# Create subgroup for the node
node_subgrp = exp_grp.create_group(node)
# Iterate over views
for m in range(len(expectations)):
if view_names is not None:
view = view_names[m]
else:
view = "%d" % m
# | corr | identifier_name |
utils.py | :
idxMask = np.random.choice(N, size=Nsamples2Mask, replace = False)
# idxMask = np.arange(Nsamples2Mask)
# print idxMask
tmp = data[m].copy()
tmp.ix[idxMask,:] = pd.np.nan
data[m] = tmp
return data
# Function to load the data
def loadData(data_opts, verbose=True):
| # Read file
file = data_opts['input_files'][m]
Y[m] = pd.read_csv(file, delimiter=data_opts["delimiter"], header=data_opts["colnames"], index_col=data_opts["rownames"]).astype(pd.np.float32)
# Y[m] = pd.read_csv(file, delimiter=data_opts["delimiter"])
print("Loaded %s with %d samples and %d features..." % (file, Y[m].shape[0], Y[m].shape[1]))
# Checking missing values on features
# print max(np.isnan(Y[m]).mean(axis=1))
# exit()
# Checking missing values on samples
# print np.isnan(Y[m]).mean(axis=1)
# exit()
# Check that the dimensions match
if len(set([Y[m].shape[0] for m in range(M)])) != 1:
if all([Y[m].shape[1] for m in range(M)]):
print("\nColumns seem to be the shared axis, transposing the data...")
for m in range(M): Y[m] = Y[m].T
else:
print("\nDimensionalities do not match, aborting. Make sure that either columns or rows are shared!")
exit()
# TO-DO: CHECK IF ANY SAMPLE HAS MISSING VALUES IN ALL VIEWS
# Sanity checks on the data
print ("\n" +"#"*46)
print("## Doing sanity checks and parsing the data ##")
print ("#"*46 + "\n")
for m in range(M):
# Removing features with complete missing values
nas = np.isnan(Y[m]).mean(axis=0)
if np.any(nas==1.):
print("Warning: %d features(s) on view %d have missing values in all samples, removing them..." % ( (nas==1.).sum(), m) )
Y[m].drop(Y[m].columns[np.where(nas==1.)], axis=1, inplace=True)
# Warning if there are features with no variance
var = Y[m].std(axis=0)
if np.any(var==0.):
print("Warning: %d features(s) on view %d have zero variance, consider removing them..." % ( (var==0.).sum(),m) )
# Y[m].drop(Y[m].columns[np.where(var==0.)], axis=1, inplace=True)
# Center the features
if data_opts['center_features'][m]:
print("Centering features for view " + str(m) + "...")
Y[m] = (Y[m] - Y[m].mean(axis=0))
# Scale the views to unit variance
if data_opts['scale_views'][m]:
print("Scaling view " + str(m) + " to unit variance...")
Y[m] = Y[m] / np.nanstd(Y[m].as_matrix())
# Scale the features to unit variance
if data_opts['scale_features'][m]:
print("Scaling features for view " + str(m) + " to unit variance...")
Y[m] = Y[m] / np.std(Y[m], axis=0, )
print("\nAfter data processing:")
for m in range(M): print("view %d has %d samples and %d features..." % (m, Y[m].shape[0], Y[m].shape[1]))
return Y
def dotd(A, B, out=None):
"""Diagonal of :math:`\mathrm A\mathrm B^\intercal`.
If ``A`` is :math:`n\times p` and ``B`` is :math:`p\times n`, it is done in :math:`O(pn)`.
Args:
A (array_like): Left matrix.
B (array_like): Right matrix.
out (:class:`numpy.ndarray`, optional): copy result to.
Returns:
:class:`numpy.ndarray`: Resulting diagonal.
"""
A = ma.asarray(A, float)
B = ma.asarray(B, float)
if A.ndim == 1 and B.ndim == 1:
if out is None:
return ma.dot(A, B)
return ma.dot(A, B, out)
if out is None:
out = ma.empty((A.shape[0], ), float)
out[:] = ma.sum(A * B.T, axis=1)
return out
def nans(shape, dtype=float):
""" Method to create an array filled with missing values """
a = np.empty(shape, dtype)
a.fill(np.nan)
return a
def corr(A,B):
""" Method to efficiently compute correlation coefficients between two matrices
PARMETERS
---------
A: np array
B: np array
"""
# Rowwise mean of input arrays & subtract from input arrays themeselves
A_mA = A - A.mean(1)[:,None]
B_mB = B - B.mean(1)[:,None]
# Sum of squares across rows
ssA = (A_mA**2).sum(1);
ssB = (B_mB**2).sum(1);
# Finally get corr coeff
return np.dot(A_mA,B_mB.T)/np.sqrt(np.dot(ssA[:,None],ssB[None]))
# NOT HERE
def logdet(X):
return np.log(np.linalg.det(X))
# UC = np.linalg.cholesky(X)
# return 2*sum(np.log(np.diag(UC)))
# NOT HERE
def ddot(d, mtx, left=True):
"""Multiply a full matrix by a diagonal matrix.
This function should always be faster than dot.
Input:
d -- 1D (N,) array (contains the diagonal elements)
mtx -- 2D (N,N) array
left: is the diagonal matrix on the left or on the right of the product?
Output:
ddot(d, mts, left=True) == dot(diag(d), mtx)
ddot(d, mts, left=False) == dot(mtx, diag(d))
"""
if left:
return (d*mtx.T).T
else:
return d*mtx
def saveParameters(model, hdf5, view_names=None):
""" Method to save the parameters of the model in an hdf5 file
PARAMETERS
----------
model: a BayesNet instance
hdf5:
view_names
"""
# Get nodes from the model
nodes = model.getNodes()
# Create groups
param_grp = hdf5.create_group("parameters")
# Iterate over nodes
for node in nodes:
# Collect node parameters
parameters = nodes[node].getParameters()
# Create node subgroup
node_subgrp = param_grp.create_group(node)
# Multi-view nodes
if type(parameters) == list:
# Loop through the views
for m in range(len(parameters)):
if view_names is not None:
tmp = view_names[m]
else:
tmp = "%d" % m
# Create subsubgroup for the view
view_subgrp = node_subgrp.create_group(tmp)
# Loop through the parameters of the view
if parameters[m] is not None:
# Variational nodes
if type(parameters[m]) == dict:
for param_name in parameters[m].keys():
if parameters[m][param_name] is not None:
view_subgrp.create_dataset(param_name, data=parameters[m][param_name].T)
# Non-variational nodes (no distributions)
elif type(parameters[m]) == np.ndarray:
view_subgrp.create_dataset("value", data=parameters[m].T)
# Single-view nodes
else:
for param_name in parameters.keys():
node_subgrp.create_dataset("%s" % (param_name), data=parameters[param_name].T)
pass
def saveExpectations(model, hdf5, view_names=None):
""" Method to save the expectations of the model in an hdf5 file
PARAMETERS
----------
model: a BayesNet instance
hdf5:
view_names:
"""
# Get nodes from the model
nodes = model.getNodes()
exp_grp = hdf5.create_group("expectations")
# Iterate over nodes
for node in nodes:
# Collect node expectations
expectations = nodes[node].getExpectations()
# Multi-view nodes
if type(expectations) == list:
# Create subgroup for the node
node_subgrp = exp_grp.create_group(node)
# Iterate over views
for m in range(len(expectations)):
if view_names is not None:
view = view_names[m]
else:
view = "%d" % m
# | """ Method to load the data
PARAMETERS
----------
data_opts: dic
verbose: boolean
"""
print ("\n")
print ("#"*18)
print ("## Loading data ##")
print ("#"*18)
print ("\n")
sleep(1)
M = len(data_opts['input_files'])
Y = [None]*M
for m in range(M):
| identifier_body |
utils.py |
data_filt = [None]*M
samples_to_keep = np.setdiff1d(range(N),samples_to_remove)
for m in range(M):
data_filt[m] = data[m].iloc[samples_to_keep]
return data_filt
def maskData(data, data_opts):
""" Method to mask values of the data,
It is mainly to test missing values and to evaluate imputation
PARAMETERS
----------
data_opts: dic
"""
print("Masking data with the following options:")
print("at random:")
print(data_opts['maskAtRandom'])
print("full cases:")
print(data_opts['maskNSamples'])
for m in range(len(data)):
# Mask values at random
D = data[m].shape[1]
N = data[m].shape[0]
p2Mask = data_opts['maskAtRandom'][m]
if p2Mask != 0:
idxMask = np.zeros(N*D)
idxMask[:int(round(N*D*p2Mask))] = 1
np.random.shuffle(idxMask)
idxMask = np.reshape(idxMask, [N, D])
data[m] = data[m].mask(idxMask==1)
# Mask samples in a complete view
Nsamples2Mask = data_opts['maskNSamples'][m]
if Nsamples2Mask != 0:
idxMask = np.random.choice(N, size=Nsamples2Mask, replace = False)
# idxMask = np.arange(Nsamples2Mask)
# print idxMask
tmp = data[m].copy()
tmp.ix[idxMask,:] = pd.np.nan
data[m] = tmp
return data
# Function to load the data
def loadData(data_opts, verbose=True):
""" Method to load the data
PARAMETERS
----------
data_opts: dic
verbose: boolean
"""
print ("\n")
print ("#"*18)
print ("## Loading data ##")
print ("#"*18)
print ("\n")
sleep(1)
M = len(data_opts['input_files'])
Y = [None]*M
for m in range(M):
# Read file
file = data_opts['input_files'][m]
Y[m] = pd.read_csv(file, delimiter=data_opts["delimiter"], header=data_opts["colnames"], index_col=data_opts["rownames"]).astype(pd.np.float32)
# Y[m] = pd.read_csv(file, delimiter=data_opts["delimiter"])
print("Loaded %s with %d samples and %d features..." % (file, Y[m].shape[0], Y[m].shape[1]))
# Checking missing values on features
# print max(np.isnan(Y[m]).mean(axis=1))
# exit()
# Checking missing values on samples
# print np.isnan(Y[m]).mean(axis=1)
# exit()
# Check that the dimensions match
if len(set([Y[m].shape[0] for m in range(M)])) != 1:
if all([Y[m].shape[1] for m in range(M)]):
print("\nColumns seem to be the shared axis, transposing the data...")
for m in range(M): Y[m] = Y[m].T
else:
print("\nDimensionalities do not match, aborting. Make sure that either columns or rows are shared!")
exit()
# TO-DO: CHECK IF ANY SAMPLE HAS MISSING VALUES IN ALL VIEWS
# Sanity checks on the data
print ("\n" +"#"*46)
print("## Doing sanity checks and parsing the data ##")
print ("#"*46 + "\n")
for m in range(M):
# Removing features with complete missing values
nas = np.isnan(Y[m]).mean(axis=0)
if np.any(nas==1.):
print("Warning: %d features(s) on view %d have missing values in all samples, removing them..." % ( (nas==1.).sum(), m) )
Y[m].drop(Y[m].columns[np.where(nas==1.)], axis=1, inplace=True)
# Warning if there are features with no variance
var = Y[m].std(axis=0)
if np.any(var==0.):
print("Warning: %d features(s) on view %d have zero variance, consider removing them..." % ( (var==0.).sum(),m) )
# Y[m].drop(Y[m].columns[np.where(var==0.)], axis=1, inplace=True)
# Center the features
if data_opts['center_features'][m]:
print("Centering features for view " + str(m) + "...")
Y[m] = (Y[m] - Y[m].mean(axis=0))
# Scale the views to unit variance
if data_opts['scale_views'][m]:
print("Scaling view " + str(m) + " to unit variance...")
Y[m] = Y[m] / np.nanstd(Y[m].as_matrix())
# Scale the features to unit variance
if data_opts['scale_features'][m]:
print("Scaling features for view " + str(m) + " to unit variance...")
Y[m] = Y[m] / np.std(Y[m], axis=0, )
print("\nAfter data processing:")
for m in range(M): print("view %d has %d samples and %d features..." % (m, Y[m].shape[0], Y[m].shape[1]))
return Y
def dotd(A, B, out=None):
"""Diagonal of :math:`\mathrm A\mathrm B^\intercal`.
If ``A`` is :math:`n\times p` and ``B`` is :math:`p\times n`, it is done in :math:`O(pn)`.
Args:
A (array_like): Left matrix.
B (array_like): Right matrix.
out (:class:`numpy.ndarray`, optional): copy result to.
Returns:
:class:`numpy.ndarray`: Resulting diagonal.
"""
A = ma.asarray(A, float)
B = ma.asarray(B, float)
if A.ndim == 1 and B.ndim == 1:
if out is None:
return ma.dot(A, B)
return ma.dot(A, B, out)
if out is None:
out = ma.empty((A.shape[0], ), float)
out[:] = ma.sum(A * B.T, axis=1)
return out
def nans(shape, dtype=float):
""" Method to create an array filled with missing values """
a = np.empty(shape, dtype)
a.fill(np.nan)
return a
def corr(A,B):
""" Method to efficiently compute correlation coefficients between two matrices
PARMETERS
---------
A: np array
B: np array
"""
# Rowwise mean of input arrays & subtract from input arrays themeselves
A_mA = A - A.mean(1)[:,None]
B_mB = B - B.mean(1)[:,None]
# Sum of squares across rows
ssA = (A_mA**2).sum(1);
ssB = (B_mB**2).sum(1);
# Finally get corr coeff
return np.dot(A_mA,B_mB.T)/np.sqrt(np.dot(ssA[:,None],ssB[None]))
# NOT HERE
def logdet(X):
return np.log(np.linalg.det(X))
# UC = np.linalg.cholesky(X)
# return 2*sum(np.log(np.diag(UC)))
# NOT HERE
def ddot(d, mtx, left=True):
"""Multiply a full matrix by a diagonal matrix.
This function should always be faster than dot.
Input:
d -- 1D (N,) array (contains the diagonal elements)
mtx -- 2D (N,N) array
left: is the diagonal matrix on the left or on the right of the product?
Output:
ddot(d, mts, left=True) == dot(diag(d), mtx)
ddot(d, mts, left=False) == dot(mtx, diag(d))
"""
if left:
return (d*mtx.T).T
else:
return d*mtx
def saveParameters(model, hdf5, view_names=None):
""" Method to save the parameters of the model in an hdf5 file
PARAMETERS
----------
model: a BayesNet instance
hdf5:
view_names
"""
# Get nodes from the model
nodes = model.getNodes()
# Create groups
param_grp = hdf5.create_group("parameters")
# Iterate over nodes
for node in nodes:
# Collect node parameters
parameters = nodes[node].getParameters()
# Create node subgroup
node_subgrp = param_grp.create_group(node)
# Multi-view nodes
if type(parameters) == list:
# Loop through the views
for m in range(len(parameters)):
if view_names is not None:
tmp = view_names[m]
else:
tmp = "%d" % m
# Create subsubgroup for the view
view_subgrp = node_subgrp.create_group(tmp)
# Loop through the parameters of the view
if parameters[m] is not None:
| print("A total of " + str(len(samples_to_remove)) + " sample(s) have at least a missing view and will be removed") | conditional_block |
|
list.js | no, error) {
$('#googleMapError').text("There was an error loading resources. Please correct and try again");
$('#googleMapError').show();
}
// Thanks to the following link for help with this function.
//http://stackoverflow.com/questions/14184956/async-google-maps-api-v3-undefined-is-not-a-function
// This function is called when google maps is ready to be loaded. (this is the callback when the map is loaded asynchronously.
/*
Uses geocoder to obtain lat and lng values for the location of interest.
Passes the location info to getBookstores, which queries Foursquare.
If getBookstores succeeds then processFrsqrBooks is called with the response from FourSquare.
*/
function loadMap() {
infowindow = new google.maps.InfoWindow({});
infowindow.isOpen = false; // property of global infowindow variable that holds the open state of the infowindow.
// Request latitide and longitude of Fisherman's wharf, SanFrancisco, CA.
var geocoder = new google.maps.Geocoder();
var address = 'fisherman\'s wharf, sanfrancisco, CA, USA';
geocoder.geocode({
'address': address
}, function(results, status) {
switch (status) {
case 'OK':
if (status == google.maps.GeocoderStatus.OK) {
map = new google.maps.Map(document.getElementById('map'), {
center: sfo,
zoom: 15,
scrollwheel: false
});
var service = new google.maps.places.PlacesService(map);
var geocoderSearchResult = results[0];
map.setCenter(geocoderSearchResult.geometry.location);
getBookstores(geocoderSearchResult).then(function(response) {
processFrsqrBooks(response);
}, function(error) {
$('#googleMapError').text("Could not load FourSquare data" + error);
$('#googleMapError').show();
console.log(error);
});
}
break;
case 'ZERO_RESULTS':
$('#googleMapError').text(status);
$('#googleMapError').show();
break;
case 'OVER_QUERY_LIMIT':
$('#googleMapError').text(status);
$('#googleMapError').show();
break;
case 'REQUEST_DENIED':
$('#googleMapError').text(status);
$('#googleMapError').show();
break;
case 'INVALID_REQUEST':
$('#googleMapError').text(status);
$('#googleMapError').show();
break;
case 'UNKNOWN_ERROR':
$('#googleMapError').text(status);
$('#googleMapError').show();
break;
}
});
}
/*
This function is called when the request to Foursquare succeeds i.e. comes back with any result.
If the status code of the response is 200, then the knockout observable arrays are populated with the
result data.
*/
function processFrsqrBooks(response) {
if (getBooksRequest.readyState === XMLHttpRequest.DONE) {
if (getBooksRequest.status === 200) {
var jsonResponse = JSON.parse(getBooksRequest.responseText);
var bkstr = []; // array, holds the frsqrItem object literal that is defined inside the loop below.
var frsqrBookItems = [];
if (jsonResponse.response.groups.length > 0) {
bookVenues = jsonResponse.response.groups[0];
items = bookVenues.items;
for (var i = 0; i < items.length; i++) {
// object that holds data for individual locations from the Foursquare response.
var frsqrItem = new Venue();
// populate the object literal with data from the response.
frsqrItem.tips = items[i].tips;
frsqrItem.name = items[i].venue.name;
frsqrItem.venueUrl = items[i].venue.url;
frsqrItem.lat = items[i].venue.location.lat;
frsqrItem.lng = items[i].venue.location.lng;
frsqrItem.index = i;
// Photos for the locations,
if (items[i].venue.photos.count > 0) {
// there is at least one photo - so construct photo url.
var groups = items[i].venue.photos.groups;
// Some Foursquare 'venues' do not have photos, so check if the location has any photos
if (groups.length > 0) {
var photoItems = groups[0].items;
if (photoItems.length > 0) {
frsqrItem.venuePhotoUrl = photoItems[0].prefix + '50x50' + photoItems[0].suffix;
}
}
}
frsqrItem.rating = items[i].venue.rating;
frsqrItem.marker = createMarker(frsqrItem);
bookstoreViewModel.venues.push(frsqrItem);
}
}
} else {
alert('There was a problem with the request.');
}
}
}
// This function sets up markes for points of interest and adds click handlers to all the markers.
function createMarker(frsqrItem) {
var content = "";
// The marker object ,
// - animation property set to DROP.
// - icon property is set to an icon from Templatic
var marker = new google.maps.Marker({
map: map,
animation: google.maps.Animation.DROP,
title: frsqrItem.name,
icon: './templatic/books-media.png',
//infowindow: new google.maps.InfoWindow(),
position: {
lat: frsqrItem.lat,
lng: frsqrItem.lng
}
});
content = content + "</br>";
content = content + "<p> " + marker.title + "</p>";
content = content + "<img src=\"" + frsqrItem.venuePhotoUrl + "\"/>";
marker.content = content;
content = '';
// add click handler to every marker.
// When a marker is clicked, the name of the location and photo is displayed.
// The animation property is set to bounce, so the marker bounces when you click on it
google.maps.event.addListener(marker, 'click', function() {
var self = this;
if (self.getAnimation() !== null) {
self.setAnimation(null);
} else {
self.setAnimation(google.maps.Animation.BOUNCE, 1400);
stopAnimation(marker);
}
// TODO: Open the infowindow only if it is not already open.
if(!infowindow.isOpen){
//The infowindow is not already open.
infowindow.setContent(self.content);
infowindow.open(self.map, this);
infowindow.isOpen = true;
}
else{
infowindow.close();
infowindow.setContent(self.content);
infowindow.open(self.map, this);
infowindow.isOpen = true;
}
});
return marker;
}
var stopAnimation = function(marker) {
setTimeout(function() {
marker.setAnimation(null);
}, 1400);
};
// code attribution: https://github.com/mdn/promises-test/blob/gh-pages/index.html
/*
This function takes the result from the geocoder request and subits a request to the Foursquare API.
*/
function | (geocoderSearchResult) {
return new Promise(function(resolve, reject) {
if (geocoderSearchResult.geometry.location) {
map.setCenter(geocoderSearchResult.geometry.location);
// Create a list and display all the results.
var cll = geocoderSearchResult.geometry.location.lat() + "," + geocoderSearchResult.geometry.location.lng();
var foursquareQuery = "https://api.foursquare.com/v2/venues/explore?client_id=F0XYIB113FEQQVQFFK5DGZ4V5PJBZA2DRNAXHFUW1G3UBE3N&client_secret=ZYY5PZ15D02DLZ0D3RGBADODPBC1KMKX4ZIQ4XNDNLUKBKEB&v=20140701&ll=" + cll + "&radius=1000&query=books&venuePhotos=1&limit=50";
getBooksRequest = new XMLHttpRequest();
if (!getBooksRequest) {
alert('Giving up getBookstores:( Cannot create an XMLHTTP instance');
return false;
}
getBooksRequest.open('GET', foursquareQuery);
getBooksRequest.responseType = 'text';
getBooksRequest.onload = function() {
if (getBooksRequest.status === 200) {
resolve(getBooksRequest.response);
} else {
reject(Error('Request did not load successfully' + getBooksRequest.statusText));
}
};
getBooksRequest.onerror = function() {
reject(Error('There was a network error'));
};
getBooksRequest.send();
} // if ends
});
}
var Venue = function() {
this.tips = '';
this.name = '';
this.venueUrl = '';
this.venuePhotoUrl = '';
this.rating = 0.0;
this.lat = 0;
this.lng = 0;
this.index = '';
this.marker = {};
this.displaySelection = function() {
if(!infowindow.isOpen){
//The infowindow is not already open.
infowindow.setContent(this.marker.content);
infowindow.open(map, this.marker);
infowindow.isOpen | getBookstores | identifier_name |
list.js | no, error) {
$('#googleMapError').text("There was an error loading resources. Please correct and try again");
$('#googleMapError').show();
}
// Thanks to the following link for help with this function.
//http://stackoverflow.com/questions/14184956/async-google-maps-api-v3-undefined-is-not-a-function
// This function is called when google maps is ready to be loaded. (this is the callback when the map is loaded asynchronously.
/*
Uses geocoder to obtain lat and lng values for the location of interest.
Passes the location info to getBookstores, which queries Foursquare.
If getBookstores succeeds then processFrsqrBooks is called with the response from FourSquare.
*/
function loadMap() {
infowindow = new google.maps.InfoWindow({});
infowindow.isOpen = false; // property of global infowindow variable that holds the open state of the infowindow.
// Request latitide and longitude of Fisherman's wharf, SanFrancisco, CA.
var geocoder = new google.maps.Geocoder();
var address = 'fisherman\'s wharf, sanfrancisco, CA, USA';
geocoder.geocode({
'address': address
}, function(results, status) {
switch (status) {
case 'OK':
if (status == google.maps.GeocoderStatus.OK) {
map = new google.maps.Map(document.getElementById('map'), {
center: sfo,
zoom: 15,
scrollwheel: false
});
var service = new google.maps.places.PlacesService(map);
var geocoderSearchResult = results[0];
map.setCenter(geocoderSearchResult.geometry.location);
getBookstores(geocoderSearchResult).then(function(response) {
processFrsqrBooks(response);
}, function(error) {
$('#googleMapError').text("Could not load FourSquare data" + error);
$('#googleMapError').show();
console.log(error);
});
}
break;
case 'ZERO_RESULTS':
$('#googleMapError').text(status);
$('#googleMapError').show();
break;
case 'OVER_QUERY_LIMIT':
$('#googleMapError').text(status);
$('#googleMapError').show();
break;
case 'REQUEST_DENIED':
$('#googleMapError').text(status);
$('#googleMapError').show();
break;
case 'INVALID_REQUEST':
$('#googleMapError').text(status);
$('#googleMapError').show();
break;
case 'UNKNOWN_ERROR':
$('#googleMapError').text(status);
$('#googleMapError').show();
break;
}
});
}
/*
This function is called when the request to Foursquare succeeds i.e. comes back with any result.
If the status code of the response is 200, then the knockout observable arrays are populated with the
result data.
*/
function processFrsqrBooks(response) {
if (getBooksRequest.readyState === XMLHttpRequest.DONE) {
if (getBooksRequest.status === 200) {
var jsonResponse = JSON.parse(getBooksRequest.responseText);
var bkstr = []; // array, holds the frsqrItem object literal that is defined inside the loop below.
var frsqrBookItems = [];
if (jsonResponse.response.groups.length > 0) {
bookVenues = jsonResponse.response.groups[0];
items = bookVenues.items;
for (var i = 0; i < items.length; i++) {
// object that holds data for individual locations from the Foursquare response.
var frsqrItem = new Venue();
// populate the object literal with data from the response.
frsqrItem.tips = items[i].tips;
frsqrItem.name = items[i].venue.name;
frsqrItem.venueUrl = items[i].venue.url;
frsqrItem.lat = items[i].venue.location.lat;
frsqrItem.lng = items[i].venue.location.lng;
frsqrItem.index = i;
// Photos for the locations,
if (items[i].venue.photos.count > 0) {
// there is at least one photo - so construct photo url.
var groups = items[i].venue.photos.groups;
// Some Foursquare 'venues' do not have photos, so check if the location has any photos
if (groups.length > 0) {
var photoItems = groups[0].items;
if (photoItems.length > 0) {
frsqrItem.venuePhotoUrl = photoItems[0].prefix + '50x50' + photoItems[0].suffix;
}
}
}
frsqrItem.rating = items[i].venue.rating;
frsqrItem.marker = createMarker(frsqrItem);
bookstoreViewModel.venues.push(frsqrItem);
}
}
} else {
alert('There was a problem with the request.');
}
}
}
// This function sets up markes for points of interest and adds click handlers to all the markers.
function createMarker(frsqrItem) {
var content = "";
// The marker object ,
// - animation property set to DROP.
// - icon property is set to an icon from Templatic
var marker = new google.maps.Marker({
map: map,
animation: google.maps.Animation.DROP,
title: frsqrItem.name,
icon: './templatic/books-media.png',
//infowindow: new google.maps.InfoWindow(),
position: {
lat: frsqrItem.lat,
lng: frsqrItem.lng
}
});
content = content + "</br>";
content = content + "<p> " + marker.title + "</p>";
content = content + "<img src=\"" + frsqrItem.venuePhotoUrl + "\"/>";
marker.content = content;
content = '';
// add click handler to every marker.
// When a marker is clicked, the name of the location and photo is displayed.
// The animation property is set to bounce, so the marker bounces when you click on it
google.maps.event.addListener(marker, 'click', function() {
var self = this;
if (self.getAnimation() !== null) {
self.setAnimation(null);
} else {
self.setAnimation(google.maps.Animation.BOUNCE, 1400);
stopAnimation(marker);
}
// TODO: Open the infowindow only if it is not already open.
if(!infowindow.isOpen){
//The infowindow is not already open.
infowindow.setContent(self.content);
infowindow.open(self.map, this);
infowindow.isOpen = true;
}
else{
infowindow.close();
infowindow.setContent(self.content);
infowindow.open(self.map, this);
infowindow.isOpen = true;
}
});
return marker;
}
var stopAnimation = function(marker) {
setTimeout(function() {
marker.setAnimation(null);
}, 1400);
};
// code attribution: https://github.com/mdn/promises-test/blob/gh-pages/index.html
/*
This function takes the result from the geocoder request and subits a request to the Foursquare API.
*/
function getBookstores(geocoderSearchResult) {
return new Promise(function(resolve, reject) {
if (geocoderSearchResult.geometry.location) {
map.setCenter(geocoderSearchResult.geometry.location);
// Create a list and display all the results.
var cll = geocoderSearchResult.geometry.location.lat() + "," + geocoderSearchResult.geometry.location.lng();
var foursquareQuery = "https://api.foursquare.com/v2/venues/explore?client_id=F0XYIB113FEQQVQFFK5DGZ4V5PJBZA2DRNAXHFUW1G3UBE3N&client_secret=ZYY5PZ15D02DLZ0D3RGBADODPBC1KMKX4ZIQ4XNDNLUKBKEB&v=20140701&ll=" + cll + "&radius=1000&query=books&venuePhotos=1&limit=50";
getBooksRequest = new XMLHttpRequest();
if (!getBooksRequest) {
alert('Giving up getBookstores:( Cannot create an XMLHTTP instance');
return false;
}
getBooksRequest.open('GET', foursquareQuery);
getBooksRequest.responseType = 'text';
getBooksRequest.onload = function() {
if (getBooksRequest.status === 200) {
resolve(getBooksRequest.response);
} else {
reject(Error('Request did not load successfully' + getBooksRequest.statusText));
}
};
getBooksRequest.onerror = function() {
reject(Error('There was a network error'));
};
getBooksRequest.send();
} // if ends
});
}
|
var Venue = function() {
this.tips = '';
this.name = '';
this.venueUrl = '';
this.venuePhotoUrl = '';
this.rating = 0.0;
this.lat = 0;
this.lng = 0;
this.index = '';
this.marker = {};
this.displaySelection = function() {
if(!infowindow.isOpen){
//The infowindow is not already open.
infowindow.setContent(this.marker.content);
infowindow.open(map, this.marker);
infowindow.isOpen | random_line_split |
|
list.js | no, error) {
$('#googleMapError').text("There was an error loading resources. Please correct and try again");
$('#googleMapError').show();
}
// Thanks to the following link for help with this function.
//http://stackoverflow.com/questions/14184956/async-google-maps-api-v3-undefined-is-not-a-function
// This function is called when google maps is ready to be loaded. (this is the callback when the map is loaded asynchronously.
/*
Uses geocoder to obtain lat and lng values for the location of interest.
Passes the location info to getBookstores, which queries Foursquare.
If getBookstores succeeds then processFrsqrBooks is called with the response from FourSquare.
*/
function loadMap() {
infowindow = new google.maps.InfoWindow({});
infowindow.isOpen = false; // property of global infowindow variable that holds the open state of the infowindow.
// Request latitide and longitude of Fisherman's wharf, SanFrancisco, CA.
var geocoder = new google.maps.Geocoder();
var address = 'fisherman\'s wharf, sanfrancisco, CA, USA';
geocoder.geocode({
'address': address
}, function(results, status) {
switch (status) {
case 'OK':
if (status == google.maps.GeocoderStatus.OK) {
map = new google.maps.Map(document.getElementById('map'), {
center: sfo,
zoom: 15,
scrollwheel: false
});
var service = new google.maps.places.PlacesService(map);
var geocoderSearchResult = results[0];
map.setCenter(geocoderSearchResult.geometry.location);
getBookstores(geocoderSearchResult).then(function(response) {
processFrsqrBooks(response);
}, function(error) {
$('#googleMapError').text("Could not load FourSquare data" + error);
$('#googleMapError').show();
console.log(error);
});
}
break;
case 'ZERO_RESULTS':
$('#googleMapError').text(status);
$('#googleMapError').show();
break;
case 'OVER_QUERY_LIMIT':
$('#googleMapError').text(status);
$('#googleMapError').show();
break;
case 'REQUEST_DENIED':
$('#googleMapError').text(status);
$('#googleMapError').show();
break;
case 'INVALID_REQUEST':
$('#googleMapError').text(status);
$('#googleMapError').show();
break;
case 'UNKNOWN_ERROR':
$('#googleMapError').text(status);
$('#googleMapError').show();
break;
}
});
}
/*
This function is called when the request to Foursquare succeeds i.e. comes back with any result.
If the status code of the response is 200, then the knockout observable arrays are populated with the
result data.
*/
function processFrsqrBooks(response) {
if (getBooksRequest.readyState === XMLHttpRequest.DONE) {
if (getBooksRequest.status === 200) {
var jsonResponse = JSON.parse(getBooksRequest.responseText);
var bkstr = []; // array, holds the frsqrItem object literal that is defined inside the loop below.
var frsqrBookItems = [];
if (jsonResponse.response.groups.length > 0) {
bookVenues = jsonResponse.response.groups[0];
items = bookVenues.items;
for (var i = 0; i < items.length; i++) {
// object that holds data for individual locations from the Foursquare response.
var frsqrItem = new Venue();
// populate the object literal with data from the response.
frsqrItem.tips = items[i].tips;
frsqrItem.name = items[i].venue.name;
frsqrItem.venueUrl = items[i].venue.url;
frsqrItem.lat = items[i].venue.location.lat;
frsqrItem.lng = items[i].venue.location.lng;
frsqrItem.index = i;
// Photos for the locations,
if (items[i].venue.photos.count > 0) {
// there is at least one photo - so construct photo url.
var groups = items[i].venue.photos.groups;
// Some Foursquare 'venues' do not have photos, so check if the location has any photos
if (groups.length > 0) {
var photoItems = groups[0].items;
if (photoItems.length > 0) {
frsqrItem.venuePhotoUrl = photoItems[0].prefix + '50x50' + photoItems[0].suffix;
}
}
}
frsqrItem.rating = items[i].venue.rating;
frsqrItem.marker = createMarker(frsqrItem);
bookstoreViewModel.venues.push(frsqrItem);
}
}
} else {
alert('There was a problem with the request.');
}
}
}
// This function sets up markes for points of interest and adds click handlers to all the markers.
function createMarker(frsqrItem) {
var content = "";
// The marker object ,
// - animation property set to DROP.
// - icon property is set to an icon from Templatic
var marker = new google.maps.Marker({
map: map,
animation: google.maps.Animation.DROP,
title: frsqrItem.name,
icon: './templatic/books-media.png',
//infowindow: new google.maps.InfoWindow(),
position: {
lat: frsqrItem.lat,
lng: frsqrItem.lng
}
});
content = content + "</br>";
content = content + "<p> " + marker.title + "</p>";
content = content + "<img src=\"" + frsqrItem.venuePhotoUrl + "\"/>";
marker.content = content;
content = '';
// add click handler to every marker.
// When a marker is clicked, the name of the location and photo is displayed.
// The animation property is set to bounce, so the marker bounces when you click on it
google.maps.event.addListener(marker, 'click', function() {
var self = this;
if (self.getAnimation() !== null) {
self.setAnimation(null);
} else {
self.setAnimation(google.maps.Animation.BOUNCE, 1400);
stopAnimation(marker);
}
// TODO: Open the infowindow only if it is not already open.
if(!infowindow.isOpen){
//The infowindow is not already open.
infowindow.setContent(self.content);
infowindow.open(self.map, this);
infowindow.isOpen = true;
}
else |
});
return marker;
}
var stopAnimation = function(marker) {
setTimeout(function() {
marker.setAnimation(null);
}, 1400);
};
// code attribution: https://github.com/mdn/promises-test/blob/gh-pages/index.html
/*
This function takes the result from the geocoder request and subits a request to the Foursquare API.
*/
function getBookstores(geocoderSearchResult) {
return new Promise(function(resolve, reject) {
if (geocoderSearchResult.geometry.location) {
map.setCenter(geocoderSearchResult.geometry.location);
// Create a list and display all the results.
var cll = geocoderSearchResult.geometry.location.lat() + "," + geocoderSearchResult.geometry.location.lng();
var foursquareQuery = "https://api.foursquare.com/v2/venues/explore?client_id=F0XYIB113FEQQVQFFK5DGZ4V5PJBZA2DRNAXHFUW1G3UBE3N&client_secret=ZYY5PZ15D02DLZ0D3RGBADODPBC1KMKX4ZIQ4XNDNLUKBKEB&v=20140701&ll=" + cll + "&radius=1000&query=books&venuePhotos=1&limit=50";
getBooksRequest = new XMLHttpRequest();
if (!getBooksRequest) {
alert('Giving up getBookstores:( Cannot create an XMLHTTP instance');
return false;
}
getBooksRequest.open('GET', foursquareQuery);
getBooksRequest.responseType = 'text';
getBooksRequest.onload = function() {
if (getBooksRequest.status === 200) {
resolve(getBooksRequest.response);
} else {
reject(Error('Request did not load successfully' + getBooksRequest.statusText));
}
};
getBooksRequest.onerror = function() {
reject(Error('There was a network error'));
};
getBooksRequest.send();
} // if ends
});
}
var Venue = function() {
this.tips = '';
this.name = '';
this.venueUrl = '';
this.venuePhotoUrl = '';
this.rating = 0.0;
this.lat = 0;
this.lng = 0;
this.index = '';
this.marker = {};
this.displaySelection = function() {
if(!infowindow.isOpen){
//The infowindow is not already open.
infowindow.setContent(this.marker.content);
infowindow.open(map, this.marker);
infowindow.isOpen = | {
infowindow.close();
infowindow.setContent(self.content);
infowindow.open(self.map, this);
infowindow.isOpen = true;
} | conditional_block |
list.js | no, error) {
$('#googleMapError').text("There was an error loading resources. Please correct and try again");
$('#googleMapError').show();
}
// Thanks to the following link for help with this function.
//http://stackoverflow.com/questions/14184956/async-google-maps-api-v3-undefined-is-not-a-function
// This function is called when google maps is ready to be loaded. (this is the callback when the map is loaded asynchronously.
/*
Uses geocoder to obtain lat and lng values for the location of interest.
Passes the location info to getBookstores, which queries Foursquare.
If getBookstores succeeds then processFrsqrBooks is called with the response from FourSquare.
*/
function loadMap() {
infowindow = new google.maps.InfoWindow({});
infowindow.isOpen = false; // property of global infowindow variable that holds the open state of the infowindow.
// Request latitide and longitude of Fisherman's wharf, SanFrancisco, CA.
var geocoder = new google.maps.Geocoder();
var address = 'fisherman\'s wharf, sanfrancisco, CA, USA';
geocoder.geocode({
'address': address
}, function(results, status) {
switch (status) {
case 'OK':
if (status == google.maps.GeocoderStatus.OK) {
map = new google.maps.Map(document.getElementById('map'), {
center: sfo,
zoom: 15,
scrollwheel: false
});
var service = new google.maps.places.PlacesService(map);
var geocoderSearchResult = results[0];
map.setCenter(geocoderSearchResult.geometry.location);
getBookstores(geocoderSearchResult).then(function(response) {
processFrsqrBooks(response);
}, function(error) {
$('#googleMapError').text("Could not load FourSquare data" + error);
$('#googleMapError').show();
console.log(error);
});
}
break;
case 'ZERO_RESULTS':
$('#googleMapError').text(status);
$('#googleMapError').show();
break;
case 'OVER_QUERY_LIMIT':
$('#googleMapError').text(status);
$('#googleMapError').show();
break;
case 'REQUEST_DENIED':
$('#googleMapError').text(status);
$('#googleMapError').show();
break;
case 'INVALID_REQUEST':
$('#googleMapError').text(status);
$('#googleMapError').show();
break;
case 'UNKNOWN_ERROR':
$('#googleMapError').text(status);
$('#googleMapError').show();
break;
}
});
}
/*
This function is called when the request to Foursquare succeeds i.e. comes back with any result.
If the status code of the response is 200, then the knockout observable arrays are populated with the
result data.
*/
function processFrsqrBooks(response) | if (items[i].venue.photos.count > 0) {
// there is at least one photo - so construct photo url.
var groups = items[i].venue.photos.groups;
// Some Foursquare 'venues' do not have photos, so check if the location has any photos
if (groups.length > 0) {
var photoItems = groups[0].items;
if (photoItems.length > 0) {
frsqrItem.venuePhotoUrl = photoItems[0].prefix + '50x50' + photoItems[0].suffix;
}
}
}
frsqrItem.rating = items[i].venue.rating;
frsqrItem.marker = createMarker(frsqrItem);
bookstoreViewModel.venues.push(frsqrItem);
}
}
} else {
alert('There was a problem with the request.');
}
}
}
// This function sets up markes for points of interest and adds click handlers to all the markers.
function createMarker(frsqrItem) {
var content = "";
// The marker object ,
// - animation property set to DROP.
// - icon property is set to an icon from Templatic
var marker = new google.maps.Marker({
map: map,
animation: google.maps.Animation.DROP,
title: frsqrItem.name,
icon: './templatic/books-media.png',
//infowindow: new google.maps.InfoWindow(),
position: {
lat: frsqrItem.lat,
lng: frsqrItem.lng
}
});
content = content + "</br>";
content = content + "<p> " + marker.title + "</p>";
content = content + "<img src=\"" + frsqrItem.venuePhotoUrl + "\"/>";
marker.content = content;
content = '';
// add click handler to every marker.
// When a marker is clicked, the name of the location and photo is displayed.
// The animation property is set to bounce, so the marker bounces when you click on it
google.maps.event.addListener(marker, 'click', function() {
var self = this;
if (self.getAnimation() !== null) {
self.setAnimation(null);
} else {
self.setAnimation(google.maps.Animation.BOUNCE, 1400);
stopAnimation(marker);
}
// TODO: Open the infowindow only if it is not already open.
if(!infowindow.isOpen){
//The infowindow is not already open.
infowindow.setContent(self.content);
infowindow.open(self.map, this);
infowindow.isOpen = true;
}
else{
infowindow.close();
infowindow.setContent(self.content);
infowindow.open(self.map, this);
infowindow.isOpen = true;
}
});
return marker;
}
var stopAnimation = function(marker) {
setTimeout(function() {
marker.setAnimation(null);
}, 1400);
};
// code attribution: https://github.com/mdn/promises-test/blob/gh-pages/index.html
/*
This function takes the result from the geocoder request and subits a request to the Foursquare API.
*/
function getBookstores(geocoderSearchResult) {
return new Promise(function(resolve, reject) {
if (geocoderSearchResult.geometry.location) {
map.setCenter(geocoderSearchResult.geometry.location);
// Create a list and display all the results.
var cll = geocoderSearchResult.geometry.location.lat() + "," + geocoderSearchResult.geometry.location.lng();
var foursquareQuery = "https://api.foursquare.com/v2/venues/explore?client_id=F0XYIB113FEQQVQFFK5DGZ4V5PJBZA2DRNAXHFUW1G3UBE3N&client_secret=ZYY5PZ15D02DLZ0D3RGBADODPBC1KMKX4ZIQ4XNDNLUKBKEB&v=20140701&ll=" + cll + "&radius=1000&query=books&venuePhotos=1&limit=50";
getBooksRequest = new XMLHttpRequest();
if (!getBooksRequest) {
alert('Giving up getBookstores:( Cannot create an XMLHTTP instance');
return false;
}
getBooksRequest.open('GET', foursquareQuery);
getBooksRequest.responseType = 'text';
getBooksRequest.onload = function() {
if (getBooksRequest.status === 200) {
resolve(getBooksRequest.response);
} else {
reject(Error('Request did not load successfully' + getBooksRequest.statusText));
}
};
getBooksRequest.onerror = function() {
reject(Error('There was a network error'));
};
getBooksRequest.send();
} // if ends
});
}
var Venue = function() {
this.tips = '';
this.name = '';
this.venueUrl = '';
this.venuePhotoUrl = '';
this.rating = 0.0;
this.lat = 0;
this.lng = 0;
this.index = '';
this.marker = {};
this.displaySelection = function() {
if(!infowindow.isOpen){
//The infowindow is not already open.
infowindow.setContent(this.marker.content);
infowindow.open(map, this.marker);
infowindow.isOpen | {
if (getBooksRequest.readyState === XMLHttpRequest.DONE) {
if (getBooksRequest.status === 200) {
var jsonResponse = JSON.parse(getBooksRequest.responseText);
var bkstr = []; // array, holds the frsqrItem object literal that is defined inside the loop below.
var frsqrBookItems = [];
if (jsonResponse.response.groups.length > 0) {
bookVenues = jsonResponse.response.groups[0];
items = bookVenues.items;
for (var i = 0; i < items.length; i++) {
// object that holds data for individual locations from the Foursquare response.
var frsqrItem = new Venue();
// populate the object literal with data from the response.
frsqrItem.tips = items[i].tips;
frsqrItem.name = items[i].venue.name;
frsqrItem.venueUrl = items[i].venue.url;
frsqrItem.lat = items[i].venue.location.lat;
frsqrItem.lng = items[i].venue.location.lng;
frsqrItem.index = i;
// Photos for the locations, | identifier_body |
shuf.rs | ),
Echo(Vec<String>),
InputRange((usize, usize)),
}
static USAGE: &str = help_usage!("shuf.md");
static ABOUT: &str = help_about!("shuf.md");
struct Options {
head_count: usize,
output: Option<String>,
random_source: Option<String>,
repeat: bool,
sep: u8,
}
mod options {
pub static ECHO: &str = "echo";
pub static INPUT_RANGE: &str = "input-range";
pub static HEAD_COUNT: &str = "head-count";
pub static OUTPUT: &str = "output";
pub static RANDOM_SOURCE: &str = "random-source";
pub static REPEAT: &str = "repeat";
pub static ZERO_TERMINATED: &str = "zero-terminated";
pub static FILE: &str = "file";
}
#[uucore::main]
pub fn uumain(args: impl uucore::Args) -> UResult<()> | .to_string(),
)
};
let options = Options {
head_count: {
let headcounts = matches
.get_many::<String>(options::HEAD_COUNT)
.unwrap_or_default()
.map(|s| s.to_owned())
.collect();
match parse_head_count(headcounts) {
Ok(val) => val,
Err(msg) => return Err(USimpleError::new(1, msg)),
}
},
output: matches.get_one::<String>(options::OUTPUT).map(String::from),
random_source: matches
.get_one::<String>(options::RANDOM_SOURCE)
.map(String::from),
repeat: matches.get_flag(options::REPEAT),
sep: if matches.get_flag(options::ZERO_TERMINATED) {
0x00_u8
} else {
0x0a_u8
},
};
match mode {
Mode::Echo(args) => {
let mut evec = args.iter().map(String::as_bytes).collect::<Vec<_>>();
find_seps(&mut evec, options.sep);
shuf_bytes(&mut evec, options)?;
}
Mode::InputRange((b, e)) => {
let rvec = (b..e).map(|x| format!("{x}")).collect::<Vec<String>>();
let mut rvec = rvec.iter().map(String::as_bytes).collect::<Vec<&[u8]>>();
shuf_bytes(&mut rvec, options)?;
}
Mode::Default(filename) => {
let fdata = read_input_file(&filename)?;
let mut fdata = vec![&fdata[..]];
find_seps(&mut fdata, options.sep);
shuf_bytes(&mut fdata, options)?;
}
}
Ok(())
}
pub fn uu_app() -> Command {
Command::new(uucore::util_name())
.about(ABOUT)
.version(crate_version!())
.override_usage(format_usage(USAGE))
.infer_long_args(true)
.args_override_self(true)
.arg(
Arg::new(options::ECHO)
.short('e')
.long(options::ECHO)
.value_name("ARG")
.help("treat each ARG as an input line")
.use_value_delimiter(false)
.num_args(0..)
.conflicts_with(options::INPUT_RANGE),
)
.arg(
Arg::new(options::INPUT_RANGE)
.short('i')
.long(options::INPUT_RANGE)
.value_name("LO-HI")
.help("treat each number LO through HI as an input line")
.conflicts_with(options::FILE),
)
.arg(
Arg::new(options::HEAD_COUNT)
.short('n')
.long(options::HEAD_COUNT)
.value_name("COUNT")
.help("output at most COUNT lines"),
)
.arg(
Arg::new(options::OUTPUT)
.short('o')
.long(options::OUTPUT)
.value_name("FILE")
.help("write result to FILE instead of standard output")
.value_hint(clap::ValueHint::FilePath),
)
.arg(
Arg::new(options::RANDOM_SOURCE)
.long(options::RANDOM_SOURCE)
.value_name("FILE")
.help("get random bytes from FILE")
.value_hint(clap::ValueHint::FilePath),
)
.arg(
Arg::new(options::REPEAT)
.short('r')
.long(options::REPEAT)
.help("output lines can be repeated")
.action(ArgAction::SetTrue),
)
.arg(
Arg::new(options::ZERO_TERMINATED)
.short('z')
.long(options::ZERO_TERMINATED)
.help("line delimiter is NUL, not newline")
.action(ArgAction::SetTrue),
)
.arg(Arg::new(options::FILE).value_hint(clap::ValueHint::FilePath))
}
fn read_input_file(filename: &str) -> UResult<Vec<u8>> {
let mut file = BufReader::new(if filename == "-" {
Box::new(stdin()) as Box<dyn Read>
} else {
let file = File::open(filename)
.map_err_context(|| format!("failed to open {}", filename.quote()))?;
Box::new(file) as Box<dyn Read>
});
let mut data = Vec::new();
file.read_to_end(&mut data)
.map_err_context(|| format!("failed reading {}", filename.quote()))?;
Ok(data)
}
fn find_seps(data: &mut Vec<&[u8]>, sep: u8) {
// need to use for loop so we don't borrow the vector as we modify it in place
// basic idea:
// * We don't care about the order of the result. This lets us slice the slices
// without making a new vector.
// * Starting from the end of the vector, we examine each element.
// * If that element contains the separator, we remove it from the vector,
// and then sub-slice it into slices that do not contain the separator.
// * We maintain the invariant throughout that each element in the vector past
// the ith element does not have any separators remaining.
for i in (0..data.len()).rev() {
if data[i].contains(&sep) {
let this = data.swap_remove(i);
let mut p = 0;
for i in memchr_iter(sep, this) {
data.push(&this[p..i]);
p = i + 1;
}
if p < this.len() {
data.push(&this[p..]);
}
}
}
}
fn shuf_bytes(input: &mut Vec<&[u8]>, opts: Options) -> UResult<()> {
let mut output = BufWriter::new(match opts.output {
None => Box::new(stdout()) as Box<dyn Write>,
Some(s) => {
let file = File::create(&s[..])
.map_err_context(|| format!("failed to open {} for writing", s.quote()))?;
Box::new(file) as Box<dyn Write>
}
});
let mut rng = match opts.random_source {
Some(r) => {
let file = File::open(&r[..])
.map_err_context(|| format!("failed to open random source {}", r.quote()))?;
WrappedRng::RngFile(rand_read_adapter::ReadRng::new(file))
}
None => WrappedRng::RngDefault(rand::thread_rng()),
};
if input.is_empty() {
return Ok(());
}
if opts.repeat {
for _ in 0..opts.head_count {
// Returns None is the slice is empty. We checked this before, so
// this is safe.
let r = input.choose(&mut rng).unwrap();
output
.write_all(r)
.map_err_context(|| "write failed".to_string())?;
output
.write_all(&[opts.sep])
.map_err_context(|| "write failed".to_string())?;
}
} else {
let (shuffled, _) = input.partial_shuffle(&mut rng, opts.head_count);
for r in shuffled {
output
.write_all(r)
.map_err_context(|| "write failed".to_string())?;
output
.write_all(&[opts.sep])
.map_err_context(|| "write failed".to_string())?;
}
}
Ok(())
}
fn parse_range(input_range: &str) -> Result<(usize, usize), String> {
if let Some((from, to)) = input_range.split_once('-') {
let begin = from
.parse::<usize>()
.map_err(|_| format!("invalid input range: {}", from.quote()))?;
| {
let args = args.collect_lossy();
let matches = uu_app().try_get_matches_from(args)?;
let mode = if let Some(args) = matches.get_many::<String>(options::ECHO) {
Mode::Echo(args.map(String::from).collect())
} else if let Some(range) = matches.get_one::<String>(options::INPUT_RANGE) {
match parse_range(range) {
Ok(m) => Mode::InputRange(m),
Err(msg) => {
return Err(USimpleError::new(1, msg));
}
}
} else {
Mode::Default(
matches
.get_one::<String>(options::FILE)
.map(|s| s.as_str())
.unwrap_or("-") | identifier_body |
shuf.rs | ),
Echo(Vec<String>),
InputRange((usize, usize)),
}
static USAGE: &str = help_usage!("shuf.md");
static ABOUT: &str = help_about!("shuf.md");
struct Options {
head_count: usize,
output: Option<String>,
random_source: Option<String>,
repeat: bool,
sep: u8,
}
mod options {
pub static ECHO: &str = "echo";
pub static INPUT_RANGE: &str = "input-range";
pub static HEAD_COUNT: &str = "head-count";
pub static OUTPUT: &str = "output";
pub static RANDOM_SOURCE: &str = "random-source";
pub static REPEAT: &str = "repeat";
pub static ZERO_TERMINATED: &str = "zero-terminated";
pub static FILE: &str = "file";
}
#[uucore::main]
pub fn | (args: impl uucore::Args) -> UResult<()> {
let args = args.collect_lossy();
let matches = uu_app().try_get_matches_from(args)?;
let mode = if let Some(args) = matches.get_many::<String>(options::ECHO) {
Mode::Echo(args.map(String::from).collect())
} else if let Some(range) = matches.get_one::<String>(options::INPUT_RANGE) {
match parse_range(range) {
Ok(m) => Mode::InputRange(m),
Err(msg) => {
return Err(USimpleError::new(1, msg));
}
}
} else {
Mode::Default(
matches
.get_one::<String>(options::FILE)
.map(|s| s.as_str())
.unwrap_or("-")
.to_string(),
)
};
let options = Options {
head_count: {
let headcounts = matches
.get_many::<String>(options::HEAD_COUNT)
.unwrap_or_default()
.map(|s| s.to_owned())
.collect();
match parse_head_count(headcounts) {
Ok(val) => val,
Err(msg) => return Err(USimpleError::new(1, msg)),
}
},
output: matches.get_one::<String>(options::OUTPUT).map(String::from),
random_source: matches
.get_one::<String>(options::RANDOM_SOURCE)
.map(String::from),
repeat: matches.get_flag(options::REPEAT),
sep: if matches.get_flag(options::ZERO_TERMINATED) {
0x00_u8
} else {
0x0a_u8
},
};
match mode {
Mode::Echo(args) => {
let mut evec = args.iter().map(String::as_bytes).collect::<Vec<_>>();
find_seps(&mut evec, options.sep);
shuf_bytes(&mut evec, options)?;
}
Mode::InputRange((b, e)) => {
let rvec = (b..e).map(|x| format!("{x}")).collect::<Vec<String>>();
let mut rvec = rvec.iter().map(String::as_bytes).collect::<Vec<&[u8]>>();
shuf_bytes(&mut rvec, options)?;
}
Mode::Default(filename) => {
let fdata = read_input_file(&filename)?;
let mut fdata = vec![&fdata[..]];
find_seps(&mut fdata, options.sep);
shuf_bytes(&mut fdata, options)?;
}
}
Ok(())
}
pub fn uu_app() -> Command {
Command::new(uucore::util_name())
.about(ABOUT)
.version(crate_version!())
.override_usage(format_usage(USAGE))
.infer_long_args(true)
.args_override_self(true)
.arg(
Arg::new(options::ECHO)
.short('e')
.long(options::ECHO)
.value_name("ARG")
.help("treat each ARG as an input line")
.use_value_delimiter(false)
.num_args(0..)
.conflicts_with(options::INPUT_RANGE),
)
.arg(
Arg::new(options::INPUT_RANGE)
.short('i')
.long(options::INPUT_RANGE)
.value_name("LO-HI")
.help("treat each number LO through HI as an input line")
.conflicts_with(options::FILE),
)
.arg(
Arg::new(options::HEAD_COUNT)
.short('n')
.long(options::HEAD_COUNT)
.value_name("COUNT")
.help("output at most COUNT lines"),
)
.arg(
Arg::new(options::OUTPUT)
.short('o')
.long(options::OUTPUT)
.value_name("FILE")
.help("write result to FILE instead of standard output")
.value_hint(clap::ValueHint::FilePath),
)
.arg(
Arg::new(options::RANDOM_SOURCE)
.long(options::RANDOM_SOURCE)
.value_name("FILE")
.help("get random bytes from FILE")
.value_hint(clap::ValueHint::FilePath),
)
.arg(
Arg::new(options::REPEAT)
.short('r')
.long(options::REPEAT)
.help("output lines can be repeated")
.action(ArgAction::SetTrue),
)
.arg(
Arg::new(options::ZERO_TERMINATED)
.short('z')
.long(options::ZERO_TERMINATED)
.help("line delimiter is NUL, not newline")
.action(ArgAction::SetTrue),
)
.arg(Arg::new(options::FILE).value_hint(clap::ValueHint::FilePath))
}
fn read_input_file(filename: &str) -> UResult<Vec<u8>> {
let mut file = BufReader::new(if filename == "-" {
Box::new(stdin()) as Box<dyn Read>
} else {
let file = File::open(filename)
.map_err_context(|| format!("failed to open {}", filename.quote()))?;
Box::new(file) as Box<dyn Read>
});
let mut data = Vec::new();
file.read_to_end(&mut data)
.map_err_context(|| format!("failed reading {}", filename.quote()))?;
Ok(data)
}
fn find_seps(data: &mut Vec<&[u8]>, sep: u8) {
// need to use for loop so we don't borrow the vector as we modify it in place
// basic idea:
// * We don't care about the order of the result. This lets us slice the slices
// without making a new vector.
// * Starting from the end of the vector, we examine each element.
// * If that element contains the separator, we remove it from the vector,
// and then sub-slice it into slices that do not contain the separator.
// * We maintain the invariant throughout that each element in the vector past
// the ith element does not have any separators remaining.
for i in (0..data.len()).rev() {
if data[i].contains(&sep) {
let this = data.swap_remove(i);
let mut p = 0;
for i in memchr_iter(sep, this) {
data.push(&this[p..i]);
p = i + 1;
}
if p < this.len() {
data.push(&this[p..]);
}
}
}
}
fn shuf_bytes(input: &mut Vec<&[u8]>, opts: Options) -> UResult<()> {
let mut output = BufWriter::new(match opts.output {
None => Box::new(stdout()) as Box<dyn Write>,
Some(s) => {
let file = File::create(&s[..])
.map_err_context(|| format!("failed to open {} for writing", s.quote()))?;
Box::new(file) as Box<dyn Write>
}
});
let mut rng = match opts.random_source {
Some(r) => {
let file = File::open(&r[..])
.map_err_context(|| format!("failed to open random source {}", r.quote()))?;
WrappedRng::RngFile(rand_read_adapter::ReadRng::new(file))
}
None => WrappedRng::RngDefault(rand::thread_rng()),
};
if input.is_empty() {
return Ok(());
}
if opts.repeat {
for _ in 0..opts.head_count {
// Returns None is the slice is empty. We checked this before, so
// this is safe.
let r = input.choose(&mut rng).unwrap();
output
.write_all(r)
.map_err_context(|| "write failed".to_string())?;
output
.write_all(&[opts.sep])
.map_err_context(|| "write failed".to_string())?;
}
} else {
let (shuffled, _) = input.partial_shuffle(&mut rng, opts.head_count);
for r in shuffled {
output
.write_all(r)
.map_err_context(|| "write failed".to_string())?;
output
.write_all(&[opts.sep])
.map_err_context(|| "write failed".to_string())?;
}
}
Ok(())
}
fn parse_range(input_range: &str) -> Result<(usize, usize), String> {
if let Some((from, to)) = input_range.split_once('-') {
let begin = from
.parse::<usize>()
.map_err(|_| format!("invalid input range: {}", from.quote()))?;
let | uumain | identifier_name |
shuf.rs | ),
Echo(Vec<String>),
InputRange((usize, usize)),
}
static USAGE: &str = help_usage!("shuf.md");
static ABOUT: &str = help_about!("shuf.md");
struct Options {
head_count: usize,
output: Option<String>,
random_source: Option<String>,
repeat: bool,
sep: u8,
}
mod options {
pub static ECHO: &str = "echo";
pub static INPUT_RANGE: &str = "input-range";
pub static HEAD_COUNT: &str = "head-count";
pub static OUTPUT: &str = "output";
pub static RANDOM_SOURCE: &str = "random-source";
pub static REPEAT: &str = "repeat";
pub static ZERO_TERMINATED: &str = "zero-terminated";
pub static FILE: &str = "file";
}
#[uucore::main]
pub fn uumain(args: impl uucore::Args) -> UResult<()> {
let args = args.collect_lossy();
let matches = uu_app().try_get_matches_from(args)?;
let mode = if let Some(args) = matches.get_many::<String>(options::ECHO) {
Mode::Echo(args.map(String::from).collect())
} else if let Some(range) = matches.get_one::<String>(options::INPUT_RANGE) {
match parse_range(range) {
Ok(m) => Mode::InputRange(m),
Err(msg) => {
return Err(USimpleError::new(1, msg));
}
}
} else {
Mode::Default(
matches
.get_one::<String>(options::FILE)
.map(|s| s.as_str())
.unwrap_or("-")
.to_string(),
)
};
let options = Options {
head_count: {
let headcounts = matches
.get_many::<String>(options::HEAD_COUNT)
.unwrap_or_default()
.map(|s| s.to_owned())
.collect();
match parse_head_count(headcounts) {
Ok(val) => val,
Err(msg) => return Err(USimpleError::new(1, msg)),
}
},
output: matches.get_one::<String>(options::OUTPUT).map(String::from),
random_source: matches
.get_one::<String>(options::RANDOM_SOURCE)
.map(String::from),
repeat: matches.get_flag(options::REPEAT),
sep: if matches.get_flag(options::ZERO_TERMINATED) {
0x00_u8
} else {
0x0a_u8
},
};
match mode {
Mode::Echo(args) => {
let mut evec = args.iter().map(String::as_bytes).collect::<Vec<_>>();
find_seps(&mut evec, options.sep);
shuf_bytes(&mut evec, options)?;
}
Mode::InputRange((b, e)) => {
let rvec = (b..e).map(|x| format!("{x}")).collect::<Vec<String>>();
let mut rvec = rvec.iter().map(String::as_bytes).collect::<Vec<&[u8]>>();
shuf_bytes(&mut rvec, options)?;
}
Mode::Default(filename) => {
let fdata = read_input_file(&filename)?;
let mut fdata = vec![&fdata[..]];
find_seps(&mut fdata, options.sep);
shuf_bytes(&mut fdata, options)?;
}
}
Ok(())
}
pub fn uu_app() -> Command {
Command::new(uucore::util_name())
.about(ABOUT) | .override_usage(format_usage(USAGE))
.infer_long_args(true)
.args_override_self(true)
.arg(
Arg::new(options::ECHO)
.short('e')
.long(options::ECHO)
.value_name("ARG")
.help("treat each ARG as an input line")
.use_value_delimiter(false)
.num_args(0..)
.conflicts_with(options::INPUT_RANGE),
)
.arg(
Arg::new(options::INPUT_RANGE)
.short('i')
.long(options::INPUT_RANGE)
.value_name("LO-HI")
.help("treat each number LO through HI as an input line")
.conflicts_with(options::FILE),
)
.arg(
Arg::new(options::HEAD_COUNT)
.short('n')
.long(options::HEAD_COUNT)
.value_name("COUNT")
.help("output at most COUNT lines"),
)
.arg(
Arg::new(options::OUTPUT)
.short('o')
.long(options::OUTPUT)
.value_name("FILE")
.help("write result to FILE instead of standard output")
.value_hint(clap::ValueHint::FilePath),
)
.arg(
Arg::new(options::RANDOM_SOURCE)
.long(options::RANDOM_SOURCE)
.value_name("FILE")
.help("get random bytes from FILE")
.value_hint(clap::ValueHint::FilePath),
)
.arg(
Arg::new(options::REPEAT)
.short('r')
.long(options::REPEAT)
.help("output lines can be repeated")
.action(ArgAction::SetTrue),
)
.arg(
Arg::new(options::ZERO_TERMINATED)
.short('z')
.long(options::ZERO_TERMINATED)
.help("line delimiter is NUL, not newline")
.action(ArgAction::SetTrue),
)
.arg(Arg::new(options::FILE).value_hint(clap::ValueHint::FilePath))
}
fn read_input_file(filename: &str) -> UResult<Vec<u8>> {
let mut file = BufReader::new(if filename == "-" {
Box::new(stdin()) as Box<dyn Read>
} else {
let file = File::open(filename)
.map_err_context(|| format!("failed to open {}", filename.quote()))?;
Box::new(file) as Box<dyn Read>
});
let mut data = Vec::new();
file.read_to_end(&mut data)
.map_err_context(|| format!("failed reading {}", filename.quote()))?;
Ok(data)
}
fn find_seps(data: &mut Vec<&[u8]>, sep: u8) {
// need to use for loop so we don't borrow the vector as we modify it in place
// basic idea:
// * We don't care about the order of the result. This lets us slice the slices
// without making a new vector.
// * Starting from the end of the vector, we examine each element.
// * If that element contains the separator, we remove it from the vector,
// and then sub-slice it into slices that do not contain the separator.
// * We maintain the invariant throughout that each element in the vector past
// the ith element does not have any separators remaining.
for i in (0..data.len()).rev() {
if data[i].contains(&sep) {
let this = data.swap_remove(i);
let mut p = 0;
for i in memchr_iter(sep, this) {
data.push(&this[p..i]);
p = i + 1;
}
if p < this.len() {
data.push(&this[p..]);
}
}
}
}
fn shuf_bytes(input: &mut Vec<&[u8]>, opts: Options) -> UResult<()> {
let mut output = BufWriter::new(match opts.output {
None => Box::new(stdout()) as Box<dyn Write>,
Some(s) => {
let file = File::create(&s[..])
.map_err_context(|| format!("failed to open {} for writing", s.quote()))?;
Box::new(file) as Box<dyn Write>
}
});
let mut rng = match opts.random_source {
Some(r) => {
let file = File::open(&r[..])
.map_err_context(|| format!("failed to open random source {}", r.quote()))?;
WrappedRng::RngFile(rand_read_adapter::ReadRng::new(file))
}
None => WrappedRng::RngDefault(rand::thread_rng()),
};
if input.is_empty() {
return Ok(());
}
if opts.repeat {
for _ in 0..opts.head_count {
// Returns None is the slice is empty. We checked this before, so
// this is safe.
let r = input.choose(&mut rng).unwrap();
output
.write_all(r)
.map_err_context(|| "write failed".to_string())?;
output
.write_all(&[opts.sep])
.map_err_context(|| "write failed".to_string())?;
}
} else {
let (shuffled, _) = input.partial_shuffle(&mut rng, opts.head_count);
for r in shuffled {
output
.write_all(r)
.map_err_context(|| "write failed".to_string())?;
output
.write_all(&[opts.sep])
.map_err_context(|| "write failed".to_string())?;
}
}
Ok(())
}
fn parse_range(input_range: &str) -> Result<(usize, usize), String> {
if let Some((from, to)) = input_range.split_once('-') {
let begin = from
.parse::<usize>()
.map_err(|_| format!("invalid input range: {}", from.quote()))?;
let | .version(crate_version!()) | random_line_split |
eda.py | attributes out into year, month, date
# - Column 'relative_humidity' and 'psi' contains value 0 --> This is quite unrealistic, so we can treat them as missing value too
# %% [markdown]
# We will take a closer look at the 'weather' column
# %%
df['weather'].value_counts()
# %% [markdown]
# Looks like 'weather' column is not so clean: there are typo errors and misspellings that will need to be cleaned up before they can be transformed to numeric values.
# %% [markdown]
# By examining the data columns above, we have concluded that there are some data cleaning/preprocessing tasks we will need to perform:
# - Column 'weather': clean up the typos, and transform to numeric data
# - Column 'guest_scooter': remove negative values
# - Column 'relative_humidity','psi': value 0 should be treated as null values
# - Column 'date': transform to numeric data e.g. year, month, date, day of week
#
# With these tasks in mind, we can now start to preprocess the data
# %% [markdown]
# ## Section 3. Data Cleaning & Preprocessing
# %% [markdown]
# #### Task 1: Column 'Weather' -- Clean up typos and transform categorical data to numeric data
# %%
df['weather']=df['weather'].str.upper()
df['weather'].value_counts()
# %%
# now we correct some of the spellings
df.loc[df['weather']=='LEAR','weather']='CLEAR'
df.loc[df['weather']=='CLAR','weather']='CLEAR'
df.loc[df['weather']=='CLUDY','weather']='CLOUDY'
df.loc[df['weather']=='LOUDY','weather']='CLOUDY'
df.loc[df['weather']=='LIHT SNOW/RAIN','weather']='LIGHT SNOW/RAIN'
df['weather'].value_counts()
# %%
# map weather to integer
map_weather = {'CLEAR': 1, 'CLOUDY': 2,'LIGHT SNOW/RAIN':3,'HEAVY SNOW/RAIN':4}
df.replace({'weather': map_weather})
# %% [markdown]
# We are set!
# %% [markdown]
# #### Task 2: Column 'guest_scooter' -- remove negative values
# %%
df['guest_scooter'].replace([-2,-1],np.nan,inplace=True)
df['guest_scooter'].describe()
# %% [markdown]
# Notice that the count has reduced. This means that some of the negative values have been converted to NULL
# %% [markdown]
# #### Task 3: Column 'relative_humidity','psi' - value 0 should be treated as null values
# %%
df['relative_humidity'].replace(0,np.nan,inplace=True)
df['relative_humidity'].describe()
| df['psi'].describe()
# %% [markdown]
# #### Task 4: Column 'date' - transform to numeric data e.g. year, month, date, day of week
# %%
# first we convert the date column to date object
dt=df['date']
df['date']=pd.to_datetime(dt)
# %%
type(df['date'].values[0])
# %% [markdown]
# Now we need to create new columns and extract the year, month, date, and weekday attributes
# %%
df['year']=df['date'].dt.year
df['month']=df['date'].dt.month
df['day']=df['date'].dt.day
df['weekday']=df['date'].apply(lambda x: x.weekday())
df.head()
# %% [markdown]
# In the 'weekday' column, we notice that the number ranges from 0 to 6. We are going to add 1 to the values.
# %%
df['weekday']=df['weekday']+1
df.head()
# %% [markdown]
# We are done with preprocessing! These steps will also be performed in the data cleaning/preprocessing phase of the MLP.
# Now on to exploring some basic relationships between the attributes.
# %% [markdown]
# ## Section 4. Data Visualization
#
#
# Bearing in mind that the final purpose of the exercise is to predict the number of active e-scooter riders. Over here, we should try to aim to look for relationships between the other attributes with the target attribute (guest_scooters).
# %%
# first we look at the distribution of ridership/hour
plt.hist(df['guest_scooter'].dropna(),bins=50)
plt.xlabel('Number of Scooters/Hr')
plt.ylabel('Frequency')
plt.show()
# %% [markdown]
# We can see that the number of scooters/hour does not follow a normal distribution. It is heavily right-tailed.
# %%
# now we want to see if there's a longitudinal trend on daily total scooter number
group_by_day=df.groupby('date').guest_scooter.sum()
group_by_day.plot()
plt.xlabel('Date')
plt.ylabel('Daily Ridership')
plt.title("Total Number of Scooters/Day")
plt.show()
# %% [markdown]
# What we can observe from above is a clear seasonal trend: numbers spike up in horter months (May - August), and drop in colder months (Oct - Apr). There are also high variations across days
# %%
# we can see the seasonal trends more clearly if we plot the total number of riders per month
group_by_month=df.groupby('month').guest_scooter.sum()
group_by_month.plot(kind='bar')
plt.xlabel('Month')
plt.ylabel('Scooter Number')
plt.title("Total number of scooters in each month over the study period")
plt.show()
# %%
sns.heatmap(pd.crosstab(df['month'],df['hr'],values=df['guest_scooter'],aggfunc='mean'),cmap='coolwarm')
plt.ylabel("Month")
plt.xlabel("Hour of day")
plt.title("Mean hourly scooters per hour")
# %%
# %%
sns.boxenplot(data=df, x="month",y="guest_scooter")
plt.title("Distribution of guest scooter numbers across months")
plt.show()
# %% [markdown]
# The number of scooters shows clear monthly and daily variations.
# What actually makes a difference in the different months and hours is the climatic conditions, this means:
# - temperature
# - feels_like_temperature
# - relative_humidity
# - psi
# - windspeed
# - weather
#
# All these conditions will have some influence in the result. What is interesting from the boxenplot above is also that there are two peaks/maximum hourly ridership in a year, one in March-May, the other in September - November. Now we know that Summer is usually the hottest time of a year -- we are talking about temperature here. If so, it looks like other factors e.g. thermal comfort, wind, and humidity might have an effect to scooters too. We will look at this in detail.
# %% [markdown]
# Before we drill in on the climatic factors, there is another temporal scale we have not explored: variations in a day.
# %%
# resample by hour of day
mean_rider_by_hr = df.groupby('hr')['guest_scooter'].mean().sort_index()
mean_temp_by_hr = df.groupby('hr')['temperature'].mean().sort_index()
mean_fltemp_by_hr=df.groupby('hr')['feels_like_temperature'].mean().sort_index()
mean_psi_by_hr=df.groupby('hr')['psi'].mean().sort_index()
mean_rh_by_hr=df.groupby('hr')['relative_humidity'].mean().sort_index()
mean_windspeed_by_hr=df.groupby('hr')['windspeed'].mean().sort_index()
# %%
# merge into a new dataframe
frame={'rider':mean_rider_by_hr,
'temp':mean_temp_by_hr,
'fltemp':mean_fltemp_by_hr,
'psi':mean_psi_by_hr,
'rh':mean_rh_by_hr,
'windspeed':mean_windspeed_by_hr}
df_by_hr=pd.DataFrame(frame)
df_by_hr.head()
# %%
# plot the variables to 4 subplots
fig,ax=plt.subplots(2,2)
fig.tight_layout(pad=0.4, w_pad=6, h_pad=2.0)
ax[0,0].bar(df_by_hr.index,df_by_hr['rider'],alpha=0.2,color='grey')
ax2=ax[0,0].twinx()
ax2.plot(df_by_hr.index,df_by_hr['temp'],color='r')
ax2.plot(df_by_hr.index,df_by_hr['fltemp'],color='orange')
ax[0,0].set_ylabel('Mean Hourly Ridership')
ax2.set_ylabel("Temperature/Feels_like_temperature")
plt.legend(['Temperature','Feels_like_Temperature'])
ax[0,1].bar(df_by_hr.index,df_by_hr['rider'],alpha=0.2,color='grey')
ax3=ax[0,1].twinx()
ax3.plot(df_by_hr.index,df_by_hr['windspeed'],color='maroon')
ax[0,1].set_ylabel('Mean Hourly Ridership')
ax3.set_ylabel("Windspeed")
plt.legend(['Windspeed'])
ax[1,0].bar(df_by_hr.index,df_by_hr['rider'],alpha=0.2,color='grey')
ax4=ax[1,0].twinx()
ax4.plot(df_by_hr.index,df_by_hr['rh'],color='green')
ax[1,0].set_ylabel('Mean Hourly Ridership')
ax4.set_ylabel('Relative Humidity')
plt.legend(['Relative Humidity'])
ax[1,1]. | # %%
df['psi'].replace(0,np.nan,inplace=True) | random_line_split |
form-field-config.type.ts | .
*
* OpenMediaVault is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*/
import { Constraint } from '~/app/shared/models/constraint.type';
import { DataStore } from '~/app/shared/models/data-store.type';
import { DatatableColumn } from '~/app/shared/models/datatable-column.type';
import { Sorter } from '~/app/shared/models/sorter.type';
export type FormFieldConfig = {
// The following field controls are supported:
// .--------------------------------------------------------------------.
// | Type | Description |
// |-------------|------------------------------------------------------|
// | confObjUuid | This is a hidden field that contains an UUID. By |
// | | default it is set to the UUID that tells the backend |
// | | that it should handle the the data to create a new |
// | | database configuration object. |
// | hidden | A hidden form field. |
// | divider | Draws a horizontal line. |
// | paragraph | Displays a title and draws a horizontal line. |
// | ... | ... |
// | container | Align child fields in horizontal order. |
// '--------------------------------------------------------------------'
type:
| 'confObjUuid'
| 'hidden'
| 'divider'
| 'paragraph'
| 'button'
| 'iconButton'
| 'textInput'
| 'folderBrowser'
| 'numberInput'
| 'checkbox'
| 'textarea'
| 'fileInput'
| 'select'
| 'sharedFolderSelect'
| 'sslCertSelect'
| 'passwordInput'
| 'datePicker'
| 'datatable'
| 'slider'
| 'container';
name?: string;
label?: string;
placeholder?: string;
tooltip?: string;
hint?: string;
value?: any;
readonly?: boolean;
// Disable the field.
// Use a tokenized string to be able to mark the field as disabled
// dynamically based on its evaluation result. The special token
// `_routeConfig` can be to access the components route configuration.
// Make sure the token will be evaluated to a boolean value.
//
// Example:
// { disabled: '{{ _routeConfig.data.editing | toboolean }}' }
// The component's route configuration:
// {
// path: 'edit/:name',
// component: UserFormPageComponent,
// data: {
// title: gettext('Edit'),
// editing: true,
// notificationTitle: gettext('Updated user "{{ name }}".')
// }
// }
disabled?: boolean | string;
// Modify the form field depending on a specified constraint. The
// constraint must be truthy to apply.
modifiers?: Array<{
type:
| 'disabled'
| 'enabled'
| 'checked'
| 'unchecked'
| 'focused'
| 'visible'
| 'hidden'
| 'value';
// Optional configuration used by modifiers. This is required by
// the 'value' modifier, e.g. '{{ <NAME> }}' to set the value
// of the given field.
typeConfig?: any;
// Apply the opposite type, e.g. `disabled` for `enabled`, if the
// constraint is falsy. Defaults to `true`.
opposite?: boolean;
// The constraint can access the current form field
// values, e.g. '{ field: '<NAME>' }'
constraint: Constraint;
}>;
autofocus?: boolean;
icon?: string;
submitValue?: boolean;
// The event name for control to update upon. Defaults to `change`.
updateOn?: 'change' | 'blur';
// The validators to ensure the form field content is
// in a valid state.
validators?: {
// Mark the field as required.
// Use a tokenized string to be able to mark the field as required
// dynamically based on its evaluation result. The special token
// `_routeConfig` can be to access the components route configuration.
// Make sure the token will be evaluated to a boolean value.
required?: boolean | string;
// When the constraint succeeds and the control has
// an empty value, then the 'required' error is set.
requiredIf?: Constraint;
minLength?: number;
maxLength?: number;
min?: number;
max?: number;
email?: boolean;
custom?: Array<FormFieldConstraintValidator>;
pattern?: {
pattern: string;
errorData?: any;
};
patternType?:
| 'userName'
| 'groupName'
| 'shareName'
| 'email'
| 'ip'
| 'ipv4'
| 'ipv6'
| 'ipList'
| 'ipNetCidr'
| 'hostName'
| 'hostNameIpNetCidr'
| 'hostNameIpNetCidrList'
| 'domainName'
| 'domainNameList'
| 'domainNameIp'
| 'domainNameIpList'
| 'port'
| 'time'
| 'sshPubKey'
| 'sshPubKeyRfc4716'
| 'sshPubKeyOpenSsh'
| 'netmask';
};
// --- container ---
fields?: Array<FormFieldConfig>;
// Fields in a container will respect the 'flex' configuration.
// Specifies the size of the field in percent.
flex?: number;
// --- button | paragraph ---
text?: string;
// --- button | iconButton ---
click?: () => void;
request?: {
// The name of the RPC service.
service: string;
// The name of the RPC.
method: string;
// The RPC parameters. The parameters will be formatted using
// the values from the parent form.
params?: Record<string, any>;
// Set `true` if the RPC is a background task.
task?: boolean;
// If a message is defined, then the UI will be blocked and
// the message is displayed while the request is running.
progressMessage?: string;
// Display a notification when the request was successful.
successNotification?: string;
// Navigate to this URL when the request was successful.
// The URL will be formatted using the values from the parent
// form. The RPC response is accessible via '_response'.
// Example:
// /foo/bar/{{ xyz }}/{{ _response['baz'] }}
// where `xyz` will be replaced by the value of the form field
// named `xyz` and `_response['baz']` by the property `baz` of
// the map/object returned by the RPC.
// Example:
// /externalRedirect/{{ _response['url'] }}
// Redirect to an external URL. The URL must not be escaped,
// this will be done automatically.
successUrl?: string;
};
// The URL will be formatted using the parent form field values.
url?: string;
// --- folderBrowser ---
dirType?: 'sharedfolder' | 'mntent';
// The name of the field that contains the UUID of the
// shared folder or mount point configuration object.
dirRefIdField?: string;
// --- numberInput | slider ---
step?: number;
| // --- numberInput | password | textInput ---
autocomplete?: string;
// Note, this button is only visible if the browser supports
// that. The following requirements must be met:
// - The HTTPS protocol is used. localhost is also supported.
// - The site is not embedded in an iFrame.
hasCopyToClipboardButton?: boolean;
// --- textarea | fileInput ---
rows?: number;
cols?: number;
// --- textarea | textInput | fileInput ---
// Use a monospaced font.
monospaced?: boolean;
// --- fileInput ---
accept?: string;
// --- select ---
multiple?: boolean;
// Defaults to 'value'.
valueField?: string;
// Defaults to 'text'.
textField?: string;
store?: DataStore;
// Add an empty option to be able to clear the selection.
hasEmptyOption?: boolean;
// The text displayed in the option with the empty value.
// Defaults to 'None'.
emptyOptionText?: string;
selectionChange?: (value: any) => void;
// --- paragraph ---
hasDivider?: boolean;
// --- datatable ---
columns?: Array<DatatableColumn>;
columnMode?: 'standard' | 'flex' | 'force';
hasActionBar?: boolean;
hasSearchField?: boolean;
hasHeader?: boolean;
hasFooter?: boolean;
selectionType?: 'none' | 'single' | 'multi';
limit?: number;
sorters?: Array<Sorter>;
actions?: Array<{
// Specifies a template which pre-configures the action button.
// add - Shows a form dialog. When the dialog is successfully
// closed, then the form values will be used to add a new
// row to the datatable.
// edit - Shows a form dialog which displays the data of the
// current selected row. The action button is only enabled
// when one row is | random_line_split |
|
lib.rs | ::fmt;
use issues::{BadIssueSeeker, Issue};
use filemap::FileMap;
use visitor::FmtVisitor;
use config::Config;
#[macro_use]
mod utils;
pub mod config;
pub mod filemap;
mod visitor;
mod checkstyle;
mod items;
mod missed_spans;
mod lists;
mod types;
mod expr;
mod imports;
mod issues;
mod rewrite;
mod string;
mod comment;
mod modules;
pub mod rustfmt_diff;
mod chains;
mod macros;
mod patterns;
const MIN_STRING: usize = 10;
// When we get scoped annotations, we should have rustfmt::skip.
const SKIP_ANNOTATION: &'static str = "rustfmt_skip";
pub trait Spanned {
fn span(&self) -> Span;
}
impl Spanned for ast::Expr {
fn span(&self) -> Span {
self.span
}
}
impl Spanned for ast::Pat {
fn span(&self) -> Span {
self.span
}
}
impl Spanned for ast::Ty {
fn span(&self) -> Span {
self.span
}
}
impl Spanned for ast::Arg {
fn span(&self) -> Span {
if items::is_named_arg(self) {
mk_sp(self.pat.span.lo, self.ty.span.hi)
} else {
self.ty.span
}
}
}
#[derive(Copy, Clone, Debug)]
pub struct Indent {
// Width of the block indent, in characters. Must be a multiple of
// Config::tab_spaces.
pub block_indent: usize,
// Alignment in characters.
pub alignment: usize,
}
impl Indent {
pub fn new(block_indent: usize, alignment: usize) -> Indent {
Indent {
block_indent: block_indent,
alignment: alignment,
}
}
pub fn empty() -> Indent {
Indent::new(0, 0)
}
pub fn block_indent(mut self, config: &Config) -> Indent {
self.block_indent += config.tab_spaces;
self
}
pub fn block_unindent(mut self, config: &Config) -> Indent {
self.block_indent -= config.tab_spaces;
self
}
pub fn width(&self) -> usize {
self.block_indent + self.alignment
}
pub fn to_string(&self, config: &Config) -> String {
let (num_tabs, num_spaces) = if config.hard_tabs {
(self.block_indent / config.tab_spaces, self.alignment)
} else {
(0, self.block_indent + self.alignment)
};
let num_chars = num_tabs + num_spaces;
let mut indent = String::with_capacity(num_chars);
for _ in 0..num_tabs {
indent.push('\t')
}
for _ in 0..num_spaces {
indent.push(' ')
}
indent
}
}
impl Add for Indent {
type Output = Indent;
fn add(self, rhs: Indent) -> Indent {
Indent {
block_indent: self.block_indent + rhs.block_indent, | }
impl Sub for Indent {
type Output = Indent;
fn sub(self, rhs: Indent) -> Indent {
Indent::new(self.block_indent - rhs.block_indent,
self.alignment - rhs.alignment)
}
}
impl Add<usize> for Indent {
type Output = Indent;
fn add(self, rhs: usize) -> Indent {
Indent::new(self.block_indent, self.alignment + rhs)
}
}
impl Sub<usize> for Indent {
type Output = Indent;
fn sub(self, rhs: usize) -> Indent {
Indent::new(self.block_indent, self.alignment - rhs)
}
}
pub enum ErrorKind {
// Line has exceeded character limit
LineOverflow,
// Line ends in whitespace
TrailingWhitespace,
// TO-DO or FIX-ME item without an issue number
BadIssue(Issue),
}
impl fmt::Display for ErrorKind {
fn fmt(&self, fmt: &mut fmt::Formatter) -> Result<(), fmt::Error> {
match *self {
ErrorKind::LineOverflow => write!(fmt, "line exceeded maximum length"),
ErrorKind::TrailingWhitespace => write!(fmt, "left behind trailing whitespace"),
ErrorKind::BadIssue(issue) => write!(fmt, "found {}", issue),
}
}
}
// Formatting errors that are identified *after* rustfmt has run.
pub struct FormattingError {
line: u32,
kind: ErrorKind,
}
impl FormattingError {
fn msg_prefix(&self) -> &str {
match self.kind {
ErrorKind::LineOverflow |
ErrorKind::TrailingWhitespace => "Rustfmt failed at",
ErrorKind::BadIssue(_) => "WARNING:",
}
}
fn msg_suffix(&self) -> &str {
match self.kind {
ErrorKind::LineOverflow |
ErrorKind::TrailingWhitespace => "(sorry)",
ErrorKind::BadIssue(_) => "",
}
}
}
pub struct FormatReport {
// Maps stringified file paths to their associated formatting errors.
file_error_map: HashMap<String, Vec<FormattingError>>,
}
impl FormatReport {
pub fn warning_count(&self) -> usize {
self.file_error_map.iter().map(|(_, ref errors)| errors.len()).fold(0, |acc, x| acc + x)
}
}
impl fmt::Display for FormatReport {
// Prints all the formatting errors.
fn fmt(&self, fmt: &mut fmt::Formatter) -> Result<(), fmt::Error> {
for (file, errors) in &self.file_error_map {
for error in errors {
try!(write!(fmt,
"{} {}:{}: {} {}\n",
error.msg_prefix(),
file,
error.line,
error.kind,
error.msg_suffix()));
}
}
Ok(())
}
}
// Formatting which depends on the AST.
fn fmt_ast(krate: &ast::Crate,
parse_session: &ParseSess,
main_file: &Path,
config: &Config)
-> FileMap {
let mut file_map = FileMap::new();
for (path, module) in modules::list_files(krate, parse_session.codemap()) {
if config.skip_children && path.as_path() != main_file {
continue;
}
let path = path.to_str().unwrap();
if config.verbose {
println!("Formatting {}", path);
}
let mut visitor = FmtVisitor::from_codemap(parse_session, config);
visitor.format_separate_mod(module);
file_map.insert(path.to_owned(), visitor.buffer);
}
file_map
}
// Formatting done on a char by char or line by line basis.
// TODO(#209) warn on bad license
// TODO(#20) other stuff for parity with make tidy
pub fn fmt_lines(file_map: &mut FileMap, config: &Config) -> FormatReport {
let mut truncate_todo = Vec::new();
let mut report = FormatReport { file_error_map: HashMap::new() };
// Iterate over the chars in the file map.
for (f, text) in file_map.iter() {
let mut trims = vec![];
let mut last_wspace: Option<usize> = None;
let mut line_len = 0;
let mut cur_line = 1;
let mut newline_count = 0;
let mut errors = vec![];
let mut issue_seeker = BadIssueSeeker::new(config.report_todo, config.report_fixme);
for (c, b) in text.chars() {
if c == '\r' {
line_len += c.len_utf8();
continue;
}
// Add warnings for bad todos/ fixmes
if let Some(issue) = issue_seeker.inspect(c) {
errors.push(FormattingError {
line: cur_line,
kind: ErrorKind::BadIssue(issue),
});
}
if c == '\n' {
// Check for (and record) trailing whitespace.
if let Some(lw) = last_wspace {
trims.push((cur_line, lw, b));
line_len -= b - lw;
}
// Check for any line width errors we couldn't correct.
if line_len > config.max_width {
errors.push(FormattingError {
line: cur_line,
kind: ErrorKind::LineOverflow,
});
}
line_len = 0;
cur_line += 1;
newline_count += 1;
last_wspace = None;
} else {
newline_count = 0;
line_len += c.len_utf8();
if c.is_whitespace() {
if last_wspace.is_none() {
last_wspace = Some(b);
}
} else {
last_wspace = None;
}
}
}
if newline_count > 1 {
debug!("track truncate: {} {} {}", f, text.len, newline_count);
truncate_todo.push((f.to_owned(), text.len - newline_count + 1))
}
for &(l, _, _) in &trims {
errors.push(FormattingError {
line: l,
kind: ErrorKind::TrailingWhitespace,
});
}
report.file_error_map.insert(f.to_owned(), errors);
}
for (f, l) in truncate_todo {
file_map.get_mut(&f).unwrap().truncate(l);
}
report
}
pub fn format_string(input: | alignment: self.alignment + rhs.alignment,
}
} | random_line_split |
lib.rs | ::fmt;
use issues::{BadIssueSeeker, Issue};
use filemap::FileMap;
use visitor::FmtVisitor;
use config::Config;
#[macro_use]
mod utils;
pub mod config;
pub mod filemap;
mod visitor;
mod checkstyle;
mod items;
mod missed_spans;
mod lists;
mod types;
mod expr;
mod imports;
mod issues;
mod rewrite;
mod string;
mod comment;
mod modules;
pub mod rustfmt_diff;
mod chains;
mod macros;
mod patterns;
const MIN_STRING: usize = 10;
// When we get scoped annotations, we should have rustfmt::skip.
const SKIP_ANNOTATION: &'static str = "rustfmt_skip";
pub trait Spanned {
fn span(&self) -> Span;
}
impl Spanned for ast::Expr {
fn | (&self) -> Span {
self.span
}
}
impl Spanned for ast::Pat {
fn span(&self) -> Span {
self.span
}
}
impl Spanned for ast::Ty {
fn span(&self) -> Span {
self.span
}
}
impl Spanned for ast::Arg {
fn span(&self) -> Span {
if items::is_named_arg(self) {
mk_sp(self.pat.span.lo, self.ty.span.hi)
} else {
self.ty.span
}
}
}
#[derive(Copy, Clone, Debug)]
pub struct Indent {
// Width of the block indent, in characters. Must be a multiple of
// Config::tab_spaces.
pub block_indent: usize,
// Alignment in characters.
pub alignment: usize,
}
impl Indent {
pub fn new(block_indent: usize, alignment: usize) -> Indent {
Indent {
block_indent: block_indent,
alignment: alignment,
}
}
pub fn empty() -> Indent {
Indent::new(0, 0)
}
pub fn block_indent(mut self, config: &Config) -> Indent {
self.block_indent += config.tab_spaces;
self
}
pub fn block_unindent(mut self, config: &Config) -> Indent {
self.block_indent -= config.tab_spaces;
self
}
pub fn width(&self) -> usize {
self.block_indent + self.alignment
}
pub fn to_string(&self, config: &Config) -> String {
let (num_tabs, num_spaces) = if config.hard_tabs {
(self.block_indent / config.tab_spaces, self.alignment)
} else {
(0, self.block_indent + self.alignment)
};
let num_chars = num_tabs + num_spaces;
let mut indent = String::with_capacity(num_chars);
for _ in 0..num_tabs {
indent.push('\t')
}
for _ in 0..num_spaces {
indent.push(' ')
}
indent
}
}
impl Add for Indent {
type Output = Indent;
fn add(self, rhs: Indent) -> Indent {
Indent {
block_indent: self.block_indent + rhs.block_indent,
alignment: self.alignment + rhs.alignment,
}
}
}
impl Sub for Indent {
type Output = Indent;
fn sub(self, rhs: Indent) -> Indent {
Indent::new(self.block_indent - rhs.block_indent,
self.alignment - rhs.alignment)
}
}
impl Add<usize> for Indent {
type Output = Indent;
fn add(self, rhs: usize) -> Indent {
Indent::new(self.block_indent, self.alignment + rhs)
}
}
impl Sub<usize> for Indent {
type Output = Indent;
fn sub(self, rhs: usize) -> Indent {
Indent::new(self.block_indent, self.alignment - rhs)
}
}
pub enum ErrorKind {
// Line has exceeded character limit
LineOverflow,
// Line ends in whitespace
TrailingWhitespace,
// TO-DO or FIX-ME item without an issue number
BadIssue(Issue),
}
impl fmt::Display for ErrorKind {
fn fmt(&self, fmt: &mut fmt::Formatter) -> Result<(), fmt::Error> {
match *self {
ErrorKind::LineOverflow => write!(fmt, "line exceeded maximum length"),
ErrorKind::TrailingWhitespace => write!(fmt, "left behind trailing whitespace"),
ErrorKind::BadIssue(issue) => write!(fmt, "found {}", issue),
}
}
}
// Formatting errors that are identified *after* rustfmt has run.
pub struct FormattingError {
line: u32,
kind: ErrorKind,
}
impl FormattingError {
fn msg_prefix(&self) -> &str {
match self.kind {
ErrorKind::LineOverflow |
ErrorKind::TrailingWhitespace => "Rustfmt failed at",
ErrorKind::BadIssue(_) => "WARNING:",
}
}
fn msg_suffix(&self) -> &str {
match self.kind {
ErrorKind::LineOverflow |
ErrorKind::TrailingWhitespace => "(sorry)",
ErrorKind::BadIssue(_) => "",
}
}
}
pub struct FormatReport {
// Maps stringified file paths to their associated formatting errors.
file_error_map: HashMap<String, Vec<FormattingError>>,
}
impl FormatReport {
pub fn warning_count(&self) -> usize {
self.file_error_map.iter().map(|(_, ref errors)| errors.len()).fold(0, |acc, x| acc + x)
}
}
impl fmt::Display for FormatReport {
// Prints all the formatting errors.
fn fmt(&self, fmt: &mut fmt::Formatter) -> Result<(), fmt::Error> {
for (file, errors) in &self.file_error_map {
for error in errors {
try!(write!(fmt,
"{} {}:{}: {} {}\n",
error.msg_prefix(),
file,
error.line,
error.kind,
error.msg_suffix()));
}
}
Ok(())
}
}
// Formatting which depends on the AST.
fn fmt_ast(krate: &ast::Crate,
parse_session: &ParseSess,
main_file: &Path,
config: &Config)
-> FileMap {
let mut file_map = FileMap::new();
for (path, module) in modules::list_files(krate, parse_session.codemap()) {
if config.skip_children && path.as_path() != main_file {
continue;
}
let path = path.to_str().unwrap();
if config.verbose {
println!("Formatting {}", path);
}
let mut visitor = FmtVisitor::from_codemap(parse_session, config);
visitor.format_separate_mod(module);
file_map.insert(path.to_owned(), visitor.buffer);
}
file_map
}
// Formatting done on a char by char or line by line basis.
// TODO(#209) warn on bad license
// TODO(#20) other stuff for parity with make tidy
pub fn fmt_lines(file_map: &mut FileMap, config: &Config) -> FormatReport {
let mut truncate_todo = Vec::new();
let mut report = FormatReport { file_error_map: HashMap::new() };
// Iterate over the chars in the file map.
for (f, text) in file_map.iter() {
let mut trims = vec![];
let mut last_wspace: Option<usize> = None;
let mut line_len = 0;
let mut cur_line = 1;
let mut newline_count = 0;
let mut errors = vec![];
let mut issue_seeker = BadIssueSeeker::new(config.report_todo, config.report_fixme);
for (c, b) in text.chars() {
if c == '\r' {
line_len += c.len_utf8();
continue;
}
// Add warnings for bad todos/ fixmes
if let Some(issue) = issue_seeker.inspect(c) {
errors.push(FormattingError {
line: cur_line,
kind: ErrorKind::BadIssue(issue),
});
}
if c == '\n' {
// Check for (and record) trailing whitespace.
if let Some(lw) = last_wspace {
trims.push((cur_line, lw, b));
line_len -= b - lw;
}
// Check for any line width errors we couldn't correct.
if line_len > config.max_width {
errors.push(FormattingError {
line: cur_line,
kind: ErrorKind::LineOverflow,
});
}
line_len = 0;
cur_line += 1;
newline_count += 1;
last_wspace = None;
} else {
newline_count = 0;
line_len += c.len_utf8();
if c.is_whitespace() {
if last_wspace.is_none() {
last_wspace = Some(b);
}
} else {
last_wspace = None;
}
}
}
if newline_count > 1 {
debug!("track truncate: {} {} {}", f, text.len, newline_count);
truncate_todo.push((f.to_owned(), text.len - newline_count + 1))
}
for &(l, _, _) in &trims {
errors.push(FormattingError {
line: l,
kind: ErrorKind::TrailingWhitespace,
});
}
report.file_error_map.insert(f.to_owned(), errors);
}
for (f, l) in truncate_todo {
file_map.get_mut(&f).unwrap().truncate(l);
}
report
}
pub fn format_string(input | span | identifier_name |
trajectory_tracking.py | errari')
data_pose_x = data.pose[indice].position.x
data_pose_y = data.pose[indice].position.y
self.data_pose= np.array([data_pose_x,data_pose_y])
return self.data_pose
def odometryCb(self,msg):
#current robot pose
x = round(msg.pose.pose.position.x,4)
y = round(msg.pose.pose.position.y,4)
theta = round(self.get_angle_pose(msg.pose.pose),4)
y=round(y-1.4*np.cos(theta+np.pi),4) ##spostiamo il punto di riferimento dal centro della macchina al centro dell' asse posteriore
x=round(x+1.4*np.sin(theta+np.pi),4) ##spostiamo il punto di riferimento dal centro della macchina al centro dell' asse posteriore
self.q = np.array([x, y, theta])
return self.q
def get_angle_pose(self, quaternion_pose):
#compute angle from quaternion
#
q = [quaternion_pose.orientation.x,
quaternion_pose.orientation.y,
quaternion_pose.orientation.z,
quaternion_pose.orientation.w]
roll, pitch, yaw = euler_from_quaternion(q)
theta = yaw
tol=0.1
if abs(abs(theta)-abs(self.thetaprec))>2*np.pi-tol and self.thetaprec!=0:
theta=theta+2*np.pi-tol
else:
pass
self.thetaprec=theta
return theta
def trajectory_generation(self, traj,a):
data = rospy.wait_for_message("/gazebo/model_states", ModelStates, timeout=5)
posizione = self.callback(data)
x = round(posizione[0],1)
y = round(posizione[1],1)
tg = Trajectory_generation()
q_i = self.get_pose()
self.trajectory=traj
if(self.trajectory == "parallel_parking" ):
(self.x_d, self.y_d,self.dotx_d,self.doty_d,self.v_d, self.w_d , self.theta_d , self.psi, self.A_park) =tg.parallel_parking_trajectory(q_i, self.t,a)
#self.A_park indica il punto di partenza della manovra di parcheggio, ovvero punto in corrispondenza del parcheggio successivo a quello libero
def get_laser(self):
flag_free=self.a.park()[0] # flag che indica se ha trovato un parcheggio libero
flag_occ=self.a.park()[1] # flag che indica se il parcheggio si trova prima(2), in corrispondenza(1) o dopo(0)
if flag_free==1:
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
return flag_free,flag_occ
def get_pose(self):
#get robot position updated from callback
x = self.q[0]
y = self.q[1]
theta = self.q[2]
return np.array([x, y, theta])
def get_error(self, T,traj):
| print("y_d:{} and y_odom:{} sample:{}".format(self.y_d[T][0],y,T))
print("theta_d:{} and theta_odom:{} sample:{}".format(self.theta_d[T],theta,T))
return np.array([float(e1), float(e2), e3])
def unicicle_Linear_control(self,traj,zeta,a):
rospy.sleep(0.1) # need small time to setup q in callback
max_t = self.t[len(self.t) - 1]
len_t = len(self.t)
self.trajectory=traj
if(self.trajectory == "parallel_parking" ):
for i in np.arange(0, len_t):
now = rospy.get_time()
err = self.get_error(i,self.trajectory)
if round(0.03*len_t)<=i<=round(0.87*len_t) : #tra il 3% e l' 87% uso il controllore
(v, w) = Linear_control_law(err, self.v_d[i], self.w_d[i],zeta,a)
else: # utilizziamo nella parte iniziale e finale le desiderate
#(evitiamo gli spike del controllore dovuti a valori prossimi allo zero di v e w)
v=self.v_d[i]
w=self.w_d[i]
# (v, w) = Linear_control_law(err, self.v_d[i], self.w_d[i],zeta,a)
print("theta_d:{} and theta_odom:{} sample:{}".format(self.theta_d[i], self.q[2] , i) )
print("v_d:{} and v:{} sample:{}".format(self.v_d[i], v , i) )
print("w_d:{} and w:{} sample:{}".format(-self.w_d[i], w , i) )
print('Errors{}'.format(err))
self.send_velocities(v, w)
diff = rospy.get_time() - now
rospy.sleep(max_t/len_t + 0.0058)
self.appov.append(v)
self.appow.append(w)
self.appox.append(self.q[0])
self.appoy.append(self.q[1])
self.appoteta.append(self.q[2])
# self.appopsi.append(math.atan(w*2.85/v))
else:
pass
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
self.plot_wmr()
def plot_wmr(self):
plot1 = plt.figure(1)
plt.title('path')
plt.plot(self.x_d,self.y_d)
plt.plot(self.appox ,self.appoy )
plt.plot()
plt.xlabel('x')
plt.ylabel('y')
plot2 = plt.figure(2)
plt.title('velocità')
plt.plot(self.t,self.v_d)
plt.plot(self.t,self.appov)
plt.xlabel('time')
plt.ylabel('Velocità lineare')
plot3 = plt.figure(3)
plt.plot(self.t,self.w_d)
plt.plot(self.t,self.appow)
plt.xlabel('time')
plt.ylabel('Velocità angolare ')
plot4 = plt.figure(4)
plt.plot(self.t,self.theta_d)
plt.plot(self.t,self.appoteta)
plt.xlabel('time')
plt.ylabel('teta ')
# plot5 = plt.figure(5)
# plt.plot(self.t,self.psi)
# plt.plot(self.t,self.appopsi)
# plt.xlabel('time')
# plt.ylabel('psi')
plt.show()
def send_velocities(self, v, w, theta=None):
twist_msg = Twist() # Creating a new message to send to the robot
twist_msg.linear.x = v
twist_msg.angular.z = -w
self.twist_pub.publish(twist_msg)
def to_point(self):
toltheta=0.2
tol=0.05
vel=2
q_i = self.get_pose()
if np.pi/2-toltheta<=q_i[2]<=toltheta+np.pi/2 or -np.pi/2-toltheta<=q_i[2]<=toltheta-np.pi/2:
while q_i[0]<=self.A_park[0][0]-tol or q_i[0]>=self.A_park[0][0]+tol:
q_i = self.get_pose()
self.send_velocities(vel,0,0)
print("x_odom:{} and x_A:{}".format(q_i[0], | if(self.trajectory == "parallel_parking" ):
(x, y, theta) = self.get_pose() #NB: i punti x e y sono sull'asse posteriore, non è il centro della macchina
else:
(a, b, theta) = self.get_pose() #prendo solo theta
x=self.data_pose[0]
y=self.data_pose[1]
#compute error
e1 = (self.x_d[T] - x) * np.cos(theta) + (self.y_d[T] - y ) * np.sin(theta)
e2 = -(self.x_d[T] - x) * np.sin(theta) + (self.y_d[T] - y ) * np.cos(theta)
# theta (derivante dall'odometria) quando va oltre 3,14 si inverte di segno (vede il 3,14 come -3.14 e va verso 0 come negativo)
e3 = self.theta_d[T] - theta if len(self.theta_d) else 0
if e3>np.pi :
e3-=2*np.pi
elif e3<-np.pi:
e3+=2*np.pi
else:
pass
print("x_d:{} and x_odom:{} sample:{}".format(self.x_d[T][0],x,T)) | identifier_body |
trajectory_tracking.py | ari')
data_pose_x = data.pose[indice].position.x
data_pose_y = data.pose[indice].position.y
self.data_pose= np.array([data_pose_x,data_pose_y])
return self.data_pose
def odometryCb(self,msg):
#current robot pose
x = round(msg.pose.pose.position.x,4)
y = round(msg.pose.pose.position.y,4)
theta = round(self.get_angle_pose(msg.pose.pose),4)
y=round(y-1.4*np.cos(theta+np.pi),4) ##spostiamo il punto di riferimento dal centro della macchina al centro dell' asse posteriore
x=round(x+1.4*np.sin(theta+np.pi),4) ##spostiamo il punto di riferimento dal centro della macchina al centro dell' asse posteriore
self.q = np.array([x, y, theta])
return self.q
def get_angle_pose(self, quaternion_pose):
#compute angle from quaternion
#
q = [quaternion_pose.orientation.x,
quaternion_pose.orientation.y,
quaternion_pose.orientation.z,
quaternion_pose.orientation.w]
roll, pitch, yaw = euler_from_quaternion(q)
theta = yaw
tol=0.1
if abs(abs(theta)-abs(self.thetaprec))>2*np.pi-tol and self.thetaprec!=0:
theta=theta+2*np.pi-tol
else:
pass
self.thetaprec=theta
return theta
def trajectory_generation(self, traj,a):
data = rospy.wait_for_message("/gazebo/model_states", ModelStates, timeout=5)
posizione = self.callback(data)
x = round(posizione[0],1)
y = round(posizione[1],1)
tg = Trajectory_generation()
q_i = self.get_pose()
self.trajectory=traj
if(self.trajectory == "parallel_parking" ):
(self.x_d, self.y_d,self.dotx_d,self.doty_d,self.v_d, self.w_d , self.theta_d , self.psi, self.A_park) =tg.parallel_parking_trajectory(q_i, self.t,a)
#self.A_park indica il punto di partenza della manovra di parcheggio, ovvero punto in corrispondenza del parcheggio successivo a quello libero
def get_laser(self):
flag_free=self.a.park()[0] # flag che indica se ha trovato un parcheggio libero
flag_occ=self.a.park()[1] # flag che indica se il parcheggio si trova prima(2), in corrispondenza(1) o dopo(0)
if flag_free==1:
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
return flag_free,flag_occ
def get_pose(self):
#get robot position updated from callback
x = self.q[0]
y = self.q[1]
theta = self.q[2]
return np.array([x, y, theta])
def get_error(self, T,traj):
if(self.trajectory == "parallel_parking" ):
(x, y, theta) = self.get_pose() #NB: i punti x e y sono sull'asse posteriore, non è il centro della macchina
else:
(a, b, theta) = self.get_pose() #prendo solo theta
x=self.data_pose[0]
y=self.data_pose[1]
#compute error
e1 = (self.x_d[T] - x) * np.cos(theta) + (self.y_d[T] - y ) * np.sin(theta)
e2 = -(self.x_d[T] - x) * np.sin(theta) + (self.y_d[T] - y ) * np.cos(theta)
# theta (derivante dall'odometria) quando va oltre 3,14 si inverte di segno (vede il 3,14 come -3.14 e va verso 0 come negativo)
e3 = self.theta_d[T] - theta if len(self.theta_d) else 0
if e3>np.pi :
e3-=2*np.pi
elif e3<-np.pi:
e3+=2*np.pi
else:
pass
print("x_d:{} and x_odom:{} sample:{}".format(self.x_d[T][0],x,T))
print("y_d:{} and y_odom:{} sample:{}".format(self.y_d[T][0],y,T))
print("theta_d:{} and theta_odom:{} sample:{}".format(self.theta_d[T],theta,T))
return np.array([float(e1), float(e2), e3])
def unicicle_Linear_control(self,traj,zeta,a):
rospy.sleep(0.1) # need small time to setup q in callback
max_t = self.t[len(self.t) - 1]
len_t = len(self.t)
self.trajectory=traj
if(self.trajectory == "parallel_parking" ):
for i in np.arange(0, len_t):
now = rospy.get_time()
err = self.get_error(i,self.trajectory)
if round(0.03*len_t)<=i<=round(0.87*len_t) : #tra il 3% e l' 87% uso il controllore
(v, w) = Linear_control_law(err, self.v_d[i], self.w_d[i],zeta,a)
else: # utilizziamo nella parte iniziale e finale le desiderate
#(evitiamo gli spike del controllore dovuti a valori prossimi allo zero di v e w)
v=self.v_d[i]
w=self.w_d[i]
# (v, w) = Linear_control_law(err, self.v_d[i], self.w_d[i],zeta,a)
print("theta_d:{} and theta_odom:{} sample:{}".format(self.theta_d[i], self.q[2] , i) )
print("v_d:{} and v:{} sample:{}".format(self.v_d[i], v , i) )
print("w_d:{} and w:{} sample:{}".format(-self.w_d[i], w , i) )
print('Errors{}'.format(err))
self.send_velocities(v, w)
diff = rospy.get_time() - now
rospy.sleep(max_t/len_t + 0.0058)
self.appov.append(v)
self.appow.append(w)
self.appox.append(self.q[0])
self.appoy.append(self.q[1])
self.appoteta.append(self.q[2])
# self.appopsi.append(math.atan(w*2.85/v))
else:
pass
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
self.plot_wmr()
def plot_wmr(self):
plot1 = plt.figure(1)
plt.title('path')
plt.plot(self.x_d,self.y_d)
plt.plot(self.appox ,self.appoy )
plt.plot()
plt.xlabel('x')
plt.ylabel('y')
plot2 = plt.figure(2)
plt.title('velocità')
plt.plot(self.t,self.v_d)
plt.plot(self.t,self.appov)
plt.xlabel('time')
plt.ylabel('Velocità lineare')
plot3 = plt.figure(3)
plt.plot(self.t,self.w_d)
plt.plot(self.t,self.appow)
plt.xlabel('time')
plt.ylabel('Velocità angolare ')
plot4 = plt.figure(4)
plt.plot(self.t,self.theta_d)
plt.plot(self.t,self.appoteta)
plt.xlabel('time')
plt.ylabel('teta ')
# plot5 = plt.figure(5)
# plt.plot(self.t,self.psi)
# plt.plot(self.t,self.appopsi)
# plt.xlabel('time')
# plt.ylabel('psi')
plt.show()
def send_velocities(self, v, w, theta=None):
twist_msg = Twist() # Creating a new message to send to the robot
twist_msg.linear.x = v
twist_msg.angular.z = -w
self.twist_pub.publish(twist_msg)
def to_p | f):
toltheta=0.2
tol=0.05
vel=2
q_i = self.get_pose()
if np.pi/2-toltheta<=q_i[2]<=toltheta+np.pi/2 or -np.pi/2-toltheta<=q_i[2]<=toltheta-np.pi/2:
while q_i[0]<=self.A_park[0][0]-tol or q_i[0]>=self.A_park[0][0]+tol:
q_i = self.get_pose()
self.send_velocities(vel,0,0)
print("x_odom:{} and x_A:{}".format(q_i[0], | oint(sel | identifier_name |
trajectory_tracking.py | errari')
data_pose_x = data.pose[indice].position.x
data_pose_y = data.pose[indice].position.y
self.data_pose= np.array([data_pose_x,data_pose_y])
return self.data_pose
def odometryCb(self,msg):
#current robot pose
x = round(msg.pose.pose.position.x,4)
y = round(msg.pose.pose.position.y,4)
theta = round(self.get_angle_pose(msg.pose.pose),4)
y=round(y-1.4*np.cos(theta+np.pi),4) ##spostiamo il punto di riferimento dal centro della macchina al centro dell' asse posteriore
x=round(x+1.4*np.sin(theta+np.pi),4) ##spostiamo il punto di riferimento dal centro della macchina al centro dell' asse posteriore
self.q = np.array([x, y, theta])
return self.q
def get_angle_pose(self, quaternion_pose):
#compute angle from quaternion
#
q = [quaternion_pose.orientation.x,
quaternion_pose.orientation.y,
quaternion_pose.orientation.z,
quaternion_pose.orientation.w]
roll, pitch, yaw = euler_from_quaternion(q)
theta = yaw
tol=0.1
if abs(abs(theta)-abs(self.thetaprec))>2*np.pi-tol and self.thetaprec!=0:
theta=theta+2*np.pi-tol
else:
pass
self.thetaprec=theta
return theta
def trajectory_generation(self, traj,a): | y = round(posizione[1],1)
tg = Trajectory_generation()
q_i = self.get_pose()
self.trajectory=traj
if(self.trajectory == "parallel_parking" ):
(self.x_d, self.y_d,self.dotx_d,self.doty_d,self.v_d, self.w_d , self.theta_d , self.psi, self.A_park) =tg.parallel_parking_trajectory(q_i, self.t,a)
#self.A_park indica il punto di partenza della manovra di parcheggio, ovvero punto in corrispondenza del parcheggio successivo a quello libero
def get_laser(self):
flag_free=self.a.park()[0] # flag che indica se ha trovato un parcheggio libero
flag_occ=self.a.park()[1] # flag che indica se il parcheggio si trova prima(2), in corrispondenza(1) o dopo(0)
if flag_free==1:
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
return flag_free,flag_occ
def get_pose(self):
#get robot position updated from callback
x = self.q[0]
y = self.q[1]
theta = self.q[2]
return np.array([x, y, theta])
def get_error(self, T,traj):
if(self.trajectory == "parallel_parking" ):
(x, y, theta) = self.get_pose() #NB: i punti x e y sono sull'asse posteriore, non è il centro della macchina
else:
(a, b, theta) = self.get_pose() #prendo solo theta
x=self.data_pose[0]
y=self.data_pose[1]
#compute error
e1 = (self.x_d[T] - x) * np.cos(theta) + (self.y_d[T] - y ) * np.sin(theta)
e2 = -(self.x_d[T] - x) * np.sin(theta) + (self.y_d[T] - y ) * np.cos(theta)
# theta (derivante dall'odometria) quando va oltre 3,14 si inverte di segno (vede il 3,14 come -3.14 e va verso 0 come negativo)
e3 = self.theta_d[T] - theta if len(self.theta_d) else 0
if e3>np.pi :
e3-=2*np.pi
elif e3<-np.pi:
e3+=2*np.pi
else:
pass
print("x_d:{} and x_odom:{} sample:{}".format(self.x_d[T][0],x,T))
print("y_d:{} and y_odom:{} sample:{}".format(self.y_d[T][0],y,T))
print("theta_d:{} and theta_odom:{} sample:{}".format(self.theta_d[T],theta,T))
return np.array([float(e1), float(e2), e3])
def unicicle_Linear_control(self,traj,zeta,a):
rospy.sleep(0.1) # need small time to setup q in callback
max_t = self.t[len(self.t) - 1]
len_t = len(self.t)
self.trajectory=traj
if(self.trajectory == "parallel_parking" ):
for i in np.arange(0, len_t):
now = rospy.get_time()
err = self.get_error(i,self.trajectory)
if round(0.03*len_t)<=i<=round(0.87*len_t) : #tra il 3% e l' 87% uso il controllore
(v, w) = Linear_control_law(err, self.v_d[i], self.w_d[i],zeta,a)
else: # utilizziamo nella parte iniziale e finale le desiderate
#(evitiamo gli spike del controllore dovuti a valori prossimi allo zero di v e w)
v=self.v_d[i]
w=self.w_d[i]
# (v, w) = Linear_control_law(err, self.v_d[i], self.w_d[i],zeta,a)
print("theta_d:{} and theta_odom:{} sample:{}".format(self.theta_d[i], self.q[2] , i) )
print("v_d:{} and v:{} sample:{}".format(self.v_d[i], v , i) )
print("w_d:{} and w:{} sample:{}".format(-self.w_d[i], w , i) )
print('Errors{}'.format(err))
self.send_velocities(v, w)
diff = rospy.get_time() - now
rospy.sleep(max_t/len_t + 0.0058)
self.appov.append(v)
self.appow.append(w)
self.appox.append(self.q[0])
self.appoy.append(self.q[1])
self.appoteta.append(self.q[2])
# self.appopsi.append(math.atan(w*2.85/v))
else:
pass
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
self.plot_wmr()
def plot_wmr(self):
plot1 = plt.figure(1)
plt.title('path')
plt.plot(self.x_d,self.y_d)
plt.plot(self.appox ,self.appoy )
plt.plot()
plt.xlabel('x')
plt.ylabel('y')
plot2 = plt.figure(2)
plt.title('velocità')
plt.plot(self.t,self.v_d)
plt.plot(self.t,self.appov)
plt.xlabel('time')
plt.ylabel('Velocità lineare')
plot3 = plt.figure(3)
plt.plot(self.t,self.w_d)
plt.plot(self.t,self.appow)
plt.xlabel('time')
plt.ylabel('Velocità angolare ')
plot4 = plt.figure(4)
plt.plot(self.t,self.theta_d)
plt.plot(self.t,self.appoteta)
plt.xlabel('time')
plt.ylabel('teta ')
# plot5 = plt.figure(5)
# plt.plot(self.t,self.psi)
# plt.plot(self.t,self.appopsi)
# plt.xlabel('time')
# plt.ylabel('psi')
plt.show()
def send_velocities(self, v, w, theta=None):
twist_msg = Twist() # Creating a new message to send to the robot
twist_msg.linear.x = v
twist_msg.angular.z = -w
self.twist_pub.publish(twist_msg)
def to_point(self):
toltheta=0.2
tol=0.05
vel=2
q_i = self.get_pose()
if np.pi/2-toltheta<=q_i[2]<=toltheta+np.pi/2 or -np.pi/2-toltheta<=q_i[2]<=toltheta-np.pi/2:
while q_i[0]<=self.A_park[0][0]-tol or q_i[0]>=self.A_park[0][0]+tol:
q_i = self.get_pose()
self.send_velocities(vel,0,0)
print("x_odom:{} and x_A:{}".format(q_i[0], |
data = rospy.wait_for_message("/gazebo/model_states", ModelStates, timeout=5)
posizione = self.callback(data)
x = round(posizione[0],1) | random_line_split |
trajectory_tracking.py | 1],1)
tg = Trajectory_generation()
q_i = self.get_pose()
self.trajectory=traj
if(self.trajectory == "parallel_parking" ):
(self.x_d, self.y_d,self.dotx_d,self.doty_d,self.v_d, self.w_d , self.theta_d , self.psi, self.A_park) =tg.parallel_parking_trajectory(q_i, self.t,a)
#self.A_park indica il punto di partenza della manovra di parcheggio, ovvero punto in corrispondenza del parcheggio successivo a quello libero
def get_laser(self):
flag_free=self.a.park()[0] # flag che indica se ha trovato un parcheggio libero
flag_occ=self.a.park()[1] # flag che indica se il parcheggio si trova prima(2), in corrispondenza(1) o dopo(0)
if flag_free==1:
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
return flag_free,flag_occ
def get_pose(self):
#get robot position updated from callback
x = self.q[0]
y = self.q[1]
theta = self.q[2]
return np.array([x, y, theta])
def get_error(self, T,traj):
if(self.trajectory == "parallel_parking" ):
(x, y, theta) = self.get_pose() #NB: i punti x e y sono sull'asse posteriore, non è il centro della macchina
else:
(a, b, theta) = self.get_pose() #prendo solo theta
x=self.data_pose[0]
y=self.data_pose[1]
#compute error
e1 = (self.x_d[T] - x) * np.cos(theta) + (self.y_d[T] - y ) * np.sin(theta)
e2 = -(self.x_d[T] - x) * np.sin(theta) + (self.y_d[T] - y ) * np.cos(theta)
# theta (derivante dall'odometria) quando va oltre 3,14 si inverte di segno (vede il 3,14 come -3.14 e va verso 0 come negativo)
e3 = self.theta_d[T] - theta if len(self.theta_d) else 0
if e3>np.pi :
e3-=2*np.pi
elif e3<-np.pi:
e3+=2*np.pi
else:
pass
print("x_d:{} and x_odom:{} sample:{}".format(self.x_d[T][0],x,T))
print("y_d:{} and y_odom:{} sample:{}".format(self.y_d[T][0],y,T))
print("theta_d:{} and theta_odom:{} sample:{}".format(self.theta_d[T],theta,T))
return np.array([float(e1), float(e2), e3])
def unicicle_Linear_control(self,traj,zeta,a):
rospy.sleep(0.1) # need small time to setup q in callback
max_t = self.t[len(self.t) - 1]
len_t = len(self.t)
self.trajectory=traj
if(self.trajectory == "parallel_parking" ):
for i in np.arange(0, len_t):
now = rospy.get_time()
err = self.get_error(i,self.trajectory)
if round(0.03*len_t)<=i<=round(0.87*len_t) : #tra il 3% e l' 87% uso il controllore
(v, w) = Linear_control_law(err, self.v_d[i], self.w_d[i],zeta,a)
else: # utilizziamo nella parte iniziale e finale le desiderate
#(evitiamo gli spike del controllore dovuti a valori prossimi allo zero di v e w)
v=self.v_d[i]
w=self.w_d[i]
# (v, w) = Linear_control_law(err, self.v_d[i], self.w_d[i],zeta,a)
print("theta_d:{} and theta_odom:{} sample:{}".format(self.theta_d[i], self.q[2] , i) )
print("v_d:{} and v:{} sample:{}".format(self.v_d[i], v , i) )
print("w_d:{} and w:{} sample:{}".format(-self.w_d[i], w , i) )
print('Errors{}'.format(err))
self.send_velocities(v, w)
diff = rospy.get_time() - now
rospy.sleep(max_t/len_t + 0.0058)
self.appov.append(v)
self.appow.append(w)
self.appox.append(self.q[0])
self.appoy.append(self.q[1])
self.appoteta.append(self.q[2])
# self.appopsi.append(math.atan(w*2.85/v))
else:
pass
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
self.plot_wmr()
def plot_wmr(self):
plot1 = plt.figure(1)
plt.title('path')
plt.plot(self.x_d,self.y_d)
plt.plot(self.appox ,self.appoy )
plt.plot()
plt.xlabel('x')
plt.ylabel('y')
plot2 = plt.figure(2)
plt.title('velocità')
plt.plot(self.t,self.v_d)
plt.plot(self.t,self.appov)
plt.xlabel('time')
plt.ylabel('Velocità lineare')
plot3 = plt.figure(3)
plt.plot(self.t,self.w_d)
plt.plot(self.t,self.appow)
plt.xlabel('time')
plt.ylabel('Velocità angolare ')
plot4 = plt.figure(4)
plt.plot(self.t,self.theta_d)
plt.plot(self.t,self.appoteta)
plt.xlabel('time')
plt.ylabel('teta ')
# plot5 = plt.figure(5)
# plt.plot(self.t,self.psi)
# plt.plot(self.t,self.appopsi)
# plt.xlabel('time')
# plt.ylabel('psi')
plt.show()
def send_velocities(self, v, w, theta=None):
twist_msg = Twist() # Creating a new message to send to the robot
twist_msg.linear.x = v
twist_msg.angular.z = -w
self.twist_pub.publish(twist_msg)
def to_point(self):
toltheta=0.2
tol=0.05
vel=2
q_i = self.get_pose()
if np.pi/2-toltheta<=q_i[2]<=toltheta+np.pi/2 or -np.pi/2-toltheta<=q_i[2]<=toltheta-np.pi/2:
while q_i[0]<=self.A_park[0][0]-tol or q_i[0]>=self.A_park[0][0]+tol:
q_i = self.get_pose()
self.send_velocities(vel,0,0)
print("x_odom:{} and x_A:{}".format(q_i[0], round(self.A_park[0][0],4)))
elif -toltheta<=q_i[2]<=toltheta or np.pi-toltheta<=q_i[2]<=toltheta+np.pi:
while q_i[1]<=self.A_park[1][0]-tol or q_i[1]>=self.A_park[1][0]+tol:
q_i = self.get_pose()
self.send_velocities(vel,0,0)
print("y_odom:{} and y_A:{}".format(q_i[1], round(self.A_park[1][0],4)))
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
self.send_velocities(0,0,0)
print("STOP")
print("Asse posteriore: x:{} and y:{}. Punto A: x:{} and y:{}".format(self.q[0], self.q[1],self.A_park[0][0],self.A_park[1][0]))
if __name__ == "__main__":
yaml_package_name = rospy.get_param('~yaml_package_name', 'object_spawner')
yaml_relative_path = rospy.get_param('~yaml_relative_path', '/config/parcheggi2.yaml')
m = g.parse_yaml(yaml_package_name,yaml_relative_path)
tt=Trajectory_tracking()
tt.t = np.linspace(0, 5, 1500)
trajectory = "parallel_parking"
toltheta=0.1
if np.pi-toltheta<=tt.q[2]<=np.pi+toltheta or -toltheta<=tt.q[2]<=toltheta:
a=tt | .data_pose[1] #estraggo y odometria
| conditional_block |
|
srvcenter.go | func (self *Result) Error() string {
b, _ := json.Marshal(self)
return string(b)
}
type ServiceConfig struct {
MaxNrConns int
MaxNrUsers int
MaxNrConnsPerUser int
MsgCache msgcache.Cache
LoginHandler evthandler.LoginHandler
LogoutHandler evthandler.LogoutHandler
MessageHandler evthandler.MessageHandler
ForwardRequestHandler evthandler.ForwardRequestHandler
ErrorHandler evthandler.ErrorHandler
// Push related web hooks
SubscribeHandler evthandler.SubscribeHandler
UnsubscribeHandler evthandler.UnsubscribeHandler
PushHandler evthandler.PushHandler
PushService push.Push
}
type writeMessageRequest struct {
user string
msg *proto.Message
ttl time.Duration
extra map[string]string
resChan chan<- []*Result
}
type serviceCenter struct {
serviceName string
config *ServiceConfig
fwdChan chan<- *server.ForwardRequest
writeReqChan chan *writeMessageRequest
connIn chan *eventConnIn
connLeave chan *eventConnLeave
subReqChan chan *server.SubscribeRequest
pushServiceLock sync.RWMutex
}
var ErrTooManyConns = errors.New("too many connections")
var ErrInvalidConnType = errors.New("invalid connection type")
func (self *serviceCenter) ReceiveForward(fwdreq *server.ForwardRequest) {
shouldFwd := false
if self.config != nil {
if self.config.ForwardRequestHandler != nil {
shouldFwd = self.config.ForwardRequestHandler.ShouldForward(fwdreq)
maxttl := self.config.ForwardRequestHandler.MaxTTL()
if fwdreq.TTL < 1*time.Second || fwdreq.TTL > maxttl {
fwdreq.TTL = maxttl
}
}
}
if !shouldFwd {
return
}
receiver := fwdreq.Receiver
extra := getPushInfo(fwdreq.Message, nil, true)
self.SendMessage(receiver, fwdreq.Message, extra, fwdreq.TTL)
}
func getPushInfo(msg *proto.Message, extra map[string]string, fwd bool) map[string]string {
if extra == nil {
extra = make(map[string]string, len(msg.Header)+3)
}
if fwd {
for k, v := range msg.Header {
if strings.HasPrefix(k, "notif.") {
if strings.HasPrefix(k, "notif.uniqush.") {
// forward message should not contain reserved fields
continue
}
extra[k] = v
delete(msg.Header, k)
}
}
extra["uniqush.sender"] = msg.Sender
extra["uniqush.sender-service"] = msg.SenderService
}
if msg.Header != nil {
if title, ok := msg.Header["title"]; ok {
if _, ok = extra["notif.msg"]; !ok {
extra["notif.msg"] = title
}
}
}
extra["notif.uniqush.msgsize"] = fmt.Sprintf("%v", msg.Size())
return extra
}
func (self *serviceCenter) shouldPush(service, username string, msg *proto.Message, extra map[string]string, fwd bool) bool {
if self.config != nil {
if self.config.PushHandler != nil {
info := getPushInfo(msg, extra, fwd)
return self.config.PushHandler.ShouldPush(service, username, info)
}
}
return false
}
func (self *serviceCenter) subscribe(req *server.SubscribeRequest) {
if req == nil {
return
}
if self.config != nil {
if self.config.PushService != nil {
if req.Subscribe {
self.config.PushService.Subscribe(req.Service, req.Username, req.Params)
} else {
self.config.PushService.Unsubscribe(req.Service, req.Username, req.Params)
}
}
}
}
func (self *serviceCenter) nrDeliveryPoints(service, username string) int {
n := 0
if self.config != nil {
if self.config.PushService != nil {
n = self.config.PushService.NrDeliveryPoints(service, username)
}
}
return n
}
func (self *serviceCenter) pushNotif(service, username string, msg *proto.Message, extra map[string]string, msgIds []string, fwd bool) {
if self.config != nil {
if self.config.PushService != nil {
info := getPushInfo(msg, extra, fwd)
err := self.config.PushService.Push(service, username, info, msgIds)
if err != nil {
self.reportError(service, username, "", "", err)
}
}
}
}
func (self *serviceCenter) reportError(service, username, connId, addr string, err error) {
if self.config != nil {
if self.config.ErrorHandler != nil {
go self.config.ErrorHandler.OnError(service, username, connId, addr, err)
}
}
}
func (self *serviceCenter) reportLogin(service, username, connId, addr string) {
if self.config != nil {
if self.config.LoginHandler != nil {
go self.config.LoginHandler.OnLogin(service, username, connId, addr)
}
}
}
func (self *serviceCenter) reportMessage(connId string, msg *proto.Message) {
if self.config != nil {
if self.config.MessageHandler != nil {
go self.config.MessageHandler.OnMessage(connId, msg)
}
}
}
func (self *serviceCenter) reportLogout(service, username, connId, addr string, err error) {
if self.config != nil {
if self.config.LogoutHandler != nil {
go self.config.LogoutHandler.OnLogout(service, username, connId, addr, err)
}
}
}
func (self *serviceCenter) cacheMessage(service, username string, msg *proto.Message, ttl time.Duration) (id string, err error) {
if self.config != nil {
if self.config.MsgCache != nil {
id, err = self.config.MsgCache.CacheMessage(service, username, msg, ttl)
}
}
return
}
type connWriteErr struct {
conn server.Conn
err error
}
func (self *serviceCenter) | (maxNrConns, maxNrConnsPerUser, maxNrUsers int) {
connMap := newTreeBasedConnMap()
nrConns := 0
for {
select {
case connInEvt := <-self.connIn:
if maxNrConns > 0 && nrConns >= maxNrConns {
if connInEvt.errChan != nil {
connInEvt.errChan <- ErrTooManyConns
}
continue
}
err := connMap.AddConn(connInEvt.conn, maxNrConnsPerUser, maxNrUsers)
if err != nil {
if connInEvt.errChan != nil {
connInEvt.errChan <- err
}
continue
}
nrConns++
if connInEvt.errChan != nil {
connInEvt.errChan <- nil
}
case leaveEvt := <-self.connLeave:
deleted := connMap.DelConn(leaveEvt.conn)
fmt.Printf("delete a connection %v under user %v; deleted: %v\n", leaveEvt.conn.UniqId(), leaveEvt.conn.Username(), deleted)
leaveEvt.conn.Close()
if deleted {
nrConns--
conn := leaveEvt.conn
self.reportLogout(conn.Service(), conn.Username(), conn.UniqId(), conn.RemoteAddr().String(), leaveEvt.err)
}
case subreq := <-self.subReqChan:
self.pushServiceLock.Lock()
self.subscribe(subreq)
self.pushServiceLock.Unlock()
case wreq := <-self.writeReqChan:
conns := connMap.GetConn(wreq.user)
res := make([]*Result, 0, len(conns))
errConns := make([]*connWriteErr, 0, len(conns))
n := 0
mid, err := self.cacheMessage(self.serviceName, wreq.user, wreq.msg, wreq.ttl)
if err != nil {
self.reportError(self.serviceName, wreq.user, "", "", err)
continue
}
for _, conn := range conns {
if conn == nil {
continue
}
var err error
sconn, ok := conn.(server.Conn)
if !ok {
continue
}
err = sconn.SendMessage(wreq.msg, wreq.extra, wreq.ttl, mid)
if err != nil {
errConns = append(errConns, &connWriteErr{sconn, err})
res = append(res, &Result{err, sconn.UniqId(), sconn.Visible()})
self.reportError(sconn.Service(), sconn.Username(), sconn.UniqId(), sconn.RemoteAddr().String(), err)
continue
} else {
res = append(res, &Result{nil, sconn.UniqId(), sconn.Visible()})
}
if sconn.Visible() {
n++
}
}
if n == 0 {
msg := wreq.msg
extra := wreq.extra
username := wreq.user
service := self.serviceName
fwd := false
if len(msg.Sender) > 0 && len(msg.SenderService) > 0 {
if msg.Sender != | process | identifier_name |
srvcenter.go | (self *Result) Error() string {
b, _ := json.Marshal(self)
return string(b)
}
type ServiceConfig struct {
MaxNrConns int
MaxNrUsers int
MaxNrConnsPerUser int
MsgCache msgcache.Cache
LoginHandler evthandler.LoginHandler
LogoutHandler evthandler.LogoutHandler
MessageHandler evthandler.MessageHandler
ForwardRequestHandler evthandler.ForwardRequestHandler
ErrorHandler evthandler.ErrorHandler
// Push related web hooks
SubscribeHandler evthandler.SubscribeHandler
UnsubscribeHandler evthandler.UnsubscribeHandler
PushHandler evthandler.PushHandler
PushService push.Push
}
type writeMessageRequest struct {
user string
msg *proto.Message
ttl time.Duration
extra map[string]string
resChan chan<- []*Result
}
type serviceCenter struct {
serviceName string
config *ServiceConfig
fwdChan chan<- *server.ForwardRequest
writeReqChan chan *writeMessageRequest
connIn chan *eventConnIn
connLeave chan *eventConnLeave
subReqChan chan *server.SubscribeRequest
pushServiceLock sync.RWMutex
}
var ErrTooManyConns = errors.New("too many connections")
var ErrInvalidConnType = errors.New("invalid connection type")
func (self *serviceCenter) ReceiveForward(fwdreq *server.ForwardRequest) |
func getPushInfo(msg *proto.Message, extra map[string]string, fwd bool) map[string]string {
if extra == nil {
extra = make(map[string]string, len(msg.Header)+3)
}
if fwd {
for k, v := range msg.Header {
if strings.HasPrefix(k, "notif.") {
if strings.HasPrefix(k, "notif.uniqush.") {
// forward message should not contain reserved fields
continue
}
extra[k] = v
delete(msg.Header, k)
}
}
extra["uniqush.sender"] = msg.Sender
extra["uniqush.sender-service"] = msg.SenderService
}
if msg.Header != nil {
if title, ok := msg.Header["title"]; ok {
if _, ok = extra["notif.msg"]; !ok {
extra["notif.msg"] = title
}
}
}
extra["notif.uniqush.msgsize"] = fmt.Sprintf("%v", msg.Size())
return extra
}
func (self *serviceCenter) shouldPush(service, username string, msg *proto.Message, extra map[string]string, fwd bool) bool {
if self.config != nil {
if self.config.PushHandler != nil {
info := getPushInfo(msg, extra, fwd)
return self.config.PushHandler.ShouldPush(service, username, info)
}
}
return false
}
func (self *serviceCenter) subscribe(req *server.SubscribeRequest) {
if req == nil {
return
}
if self.config != nil {
if self.config.PushService != nil {
if req.Subscribe {
self.config.PushService.Subscribe(req.Service, req.Username, req.Params)
} else {
self.config.PushService.Unsubscribe(req.Service, req.Username, req.Params)
}
}
}
}
func (self *serviceCenter) nrDeliveryPoints(service, username string) int {
n := 0
if self.config != nil {
if self.config.PushService != nil {
n = self.config.PushService.NrDeliveryPoints(service, username)
}
}
return n
}
func (self *serviceCenter) pushNotif(service, username string, msg *proto.Message, extra map[string]string, msgIds []string, fwd bool) {
if self.config != nil {
if self.config.PushService != nil {
info := getPushInfo(msg, extra, fwd)
err := self.config.PushService.Push(service, username, info, msgIds)
if err != nil {
self.reportError(service, username, "", "", err)
}
}
}
}
func (self *serviceCenter) reportError(service, username, connId, addr string, err error) {
if self.config != nil {
if self.config.ErrorHandler != nil {
go self.config.ErrorHandler.OnError(service, username, connId, addr, err)
}
}
}
func (self *serviceCenter) reportLogin(service, username, connId, addr string) {
if self.config != nil {
if self.config.LoginHandler != nil {
go self.config.LoginHandler.OnLogin(service, username, connId, addr)
}
}
}
func (self *serviceCenter) reportMessage(connId string, msg *proto.Message) {
if self.config != nil {
if self.config.MessageHandler != nil {
go self.config.MessageHandler.OnMessage(connId, msg)
}
}
}
func (self *serviceCenter) reportLogout(service, username, connId, addr string, err error) {
if self.config != nil {
if self.config.LogoutHandler != nil {
go self.config.LogoutHandler.OnLogout(service, username, connId, addr, err)
}
}
}
func (self *serviceCenter) cacheMessage(service, username string, msg *proto.Message, ttl time.Duration) (id string, err error) {
if self.config != nil {
if self.config.MsgCache != nil {
id, err = self.config.MsgCache.CacheMessage(service, username, msg, ttl)
}
}
return
}
type connWriteErr struct {
conn server.Conn
err error
}
func (self *serviceCenter) process(maxNrConns, maxNrConnsPerUser, maxNrUsers int) {
connMap := newTreeBasedConnMap()
nrConns := 0
for {
select {
case connInEvt := <-self.connIn:
if maxNrConns > 0 && nrConns >= maxNrConns {
if connInEvt.errChan != nil {
connInEvt.errChan <- ErrTooManyConns
}
continue
}
err := connMap.AddConn(connInEvt.conn, maxNrConnsPerUser, maxNrUsers)
if err != nil {
if connInEvt.errChan != nil {
connInEvt.errChan <- err
}
continue
}
nrConns++
if connInEvt.errChan != nil {
connInEvt.errChan <- nil
}
case leaveEvt := <-self.connLeave:
deleted := connMap.DelConn(leaveEvt.conn)
fmt.Printf("delete a connection %v under user %v; deleted: %v\n", leaveEvt.conn.UniqId(), leaveEvt.conn.Username(), deleted)
leaveEvt.conn.Close()
if deleted {
nrConns--
conn := leaveEvt.conn
self.reportLogout(conn.Service(), conn.Username(), conn.UniqId(), conn.RemoteAddr().String(), leaveEvt.err)
}
case subreq := <-self.subReqChan:
self.pushServiceLock.Lock()
self.subscribe(subreq)
self.pushServiceLock.Unlock()
case wreq := <-self.writeReqChan:
conns := connMap.GetConn(wreq.user)
res := make([]*Result, 0, len(conns))
errConns := make([]*connWriteErr, 0, len(conns))
n := 0
mid, err := self.cacheMessage(self.serviceName, wreq.user, wreq.msg, wreq.ttl)
if err != nil {
self.reportError(self.serviceName, wreq.user, "", "", err)
continue
}
for _, conn := range conns {
if conn == nil {
continue
}
var err error
sconn, ok := conn.(server.Conn)
if !ok {
continue
}
err = sconn.SendMessage(wreq.msg, wreq.extra, wreq.ttl, mid)
if err != nil {
errConns = append(errConns, &connWriteErr{sconn, err})
res = append(res, &Result{err, sconn.UniqId(), sconn.Visible()})
self.reportError(sconn.Service(), sconn.Username(), sconn.UniqId(), sconn.RemoteAddr().String(), err)
continue
} else {
res = append(res, &Result{nil, sconn.UniqId(), sconn.Visible()})
}
if sconn.Visible() {
n++
}
}
if n == 0 {
msg := wreq.msg
extra := wreq.extra
username := wreq.user
service := self.serviceName
fwd := false
if len(msg.Sender) > 0 && len(msg.SenderService) > 0 {
if msg.Sender != | {
shouldFwd := false
if self.config != nil {
if self.config.ForwardRequestHandler != nil {
shouldFwd = self.config.ForwardRequestHandler.ShouldForward(fwdreq)
maxttl := self.config.ForwardRequestHandler.MaxTTL()
if fwdreq.TTL < 1*time.Second || fwdreq.TTL > maxttl {
fwdreq.TTL = maxttl
}
}
}
if !shouldFwd {
return
}
receiver := fwdreq.Receiver
extra := getPushInfo(fwdreq.Message, nil, true)
self.SendMessage(receiver, fwdreq.Message, extra, fwdreq.TTL)
} | identifier_body |
srvcenter.go | ForwardRequestHandler evthandler.ForwardRequestHandler
ErrorHandler evthandler.ErrorHandler
// Push related web hooks
SubscribeHandler evthandler.SubscribeHandler
UnsubscribeHandler evthandler.UnsubscribeHandler
PushHandler evthandler.PushHandler
PushService push.Push
}
type writeMessageRequest struct {
user string
msg *proto.Message
ttl time.Duration
extra map[string]string
resChan chan<- []*Result
}
type serviceCenter struct {
serviceName string
config *ServiceConfig
fwdChan chan<- *server.ForwardRequest
writeReqChan chan *writeMessageRequest
connIn chan *eventConnIn
connLeave chan *eventConnLeave
subReqChan chan *server.SubscribeRequest
pushServiceLock sync.RWMutex
}
var ErrTooManyConns = errors.New("too many connections")
var ErrInvalidConnType = errors.New("invalid connection type")
func (self *serviceCenter) ReceiveForward(fwdreq *server.ForwardRequest) {
shouldFwd := false
if self.config != nil {
if self.config.ForwardRequestHandler != nil {
shouldFwd = self.config.ForwardRequestHandler.ShouldForward(fwdreq)
maxttl := self.config.ForwardRequestHandler.MaxTTL()
if fwdreq.TTL < 1*time.Second || fwdreq.TTL > maxttl {
fwdreq.TTL = maxttl
}
}
}
if !shouldFwd {
return
}
receiver := fwdreq.Receiver
extra := getPushInfo(fwdreq.Message, nil, true)
self.SendMessage(receiver, fwdreq.Message, extra, fwdreq.TTL)
}
func getPushInfo(msg *proto.Message, extra map[string]string, fwd bool) map[string]string {
if extra == nil {
extra = make(map[string]string, len(msg.Header)+3)
}
if fwd {
for k, v := range msg.Header {
if strings.HasPrefix(k, "notif.") {
if strings.HasPrefix(k, "notif.uniqush.") {
// forward message should not contain reserved fields
continue
}
extra[k] = v
delete(msg.Header, k)
}
}
extra["uniqush.sender"] = msg.Sender
extra["uniqush.sender-service"] = msg.SenderService
}
if msg.Header != nil {
if title, ok := msg.Header["title"]; ok {
if _, ok = extra["notif.msg"]; !ok {
extra["notif.msg"] = title
}
}
}
extra["notif.uniqush.msgsize"] = fmt.Sprintf("%v", msg.Size())
return extra
}
func (self *serviceCenter) shouldPush(service, username string, msg *proto.Message, extra map[string]string, fwd bool) bool {
if self.config != nil {
if self.config.PushHandler != nil {
info := getPushInfo(msg, extra, fwd)
return self.config.PushHandler.ShouldPush(service, username, info)
}
}
return false
}
func (self *serviceCenter) subscribe(req *server.SubscribeRequest) {
if req == nil {
return
}
if self.config != nil {
if self.config.PushService != nil {
if req.Subscribe {
self.config.PushService.Subscribe(req.Service, req.Username, req.Params)
} else {
self.config.PushService.Unsubscribe(req.Service, req.Username, req.Params)
}
}
}
}
func (self *serviceCenter) nrDeliveryPoints(service, username string) int {
n := 0
if self.config != nil {
if self.config.PushService != nil {
n = self.config.PushService.NrDeliveryPoints(service, username)
}
}
return n
}
func (self *serviceCenter) pushNotif(service, username string, msg *proto.Message, extra map[string]string, msgIds []string, fwd bool) {
if self.config != nil {
if self.config.PushService != nil {
info := getPushInfo(msg, extra, fwd)
err := self.config.PushService.Push(service, username, info, msgIds)
if err != nil {
self.reportError(service, username, "", "", err)
}
}
}
}
func (self *serviceCenter) reportError(service, username, connId, addr string, err error) {
if self.config != nil {
if self.config.ErrorHandler != nil {
go self.config.ErrorHandler.OnError(service, username, connId, addr, err)
}
}
}
func (self *serviceCenter) reportLogin(service, username, connId, addr string) {
if self.config != nil {
if self.config.LoginHandler != nil {
go self.config.LoginHandler.OnLogin(service, username, connId, addr)
}
}
}
func (self *serviceCenter) reportMessage(connId string, msg *proto.Message) {
if self.config != nil {
if self.config.MessageHandler != nil {
go self.config.MessageHandler.OnMessage(connId, msg)
}
}
}
func (self *serviceCenter) reportLogout(service, username, connId, addr string, err error) {
if self.config != nil {
if self.config.LogoutHandler != nil {
go self.config.LogoutHandler.OnLogout(service, username, connId, addr, err)
}
}
}
func (self *serviceCenter) cacheMessage(service, username string, msg *proto.Message, ttl time.Duration) (id string, err error) {
if self.config != nil {
if self.config.MsgCache != nil {
id, err = self.config.MsgCache.CacheMessage(service, username, msg, ttl)
}
}
return
}
type connWriteErr struct {
conn server.Conn
err error
}
func (self *serviceCenter) process(maxNrConns, maxNrConnsPerUser, maxNrUsers int) {
connMap := newTreeBasedConnMap()
nrConns := 0
for {
select {
case connInEvt := <-self.connIn:
if maxNrConns > 0 && nrConns >= maxNrConns {
if connInEvt.errChan != nil {
connInEvt.errChan <- ErrTooManyConns
}
continue
}
err := connMap.AddConn(connInEvt.conn, maxNrConnsPerUser, maxNrUsers)
if err != nil {
if connInEvt.errChan != nil {
connInEvt.errChan <- err
}
continue
}
nrConns++
if connInEvt.errChan != nil {
connInEvt.errChan <- nil
}
case leaveEvt := <-self.connLeave:
deleted := connMap.DelConn(leaveEvt.conn)
fmt.Printf("delete a connection %v under user %v; deleted: %v\n", leaveEvt.conn.UniqId(), leaveEvt.conn.Username(), deleted)
leaveEvt.conn.Close()
if deleted {
nrConns--
conn := leaveEvt.conn
self.reportLogout(conn.Service(), conn.Username(), conn.UniqId(), conn.RemoteAddr().String(), leaveEvt.err)
}
case subreq := <-self.subReqChan:
self.pushServiceLock.Lock()
self.subscribe(subreq)
self.pushServiceLock.Unlock()
case wreq := <-self.writeReqChan:
conns := connMap.GetConn(wreq.user)
res := make([]*Result, 0, len(conns))
errConns := make([]*connWriteErr, 0, len(conns))
n := 0
mid, err := self.cacheMessage(self.serviceName, wreq.user, wreq.msg, wreq.ttl)
if err != nil {
self.reportError(self.serviceName, wreq.user, "", "", err)
continue
}
for _, conn := range conns {
if conn == nil {
continue
}
var err error
sconn, ok := conn.(server.Conn)
if !ok {
continue
}
err = sconn.SendMessage(wreq.msg, wreq.extra, wreq.ttl, mid)
if err != nil {
errConns = append(errConns, &connWriteErr{sconn, err})
res = append(res, &Result{err, sconn.UniqId(), sconn.Visible()})
self.reportError(sconn.Service(), sconn.Username(), sconn.UniqId(), sconn.RemoteAddr().String(), err)
continue
} else {
res = append(res, &Result{nil, sconn.UniqId(), sconn.Visible()})
}
if sconn.Visible() {
n++
}
}
if n == 0 | {
msg := wreq.msg
extra := wreq.extra
username := wreq.user
service := self.serviceName
fwd := false
if len(msg.Sender) > 0 && len(msg.SenderService) > 0 {
if msg.Sender != username || msg.SenderService != service {
fwd = true
}
}
go func() {
should := self.shouldPush(service, username, msg, extra, fwd)
if !should {
return
}
self.pushServiceLock.RLock()
defer self.pushServiceLock.RUnlock()
n := self.nrDeliveryPoints(service, username)
if n <= 0 { | conditional_block |
|
srvcenter.go | getPushInfo(fwdreq.Message, nil, true)
self.SendMessage(receiver, fwdreq.Message, extra, fwdreq.TTL)
}
func getPushInfo(msg *proto.Message, extra map[string]string, fwd bool) map[string]string {
if extra == nil {
extra = make(map[string]string, len(msg.Header)+3)
}
if fwd {
for k, v := range msg.Header {
if strings.HasPrefix(k, "notif.") {
if strings.HasPrefix(k, "notif.uniqush.") {
// forward message should not contain reserved fields
continue
}
extra[k] = v
delete(msg.Header, k)
}
}
extra["uniqush.sender"] = msg.Sender
extra["uniqush.sender-service"] = msg.SenderService
}
if msg.Header != nil {
if title, ok := msg.Header["title"]; ok {
if _, ok = extra["notif.msg"]; !ok {
extra["notif.msg"] = title
}
}
}
extra["notif.uniqush.msgsize"] = fmt.Sprintf("%v", msg.Size())
return extra
}
func (self *serviceCenter) shouldPush(service, username string, msg *proto.Message, extra map[string]string, fwd bool) bool {
if self.config != nil {
if self.config.PushHandler != nil {
info := getPushInfo(msg, extra, fwd)
return self.config.PushHandler.ShouldPush(service, username, info)
}
}
return false
}
func (self *serviceCenter) subscribe(req *server.SubscribeRequest) {
if req == nil {
return
}
if self.config != nil {
if self.config.PushService != nil {
if req.Subscribe {
self.config.PushService.Subscribe(req.Service, req.Username, req.Params)
} else {
self.config.PushService.Unsubscribe(req.Service, req.Username, req.Params)
}
}
}
}
func (self *serviceCenter) nrDeliveryPoints(service, username string) int {
n := 0
if self.config != nil {
if self.config.PushService != nil {
n = self.config.PushService.NrDeliveryPoints(service, username)
}
}
return n
}
func (self *serviceCenter) pushNotif(service, username string, msg *proto.Message, extra map[string]string, msgIds []string, fwd bool) {
if self.config != nil {
if self.config.PushService != nil {
info := getPushInfo(msg, extra, fwd)
err := self.config.PushService.Push(service, username, info, msgIds)
if err != nil {
self.reportError(service, username, "", "", err)
}
}
}
}
func (self *serviceCenter) reportError(service, username, connId, addr string, err error) {
if self.config != nil {
if self.config.ErrorHandler != nil {
go self.config.ErrorHandler.OnError(service, username, connId, addr, err)
}
}
}
func (self *serviceCenter) reportLogin(service, username, connId, addr string) {
if self.config != nil {
if self.config.LoginHandler != nil {
go self.config.LoginHandler.OnLogin(service, username, connId, addr)
}
}
}
func (self *serviceCenter) reportMessage(connId string, msg *proto.Message) {
if self.config != nil {
if self.config.MessageHandler != nil {
go self.config.MessageHandler.OnMessage(connId, msg)
}
}
}
func (self *serviceCenter) reportLogout(service, username, connId, addr string, err error) {
if self.config != nil {
if self.config.LogoutHandler != nil {
go self.config.LogoutHandler.OnLogout(service, username, connId, addr, err)
}
}
}
func (self *serviceCenter) cacheMessage(service, username string, msg *proto.Message, ttl time.Duration) (id string, err error) {
if self.config != nil {
if self.config.MsgCache != nil {
id, err = self.config.MsgCache.CacheMessage(service, username, msg, ttl)
}
}
return
}
type connWriteErr struct {
conn server.Conn
err error
}
func (self *serviceCenter) process(maxNrConns, maxNrConnsPerUser, maxNrUsers int) {
connMap := newTreeBasedConnMap()
nrConns := 0
for {
select {
case connInEvt := <-self.connIn:
if maxNrConns > 0 && nrConns >= maxNrConns {
if connInEvt.errChan != nil {
connInEvt.errChan <- ErrTooManyConns
}
continue
}
err := connMap.AddConn(connInEvt.conn, maxNrConnsPerUser, maxNrUsers)
if err != nil {
if connInEvt.errChan != nil {
connInEvt.errChan <- err
}
continue
}
nrConns++
if connInEvt.errChan != nil {
connInEvt.errChan <- nil
}
case leaveEvt := <-self.connLeave:
deleted := connMap.DelConn(leaveEvt.conn)
fmt.Printf("delete a connection %v under user %v; deleted: %v\n", leaveEvt.conn.UniqId(), leaveEvt.conn.Username(), deleted)
leaveEvt.conn.Close()
if deleted {
nrConns--
conn := leaveEvt.conn
self.reportLogout(conn.Service(), conn.Username(), conn.UniqId(), conn.RemoteAddr().String(), leaveEvt.err)
}
case subreq := <-self.subReqChan:
self.pushServiceLock.Lock()
self.subscribe(subreq)
self.pushServiceLock.Unlock()
case wreq := <-self.writeReqChan:
conns := connMap.GetConn(wreq.user)
res := make([]*Result, 0, len(conns))
errConns := make([]*connWriteErr, 0, len(conns))
n := 0
mid, err := self.cacheMessage(self.serviceName, wreq.user, wreq.msg, wreq.ttl)
if err != nil {
self.reportError(self.serviceName, wreq.user, "", "", err)
continue
}
for _, conn := range conns {
if conn == nil {
continue
}
var err error
sconn, ok := conn.(server.Conn)
if !ok {
continue
}
err = sconn.SendMessage(wreq.msg, wreq.extra, wreq.ttl, mid)
if err != nil {
errConns = append(errConns, &connWriteErr{sconn, err})
res = append(res, &Result{err, sconn.UniqId(), sconn.Visible()})
self.reportError(sconn.Service(), sconn.Username(), sconn.UniqId(), sconn.RemoteAddr().String(), err)
continue
} else {
res = append(res, &Result{nil, sconn.UniqId(), sconn.Visible()})
}
if sconn.Visible() {
n++
}
}
if n == 0 {
msg := wreq.msg
extra := wreq.extra
username := wreq.user
service := self.serviceName
fwd := false
if len(msg.Sender) > 0 && len(msg.SenderService) > 0 {
if msg.Sender != username || msg.SenderService != service {
fwd = true
}
}
go func() {
should := self.shouldPush(service, username, msg, extra, fwd)
if !should {
return
}
self.pushServiceLock.RLock()
defer self.pushServiceLock.RUnlock()
n := self.nrDeliveryPoints(service, username)
if n <= 0 {
return
}
msgIds := []string{mid}
self.pushNotif(service, username, msg, extra, msgIds, fwd)
}()
}
if wreq.resChan != nil {
wreq.resChan <- res
}
// close all connections with error:
go func() {
for _, e := range errConns {
fmt.Printf("Need to remove connection %v\n", e.conn.UniqId())
self.connLeave <- &eventConnLeave{conn: e.conn, err: e.err}
}
}()
}
}
}
func (self *serviceCenter) SendMessage(username string, msg *proto.Message, extra map[string]string, ttl time.Duration) []*Result {
req := new(writeMessageRequest)
ch := make(chan []*Result)
req.msg = msg
req.user = username
req.ttl = ttl
req.resChan = ch
req.extra = extra
self.writeReqChan <- req
res := <-ch
return res
}
func (self *serviceCenter) serveConn(conn server.Conn) {
conn.SetForwardRequestChannel(self.fwdChan)
conn.SetSubscribeRequestChan(self.subReqChan)
var err error
defer func() {
self.connLeave <- &eventConnLeave{conn: conn, err: err}
}()
for {
var msg *proto.Message
msg, err = conn.ReadMessage()
if err != nil {
return
}
self.reportMessage(conn.UniqId(), msg) | }
}
func (self *serviceCenter) NewConn(conn server.Conn) error { | random_line_split |
|
wireguard.go | OrPromptImpl(ctx, nth, prompt, true)
}
func orgByArg(cmdCtx *cmdctx.CmdContext) (*api.Organization, error) {
ctx := cmdCtx.Command.Context()
client := cmdCtx.Client.API()
if len(cmdCtx.Args) == 0 {
org, err := selectOrganization(ctx, client, "", nil)
if err != nil {
return nil, err
}
return org, nil
}
return client.FindOrganizationBySlug(ctx, cmdCtx.Args[0])
}
func runWireGuardList(cmdCtx *cmdctx.CmdContext) error {
ctx := cmdCtx.Command.Context()
client := cmdCtx.Client.API()
org, err := orgByArg(cmdCtx)
if err != nil {
return err
}
peers, err := client.GetWireGuardPeers(ctx, org.Slug)
if err != nil {
return err
}
if cmdCtx.OutputJSON() {
cmdCtx.WriteJSON(peers)
return nil
}
table := tablewriter.NewWriter(cmdCtx.Out)
table.SetHeader([]string{
"Name",
"Region",
"Peer IP",
})
for _, peer := range peers {
table.Append([]string{peer.Name, peer.Region, peer.Peerip})
}
table.Render()
return nil
}
func generateWgConf(peer *api.CreatedWireGuardPeer, privkey string, w io.Writer) {
templateStr := `
[Interface]
PrivateKey = {{.Meta.Privkey}}
Address = {{.Peer.Peerip}}/120
DNS = {{.Meta.DNS}}
[Peer]
PublicKey = {{.Peer.Pubkey}}
AllowedIPs = {{.Meta.AllowedIPs}}
Endpoint = {{.Peer.Endpointip}}:51820
PersistentKeepalive = 15
`
data := struct {
Peer *api.CreatedWireGuardPeer
Meta struct {
Privkey string
AllowedIPs string
DNS string
}
}{
Peer: peer,
}
addr := net.ParseIP(peer.Peerip).To16()
for i := 6; i < 16; i++ {
addr[i] = 0
}
// BUG(tqbf): can't stay this way
data.Meta.AllowedIPs = fmt.Sprintf("%s/48", addr)
addr[15] = 3
data.Meta.DNS = fmt.Sprintf("%s", addr)
data.Meta.Privkey = privkey
tmpl := template.Must(template.New("name").Parse(templateStr))
tmpl.Execute(w, &data)
}
func resolveOutputWriter(ctx *cmdctx.CmdContext, idx int, prompt string) (w io.WriteCloser, mustClose bool, err error) {
var (
f *os.File
filename string
)
for {
filename, err = argOrPromptLoop(ctx, idx, prompt, filename)
if err != nil {
return nil, false, err
}
if filename == "" {
fmt.Println("Provide a filename (or 'stdout')")
continue
}
if filename == "stdout" {
return os.Stdout, false, nil
}
f, err = os.OpenFile(filename, os.O_WRONLY|os.O_CREATE|os.O_EXCL, 0600)
if err == nil {
return f, true, nil
}
fmt.Printf("Can't create '%s': %s\n", filename, err)
}
}
func runWireGuardResetPeer(ctx *cmdctx.CmdContext) error {
org, err := orgByArg(ctx)
if err != nil {
return err
}
client := ctx.Client.API()
agentclient, err := agent.Establish(context.Background(), client)
if err != nil {
return err
}
conf, err := agentclient.Reestablish(context.Background(), org.Slug)
if err != nil {
return err
}
fmt.Printf("New WireGuard peer for organization '%s': '%s'\n", org.Slug, conf.WireGuardState.Name)
return nil
}
func runWireGuardCreate(ctx *cmdctx.CmdContext) error {
org, err := orgByArg(ctx)
if err != nil {
return err
}
var region string
var name string
if len(ctx.Args) > 1 && ctx.Args[1] != "" {
region = ctx.Args[1]
}
if len(ctx.Args) > 2 && ctx.Args[2] != "" {
name = ctx.Args[2]
}
state, err := wireguard.Create(ctx.Client.API(), org, region, name)
if err != nil {
return err
}
data := &state.Peer
fmt.Printf(`
!!!! WARNING: Output includes private key. Private keys cannot be recovered !!!!
!!!! after creating the peer; if you lose the key, you'll need to remove !!!!
!!!! and re-add the peering connection. !!!!
`)
w, shouldClose, err := resolveOutputWriter(ctx, 3, "Filename to store WireGuard configuration in, or 'stdout': ")
if err != nil {
return err
}
if shouldClose {
defer w.Close()
}
generateWgConf(data, state.LocalPrivate, w)
if shouldClose {
filename := w.(*os.File).Name()
fmt.Printf("Wrote WireGuard configuration to %s; load in your WireGuard client\n", filename)
}
return nil
}
func runWireGuardRemove(cmdCtx *cmdctx.CmdContext) error {
ctx := cmdCtx.Command.Context()
client := cmdCtx.Client.API()
org, err := orgByArg(cmdCtx)
if err != nil {
return err
}
var name string
if len(cmdCtx.Args) >= 2 {
name = cmdCtx.Args[1]
} else {
name, err = selectWireGuardPeer(ctx, cmdCtx.Client.API(), org.Slug)
if err != nil {
return err
}
}
fmt.Printf("Removing WireGuard peer \"%s\" for organization %s\n", name, org.Slug)
err = client.RemoveWireGuardPeer(ctx, org, name)
if err != nil {
return err
}
fmt.Println("Removed peer.")
return wireguard.PruneInvalidPeers(ctx, cmdCtx.Client.API())
}
func runWireGuardStat(cmdCtx *cmdctx.CmdContext) error {
ctx := cmdCtx.Command.Context()
client := cmdCtx.Client.API()
org, err := orgByArg(cmdCtx)
if err != nil {
return err
}
var name string
if len(cmdCtx.Args) >= 2 {
name = cmdCtx.Args[1]
} else {
name, err = selectWireGuardPeer(ctx, cmdCtx.Client.API(), org.Slug)
if err != nil {
return err
}
}
status, err := client.GetWireGuardPeerStatus(ctx, org.Slug, name)
if err != nil {
return err
}
fmt.Printf("Alive: %+v\n", status.Live)
if status.WgError != "" {
fmt.Printf("Gateway error: %s\n", status.WgError)
}
if !status.Live {
return nil
}
if status.Endpoint != "" {
fmt.Printf("Last Source Address: %s\n", status.Endpoint)
}
ago := ""
if status.SinceAdded != "" {
ago = " (" + status.SinceAdded + " ago)"
}
if status.LastHandshake != "" {
fmt.Printf("Last Handshake At: %s%s\n", status.LastHandshake, ago)
}
ago = ""
if status.SinceHandshake != "" {
ago = " (" + status.SinceHandshake + " ago)"
}
fmt.Printf("Installed On Gateway At: %s%s\n", status.Added, ago)
fmt.Printf("Traffic: rx:%d tx:%d\n", status.Rx, status.Tx)
return nil
}
func runWireGuardTokenList(cmdCtx *cmdctx.CmdContext) error {
ctx := cmdCtx.Command.Context()
client := cmdCtx.Client.API()
org, err := orgByArg(cmdCtx)
if err != nil {
return err
}
tokens, err := client.GetDelegatedWireGuardTokens(ctx, org.Slug)
if err != nil {
return err
}
if cmdCtx.OutputJSON() {
cmdCtx.WriteJSON(tokens)
return nil
}
table := tablewriter.NewWriter(cmdCtx.Out)
table.SetHeader([]string{
"Name",
})
for _, peer := range tokens {
table.Append([]string{peer.Name})
}
table.Render()
return nil
}
func runWireGuardTokenCreate(cmdCtx *cmdctx.CmdContext) error {
ctx := cmdCtx.Command.Context()
client := cmdCtx.Client.API()
org, err := orgByArg(cmdCtx)
if err != nil {
return err
}
name, err := argOrPrompt(cmdCtx, 1, "Memorable name for WireGuard token: ")
if err != nil {
return err
}
data, err := client.CreateDelegatedWireGuardToken(ctx, org, name)
if err != nil |
fmt.Printf(`
!!!! WARNING: Output includes credential information. Credentials cannot !!!!
!!!! be recovered after creation; if you lose the token, you'll need to !!!!
!!!! remove and and re-add it. !!!!
To use a token to create a WireGuard connection, you can use curl:
curl -v --request POST
-H "Authorization: Bearer ${WG_TOKEN}"
-H "Content-Type: application/json"
--data | {
return err
} | conditional_block |
wireguard.go | OrPromptImpl(ctx, nth, prompt, true)
}
func orgByArg(cmdCtx *cmdctx.CmdContext) (*api.Organization, error) {
ctx := cmdCtx.Command.Context()
client := cmdCtx.Client.API()
if len(cmdCtx.Args) == 0 {
org, err := selectOrganization(ctx, client, "", nil)
if err != nil {
return nil, err
}
return org, nil
}
return client.FindOrganizationBySlug(ctx, cmdCtx.Args[0])
}
func runWireGuardList(cmdCtx *cmdctx.CmdContext) error {
ctx := cmdCtx.Command.Context()
client := cmdCtx.Client.API()
org, err := orgByArg(cmdCtx)
if err != nil {
return err
}
peers, err := client.GetWireGuardPeers(ctx, org.Slug)
if err != nil {
return err
}
if cmdCtx.OutputJSON() {
cmdCtx.WriteJSON(peers)
return nil
}
table := tablewriter.NewWriter(cmdCtx.Out)
table.SetHeader([]string{
"Name",
"Region",
"Peer IP",
})
for _, peer := range peers {
table.Append([]string{peer.Name, peer.Region, peer.Peerip})
}
table.Render()
return nil
}
func generateWgConf(peer *api.CreatedWireGuardPeer, privkey string, w io.Writer) {
templateStr := `
[Interface]
PrivateKey = {{.Meta.Privkey}}
Address = {{.Peer.Peerip}}/120
DNS = {{.Meta.DNS}}
[Peer]
PublicKey = {{.Peer.Pubkey}}
AllowedIPs = {{.Meta.AllowedIPs}}
Endpoint = {{.Peer.Endpointip}}:51820
PersistentKeepalive = 15
`
data := struct {
Peer *api.CreatedWireGuardPeer
Meta struct {
Privkey string
AllowedIPs string
DNS string
}
}{
Peer: peer,
}
addr := net.ParseIP(peer.Peerip).To16()
for i := 6; i < 16; i++ {
addr[i] = 0
}
// BUG(tqbf): can't stay this way
data.Meta.AllowedIPs = fmt.Sprintf("%s/48", addr)
addr[15] = 3
data.Meta.DNS = fmt.Sprintf("%s", addr)
data.Meta.Privkey = privkey
tmpl := template.Must(template.New("name").Parse(templateStr))
tmpl.Execute(w, &data)
}
func | (ctx *cmdctx.CmdContext, idx int, prompt string) (w io.WriteCloser, mustClose bool, err error) {
var (
f *os.File
filename string
)
for {
filename, err = argOrPromptLoop(ctx, idx, prompt, filename)
if err != nil {
return nil, false, err
}
if filename == "" {
fmt.Println("Provide a filename (or 'stdout')")
continue
}
if filename == "stdout" {
return os.Stdout, false, nil
}
f, err = os.OpenFile(filename, os.O_WRONLY|os.O_CREATE|os.O_EXCL, 0600)
if err == nil {
return f, true, nil
}
fmt.Printf("Can't create '%s': %s\n", filename, err)
}
}
func runWireGuardResetPeer(ctx *cmdctx.CmdContext) error {
org, err := orgByArg(ctx)
if err != nil {
return err
}
client := ctx.Client.API()
agentclient, err := agent.Establish(context.Background(), client)
if err != nil {
return err
}
conf, err := agentclient.Reestablish(context.Background(), org.Slug)
if err != nil {
return err
}
fmt.Printf("New WireGuard peer for organization '%s': '%s'\n", org.Slug, conf.WireGuardState.Name)
return nil
}
func runWireGuardCreate(ctx *cmdctx.CmdContext) error {
org, err := orgByArg(ctx)
if err != nil {
return err
}
var region string
var name string
if len(ctx.Args) > 1 && ctx.Args[1] != "" {
region = ctx.Args[1]
}
if len(ctx.Args) > 2 && ctx.Args[2] != "" {
name = ctx.Args[2]
}
state, err := wireguard.Create(ctx.Client.API(), org, region, name)
if err != nil {
return err
}
data := &state.Peer
fmt.Printf(`
!!!! WARNING: Output includes private key. Private keys cannot be recovered !!!!
!!!! after creating the peer; if you lose the key, you'll need to remove !!!!
!!!! and re-add the peering connection. !!!!
`)
w, shouldClose, err := resolveOutputWriter(ctx, 3, "Filename to store WireGuard configuration in, or 'stdout': ")
if err != nil {
return err
}
if shouldClose {
defer w.Close()
}
generateWgConf(data, state.LocalPrivate, w)
if shouldClose {
filename := w.(*os.File).Name()
fmt.Printf("Wrote WireGuard configuration to %s; load in your WireGuard client\n", filename)
}
return nil
}
func runWireGuardRemove(cmdCtx *cmdctx.CmdContext) error {
ctx := cmdCtx.Command.Context()
client := cmdCtx.Client.API()
org, err := orgByArg(cmdCtx)
if err != nil {
return err
}
var name string
if len(cmdCtx.Args) >= 2 {
name = cmdCtx.Args[1]
} else {
name, err = selectWireGuardPeer(ctx, cmdCtx.Client.API(), org.Slug)
if err != nil {
return err
}
}
fmt.Printf("Removing WireGuard peer \"%s\" for organization %s\n", name, org.Slug)
err = client.RemoveWireGuardPeer(ctx, org, name)
if err != nil {
return err
}
fmt.Println("Removed peer.")
return wireguard.PruneInvalidPeers(ctx, cmdCtx.Client.API())
}
func runWireGuardStat(cmdCtx *cmdctx.CmdContext) error {
ctx := cmdCtx.Command.Context()
client := cmdCtx.Client.API()
org, err := orgByArg(cmdCtx)
if err != nil {
return err
}
var name string
if len(cmdCtx.Args) >= 2 {
name = cmdCtx.Args[1]
} else {
name, err = selectWireGuardPeer(ctx, cmdCtx.Client.API(), org.Slug)
if err != nil {
return err
}
}
status, err := client.GetWireGuardPeerStatus(ctx, org.Slug, name)
if err != nil {
return err
}
fmt.Printf("Alive: %+v\n", status.Live)
if status.WgError != "" {
fmt.Printf("Gateway error: %s\n", status.WgError)
}
if !status.Live {
return nil
}
if status.Endpoint != "" {
fmt.Printf("Last Source Address: %s\n", status.Endpoint)
}
ago := ""
if status.SinceAdded != "" {
ago = " (" + status.SinceAdded + " ago)"
}
if status.LastHandshake != "" {
fmt.Printf("Last Handshake At: %s%s\n", status.LastHandshake, ago)
}
ago = ""
if status.SinceHandshake != "" {
ago = " (" + status.SinceHandshake + " ago)"
}
fmt.Printf("Installed On Gateway At: %s%s\n", status.Added, ago)
fmt.Printf("Traffic: rx:%d tx:%d\n", status.Rx, status.Tx)
return nil
}
func runWireGuardTokenList(cmdCtx *cmdctx.CmdContext) error {
ctx := cmdCtx.Command.Context()
client := cmdCtx.Client.API()
org, err := orgByArg(cmdCtx)
if err != nil {
return err
}
tokens, err := client.GetDelegatedWireGuardTokens(ctx, org.Slug)
if err != nil {
return err
}
if cmdCtx.OutputJSON() {
cmdCtx.WriteJSON(tokens)
return nil
}
table := tablewriter.NewWriter(cmdCtx.Out)
table.SetHeader([]string{
"Name",
})
for _, peer := range tokens {
table.Append([]string{peer.Name})
}
table.Render()
return nil
}
func runWireGuardTokenCreate(cmdCtx *cmdctx.CmdContext) error {
ctx := cmdCtx.Command.Context()
client := cmdCtx.Client.API()
org, err := orgByArg(cmdCtx)
if err != nil {
return err
}
name, err := argOrPrompt(cmdCtx, 1, "Memorable name for WireGuard token: ")
if err != nil {
return err
}
data, err := client.CreateDelegatedWireGuardToken(ctx, org, name)
if err != nil {
return err
}
fmt.Printf(`
!!!! WARNING: Output includes credential information. Credentials cannot !!!!
!!!! be recovered after creation; if you lose the token, you'll need to !!!!
!!!! remove and and re-add it. !!!!
To use a token to create a WireGuard connection, you can use curl:
curl -v --request POST
-H "Authorization: Bearer ${WG_TOKEN}"
-H "Content-Type: application/json"
--data '{" | resolveOutputWriter | identifier_name |
wireguard.go | argOrPromptImpl(ctx, nth, prompt, true)
}
func orgByArg(cmdCtx *cmdctx.CmdContext) (*api.Organization, error) {
ctx := cmdCtx.Command.Context()
client := cmdCtx.Client.API()
if len(cmdCtx.Args) == 0 {
org, err := selectOrganization(ctx, client, "", nil)
if err != nil {
return nil, err
}
return org, nil
}
return client.FindOrganizationBySlug(ctx, cmdCtx.Args[0])
}
func runWireGuardList(cmdCtx *cmdctx.CmdContext) error {
ctx := cmdCtx.Command.Context()
client := cmdCtx.Client.API()
org, err := orgByArg(cmdCtx)
if err != nil {
return err
}
peers, err := client.GetWireGuardPeers(ctx, org.Slug)
if err != nil {
return err
}
if cmdCtx.OutputJSON() {
cmdCtx.WriteJSON(peers)
return nil
}
table := tablewriter.NewWriter(cmdCtx.Out)
table.SetHeader([]string{
"Name",
"Region",
"Peer IP",
})
for _, peer := range peers {
table.Append([]string{peer.Name, peer.Region, peer.Peerip})
}
table.Render()
return nil
}
func generateWgConf(peer *api.CreatedWireGuardPeer, privkey string, w io.Writer) {
templateStr := `
[Interface]
PrivateKey = {{.Meta.Privkey}}
Address = {{.Peer.Peerip}}/120
DNS = {{.Meta.DNS}}
[Peer]
PublicKey = {{.Peer.Pubkey}}
AllowedIPs = {{.Meta.AllowedIPs}}
Endpoint = {{.Peer.Endpointip}}:51820
PersistentKeepalive = 15
`
data := struct {
Peer *api.CreatedWireGuardPeer
Meta struct {
Privkey string
AllowedIPs string
DNS string
}
}{
Peer: peer,
}
addr := net.ParseIP(peer.Peerip).To16()
for i := 6; i < 16; i++ {
addr[i] = 0
}
// BUG(tqbf): can't stay this way
data.Meta.AllowedIPs = fmt.Sprintf("%s/48", addr)
addr[15] = 3
data.Meta.DNS = fmt.Sprintf("%s", addr)
data.Meta.Privkey = privkey
tmpl := template.Must(template.New("name").Parse(templateStr))
tmpl.Execute(w, &data)
}
func resolveOutputWriter(ctx *cmdctx.CmdContext, idx int, prompt string) (w io.WriteCloser, mustClose bool, err error) {
var (
f *os.File
filename string
)
for {
filename, err = argOrPromptLoop(ctx, idx, prompt, filename)
if err != nil {
return nil, false, err
}
if filename == "" {
fmt.Println("Provide a filename (or 'stdout')")
continue
}
if filename == "stdout" {
return os.Stdout, false, nil
}
f, err = os.OpenFile(filename, os.O_WRONLY|os.O_CREATE|os.O_EXCL, 0600)
if err == nil {
return f, true, nil
}
fmt.Printf("Can't create '%s': %s\n", filename, err)
}
}
func runWireGuardResetPeer(ctx *cmdctx.CmdContext) error {
org, err := orgByArg(ctx)
if err != nil {
return err
}
client := ctx.Client.API()
agentclient, err := agent.Establish(context.Background(), client)
if err != nil {
return err
}
conf, err := agentclient.Reestablish(context.Background(), org.Slug)
if err != nil {
return err
}
fmt.Printf("New WireGuard peer for organization '%s': '%s'\n", org.Slug, conf.WireGuardState.Name)
return nil
}
func runWireGuardCreate(ctx *cmdctx.CmdContext) error {
org, err := orgByArg(ctx)
if err != nil {
return err
}
var region string
var name string
if len(ctx.Args) > 1 && ctx.Args[1] != "" {
region = ctx.Args[1]
}
if len(ctx.Args) > 2 && ctx.Args[2] != "" {
name = ctx.Args[2]
}
state, err := wireguard.Create(ctx.Client.API(), org, region, name)
if err != nil {
return err
}
data := &state.Peer
fmt.Printf(`
!!!! WARNING: Output includes private key. Private keys cannot be recovered !!!!
!!!! after creating the peer; if you lose the key, you'll need to remove !!!!
!!!! and re-add the peering connection. !!!!
`)
w, shouldClose, err := resolveOutputWriter(ctx, 3, "Filename to store WireGuard configuration in, or 'stdout': ")
if err != nil {
return err
}
if shouldClose {
defer w.Close()
}
generateWgConf(data, state.LocalPrivate, w)
if shouldClose {
filename := w.(*os.File).Name()
fmt.Printf("Wrote WireGuard configuration to %s; load in your WireGuard client\n", filename)
}
return nil
}
func runWireGuardRemove(cmdCtx *cmdctx.CmdContext) error {
ctx := cmdCtx.Command.Context()
client := cmdCtx.Client.API()
org, err := orgByArg(cmdCtx)
if err != nil {
return err
}
var name string
if len(cmdCtx.Args) >= 2 {
name = cmdCtx.Args[1]
} else {
name, err = selectWireGuardPeer(ctx, cmdCtx.Client.API(), org.Slug)
if err != nil {
return err
}
}
fmt.Printf("Removing WireGuard peer \"%s\" for organization %s\n", name, org.Slug)
err = client.RemoveWireGuardPeer(ctx, org, name)
if err != nil {
return err
}
| }
func runWireGuardStat(cmdCtx *cmdctx.CmdContext) error {
ctx := cmdCtx.Command.Context()
client := cmdCtx.Client.API()
org, err := orgByArg(cmdCtx)
if err != nil {
return err
}
var name string
if len(cmdCtx.Args) >= 2 {
name = cmdCtx.Args[1]
} else {
name, err = selectWireGuardPeer(ctx, cmdCtx.Client.API(), org.Slug)
if err != nil {
return err
}
}
status, err := client.GetWireGuardPeerStatus(ctx, org.Slug, name)
if err != nil {
return err
}
fmt.Printf("Alive: %+v\n", status.Live)
if status.WgError != "" {
fmt.Printf("Gateway error: %s\n", status.WgError)
}
if !status.Live {
return nil
}
if status.Endpoint != "" {
fmt.Printf("Last Source Address: %s\n", status.Endpoint)
}
ago := ""
if status.SinceAdded != "" {
ago = " (" + status.SinceAdded + " ago)"
}
if status.LastHandshake != "" {
fmt.Printf("Last Handshake At: %s%s\n", status.LastHandshake, ago)
}
ago = ""
if status.SinceHandshake != "" {
ago = " (" + status.SinceHandshake + " ago)"
}
fmt.Printf("Installed On Gateway At: %s%s\n", status.Added, ago)
fmt.Printf("Traffic: rx:%d tx:%d\n", status.Rx, status.Tx)
return nil
}
func runWireGuardTokenList(cmdCtx *cmdctx.CmdContext) error {
ctx := cmdCtx.Command.Context()
client := cmdCtx.Client.API()
org, err := orgByArg(cmdCtx)
if err != nil {
return err
}
tokens, err := client.GetDelegatedWireGuardTokens(ctx, org.Slug)
if err != nil {
return err
}
if cmdCtx.OutputJSON() {
cmdCtx.WriteJSON(tokens)
return nil
}
table := tablewriter.NewWriter(cmdCtx.Out)
table.SetHeader([]string{
"Name",
})
for _, peer := range tokens {
table.Append([]string{peer.Name})
}
table.Render()
return nil
}
func runWireGuardTokenCreate(cmdCtx *cmdctx.CmdContext) error {
ctx := cmdCtx.Command.Context()
client := cmdCtx.Client.API()
org, err := orgByArg(cmdCtx)
if err != nil {
return err
}
name, err := argOrPrompt(cmdCtx, 1, "Memorable name for WireGuard token: ")
if err != nil {
return err
}
data, err := client.CreateDelegatedWireGuardToken(ctx, org, name)
if err != nil {
return err
}
fmt.Printf(`
!!!! WARNING: Output includes credential information. Credentials cannot !!!!
!!!! be recovered after creation; if you lose the token, you'll need to !!!!
!!!! remove and and re-add it. !!!!
To use a token to create a WireGuard connection, you can use curl:
curl -v --request POST
-H "Authorization: Bearer ${WG_TOKEN}"
-H "Content-Type: application/json"
--data '{" | fmt.Println("Removed peer.")
return wireguard.PruneInvalidPeers(ctx, cmdCtx.Client.API()) | random_line_split |
wireguard.go | != nil {
return err
}
fmt.Printf("New WireGuard peer for organization '%s': '%s'\n", org.Slug, conf.WireGuardState.Name)
return nil
}
func runWireGuardCreate(ctx *cmdctx.CmdContext) error {
org, err := orgByArg(ctx)
if err != nil {
return err
}
var region string
var name string
if len(ctx.Args) > 1 && ctx.Args[1] != "" {
region = ctx.Args[1]
}
if len(ctx.Args) > 2 && ctx.Args[2] != "" {
name = ctx.Args[2]
}
state, err := wireguard.Create(ctx.Client.API(), org, region, name)
if err != nil {
return err
}
data := &state.Peer
fmt.Printf(`
!!!! WARNING: Output includes private key. Private keys cannot be recovered !!!!
!!!! after creating the peer; if you lose the key, you'll need to remove !!!!
!!!! and re-add the peering connection. !!!!
`)
w, shouldClose, err := resolveOutputWriter(ctx, 3, "Filename to store WireGuard configuration in, or 'stdout': ")
if err != nil {
return err
}
if shouldClose {
defer w.Close()
}
generateWgConf(data, state.LocalPrivate, w)
if shouldClose {
filename := w.(*os.File).Name()
fmt.Printf("Wrote WireGuard configuration to %s; load in your WireGuard client\n", filename)
}
return nil
}
func runWireGuardRemove(cmdCtx *cmdctx.CmdContext) error {
ctx := cmdCtx.Command.Context()
client := cmdCtx.Client.API()
org, err := orgByArg(cmdCtx)
if err != nil {
return err
}
var name string
if len(cmdCtx.Args) >= 2 {
name = cmdCtx.Args[1]
} else {
name, err = selectWireGuardPeer(ctx, cmdCtx.Client.API(), org.Slug)
if err != nil {
return err
}
}
fmt.Printf("Removing WireGuard peer \"%s\" for organization %s\n", name, org.Slug)
err = client.RemoveWireGuardPeer(ctx, org, name)
if err != nil {
return err
}
fmt.Println("Removed peer.")
return wireguard.PruneInvalidPeers(ctx, cmdCtx.Client.API())
}
func runWireGuardStat(cmdCtx *cmdctx.CmdContext) error {
ctx := cmdCtx.Command.Context()
client := cmdCtx.Client.API()
org, err := orgByArg(cmdCtx)
if err != nil {
return err
}
var name string
if len(cmdCtx.Args) >= 2 {
name = cmdCtx.Args[1]
} else {
name, err = selectWireGuardPeer(ctx, cmdCtx.Client.API(), org.Slug)
if err != nil {
return err
}
}
status, err := client.GetWireGuardPeerStatus(ctx, org.Slug, name)
if err != nil {
return err
}
fmt.Printf("Alive: %+v\n", status.Live)
if status.WgError != "" {
fmt.Printf("Gateway error: %s\n", status.WgError)
}
if !status.Live {
return nil
}
if status.Endpoint != "" {
fmt.Printf("Last Source Address: %s\n", status.Endpoint)
}
ago := ""
if status.SinceAdded != "" {
ago = " (" + status.SinceAdded + " ago)"
}
if status.LastHandshake != "" {
fmt.Printf("Last Handshake At: %s%s\n", status.LastHandshake, ago)
}
ago = ""
if status.SinceHandshake != "" {
ago = " (" + status.SinceHandshake + " ago)"
}
fmt.Printf("Installed On Gateway At: %s%s\n", status.Added, ago)
fmt.Printf("Traffic: rx:%d tx:%d\n", status.Rx, status.Tx)
return nil
}
func runWireGuardTokenList(cmdCtx *cmdctx.CmdContext) error {
ctx := cmdCtx.Command.Context()
client := cmdCtx.Client.API()
org, err := orgByArg(cmdCtx)
if err != nil {
return err
}
tokens, err := client.GetDelegatedWireGuardTokens(ctx, org.Slug)
if err != nil {
return err
}
if cmdCtx.OutputJSON() {
cmdCtx.WriteJSON(tokens)
return nil
}
table := tablewriter.NewWriter(cmdCtx.Out)
table.SetHeader([]string{
"Name",
})
for _, peer := range tokens {
table.Append([]string{peer.Name})
}
table.Render()
return nil
}
func runWireGuardTokenCreate(cmdCtx *cmdctx.CmdContext) error {
ctx := cmdCtx.Command.Context()
client := cmdCtx.Client.API()
org, err := orgByArg(cmdCtx)
if err != nil {
return err
}
name, err := argOrPrompt(cmdCtx, 1, "Memorable name for WireGuard token: ")
if err != nil {
return err
}
data, err := client.CreateDelegatedWireGuardToken(ctx, org, name)
if err != nil {
return err
}
fmt.Printf(`
!!!! WARNING: Output includes credential information. Credentials cannot !!!!
!!!! be recovered after creation; if you lose the token, you'll need to !!!!
!!!! remove and and re-add it. !!!!
To use a token to create a WireGuard connection, you can use curl:
curl -v --request POST
-H "Authorization: Bearer ${WG_TOKEN}"
-H "Content-Type: application/json"
--data '{"name": "node-1", \
"group": "k8s", \
"pubkey": "'"${WG_PUBKEY}"'", \
"region": "dev"}'
http://fly.io/api/v3/wire_guard_peers
We'll return 'us' (our local 6PN address), 'them' (the gateway IP address),
and 'pubkey' (the public key of the gateway), which you can inject into a
"wg.con".
`)
w, shouldClose, err := resolveOutputWriter(cmdCtx, 2, "Filename to store WireGuard token in, or 'stdout': ")
if err != nil {
return err
}
if shouldClose {
defer w.Close()
}
fmt.Fprintf(w, "FLY_WIREGUARD_TOKEN=%s\n", data.Token)
return nil
}
func runWireGuardTokenDelete(cmdCtx *cmdctx.CmdContext) error {
ctx := cmdCtx.Command.Context()
client := cmdCtx.Client.API()
org, err := orgByArg(cmdCtx)
if err != nil {
return err
}
kv, err := argOrPrompt(cmdCtx, 1, "'name:<name>' or token:<token>': ")
if err != nil {
return err
}
tup := strings.SplitN(kv, ":", 2)
if len(tup) != 2 || (tup[0] != "name" && tup[0] != "token") {
return fmt.Errorf("format is name:<name> or token:<token>")
}
fmt.Printf("Removing WireGuard token \"%s\" for organization %s\n", kv, org.Slug)
if tup[0] == "name" {
err = client.DeleteDelegatedWireGuardToken(ctx, org, &tup[1], nil)
} else {
err = client.DeleteDelegatedWireGuardToken(ctx, org, nil, &tup[1])
}
if err != nil {
return err
}
fmt.Println("Removed token.")
return nil
}
func tokenRequest(method, path, token string, data interface{}) (*http.Response, error) {
buf := &bytes.Buffer{}
if err := json.NewEncoder(buf).Encode(data); err != nil {
return nil, err
}
req, err := http.NewRequest(method,
fmt.Sprintf("https://fly.io/api/v3/wire_guard_peers%s", path),
buf)
if err != nil {
return nil, err
}
req.Header.Add("Authorization", "Bearer "+token)
req.Header.Add("Content-Type", "application/json")
return (&http.Client{}).Do(req)
}
type StartPeerJson struct {
Name string `json:"name"`
Group string `json:"group"`
Pubkey string `json:"pubkey"`
Region string `json:"region"`
}
type UpdatePeerJson struct {
Pubkey string `json:"pubkey"`
}
type PeerStatusJson struct {
Us string `json:"us"`
Them string `json:"them"`
Pubkey string `json:"key"`
Error string `json:"error"`
}
func generateTokenConf(ctx *cmdctx.CmdContext, idx int, stat *PeerStatusJson, privkey string) error | {
fmt.Printf(`
!!!! WARNING: Output includes private key. Private keys cannot be recovered !!!!
!!!! after creating the peer; if you lose the key, you'll need to rekey !!!!
!!!! the peering connection. !!!!
`)
w, shouldClose, err := resolveOutputWriter(ctx, idx, "Filename to store WireGuard configuration in, or 'stdout': ")
if err != nil {
return err
}
if shouldClose {
defer w.Close()
}
generateWgConf(&api.CreatedWireGuardPeer{
Peerip: stat.Us,
Pubkey: stat.Pubkey,
Endpointip: stat.Them,
}, privkey, w) | identifier_body |
|
nba-game-list.ts | Date();
//we use America/Los_Angeles time zone because L.A. game start at last in one day.
let SpecificDateTimeArray: any[] = this.GetSpecificTimeZoneFormat(nowLocalTime, -7);
this.ChangedDate = SpecificDateTimeArray[0] + SpecificDateTimeArray[1] + SpecificDateTimeArray[2];
this.gameDate = SpecificDateTimeArray[0] + '-' + SpecificDateTimeArray[1] + '-' + SpecificDateTimeArray[2];
this.selectedYear = SpecificDateTimeArray[0];
this.GetNBAGameList(this.ChangedDate).then(() => null).catch(this.handleError);
}
doRefresh(refresher: Refresher) {
this.GetNBAGameList(this.ChangedDate).then(() => refresher.complete()).catch(this.handleError);
}
OpenDatePicker(): void {
let options: DatePickerOptions;
if (this.platform.is('android')) {
options = {
date: new Date(),
mode: 'datetime',
androidTheme : 3
}
}
else {
options = {
date: new Date(),
mode: 'datetime'
}
}
DatePicker.show(options).then(
date => {
let SpecificDateTimeArray: any[] = this.GetSpecificTimeZoneFormat(date, -7);
this.ChangedDate = SpecificDateTimeArray[0] + SpecificDateTimeArray[1] + SpecificDateTimeArray[2];
this.gameDate = SpecificDateTimeArray[0] + '-' + SpecificDateTimeArray[1] + '-' + SpecificDateTimeArray[2];
this.selectedYear = SpecificDateTimeArray[0];
this.GetNBAGameList(this.ChangedDate).then(() => null).catch(this.handleError);
},
error => console.log('Error occurred while getting date: ', error)
);
}
GameItemTapped(event, GameItem) {
this.navCtrl.push(NBAGameDetailPage, {
GameItem: GameItem,
SelectedYear: this.selectedYear
});
}
/**
* @Param: nowDateTime, UTC
* @Example: UTC => +8: "Asia/Shanghai", -4: "America/New_York", -7: "America/Los_Angeles"
* @Remark: "Asia/Shanghai" => GMT+8 (CST)
* "America/New_York" => GMT-4 (EDT)
* "America/Los_Angeles" => GMT-7 (PDT)
*/
private GetSpecificTimeZoneFormat(SpecificLocalDateTime: Date, UTC: number): any[] {
let YMDDateArray: any[] = []; //need to assign initial value or occur error : cannot access property push of undefined
//Return the number of milliseconds since 1970/01/01
let localTime: number = SpecificLocalDateTime.getTime();
//Return the timezone difference between UTC and Local Time
//By default, this method returns the time zone offset in minutes, so convert this value to milliseconds for easier manipulation:
//Note that a negative return value from getTimezoneOffset() indicates that the current location is ahead of UTC,
//while a positive value indicates that the location is behind UTC. (ex: Taiwan is UTC+8 => 8 * 60 = -480)
//Obtain local UTC offset and convert to msec
let localOffset: number = SpecificLocalDateTime.getTimezoneOffset() * 60000;
//Obtain UTC time in msec
let utc = localTime + localOffset;
// Obtain and add destination's UTC time offset. For example, America/Los_Angeles which is UTC - 7 hours
let offset = UTC;
//Note: In case you're wondering how I arrived at 3600000 as the multiplication factor, remember that
//1000 millseconds = 1 second, and 1 hour = 3600 seconds. Therefore, converting hours to milliseconds
//involves multiplying by 3600 * 1000 = 3600000.
let timezone = utc + (3600000 * offset);
let anotherTimeZoneTime = new Date(timezone);
let nowYear: string = anotherTimeZoneTime.getFullYear().toString();
//Months are zero based
let nowMonth: string = (anotherTimeZoneTime.getMonth() + 1) < 10 ? "0" + (anotherTimeZoneTime.getMonth() + 1).toString() : (anotherTimeZoneTime.getMonth() + 1).toString();
let nowDate: string = anotherTimeZoneTime.getDate() < 10 ? "0" + anotherTimeZoneTime.getDate().toString() : anotherTimeZoneTime.getDate().toString();
//let nowHour: string = anotherTimeZoneTime.getHours() < 10 ? "0" + anotherTimeZoneTime.getHours().toString() : anotherTimeZoneTime.getHours().toString();
//let nowMinute: string = anotherTimeZoneTime.getMinutes().toString();
//let nowSecond: string = anotherTimeZoneTime.getSeconds().toString();
YMDDateArray.push(nowYear, nowMonth, nowDate);
return YMDDateArray;
//[0]:year, [1]:month, [2]:day
}
private GetNBAGameList(GameDate: string): Promise<any> {
return this.NBAdataservices.GetGameGeneral(GameDate)
.then(gameGeneral => {
//before everytime we get data, we should clear the former game data
this.NBAGameList.length = 0;
let unstartArray: any[] = gameGeneral['unstart'];
let liveArray: any[] = gameGeneral['live'];
let overArray: any[] = gameGeneral['over'];
this.gameCount = unstartArray.length + liveArray.length + overArray.length;
if(this.gameCount !== 0) {
//The concat() method is used to join two or more arrays.
let totalArray: any[] = unstartArray.concat(liveArray, overArray);
totalArray.forEach(EachGameitem => {
let GameID: string = EachGameitem['gameID'];
let GameDate: string = EachGameitem['gameDate'];
let HomeTeamID: string = EachGameitem['home']['TeamID'];
let HomeAbbr: string = EachGameitem['home']['TeamAbbr'];
let VisitorTeamID: string = EachGameitem['visitor']['TeamID'];
let VisitorAbbr: string = EachGameitem['visitor']['TeamAbbr'];
let GameProcess: string;
let GameType: string = EachGameitem['GameProcessType'];
switch(GameType) {
case 'unstart':
GameProcess = EachGameitem['GameDate'].replace(/\s*ET\s*/, '');
break;
case 'live':
GameProcess = EachGameitem['process']['Quarter'] + ' ';
GameProcess += EachGameitem['process']['Time'].replace(/\s+/, '');
break;
case 'over':
GameProcess = 'Final';
break;
default:
return;
}
let HomeScore: number = EachGameitem['home']['Score'];
let VisitorScore: number = EachGameitem['visitor']['Score'];
//Get City, Team Name, Logo, Color from TeamMap services based on Abbreviation, so use Array.filter to get match result
//Only return an array that count is 1 because Abbreviation is unique.
let HomeMatchAbbrArray: any[] = this.NBATeamMapData.filter(data => data['Abbreviation'] === HomeAbbr);
let VisitorMatchAbbrArray: any[] = this.NBATeamMapData.filter(data => data['Abbreviation'] === VisitorAbbr);
let HomeTeamColor: string = HomeMatchAbbrArray[0]['Color'];
let HomeTeamCity: string = HomeMatchAbbrArray[0]['City'];
let VisitorTeamCity: string = VisitorMatchAbbrArray[0]['City'];
let HomeTeamName: string = HomeMatchAbbrArray[0]['Team'];
let VisitorTeamName: string = VisitorMatchAbbrArray[0]['Team'];
let HomeTeamLogo: string = HomeMatchAbbrArray[0]['imgGameUrl'];
let VisitorTeamLogo: string = VisitorMatchAbbrArray[0]['imgGameUrl'];
//Finally, we create an array contains all information we want to show in html, and use this array
//as the datasource of ion-card, which use *ngFor to produce game list.
this.NBAGameList.push({
GameID: GameID,
GameDate: GameDate,
HomeTeamID: HomeTeamID,
HomeTeamColor: HomeTeamColor,
HomeTeamCity: HomeTeamCity,
VisitorTeamID: VisitorTeamID,
VisitorTeamCity: VisitorTeamCity,
HomeTeamName: HomeTeamName,
VisitorTeamName: VisitorTeamName,
HomeTeamLogo: HomeTeamLogo,
VisitorTeamLogo: VisitorTeamLogo,
GameProcess: GameProcess,
HomeScore: HomeScore,
VisitorScore: VisitorScore
});
});
}
else | {
this.gameCount = 0;
let alert = this.GamealertCtrl.create({
title: 'Oops!',
subTitle: 'There are not any game today or the day you select.',
buttons: ['OK']
});
alert.present();
} | conditional_block |
|
nba-game-list.ts | ameDetailPage } from '../nba-game-detail/nba-game-detail';
@Component({
templateUrl: 'build/pages/nba-game-list/nba-game-list.html'
})
export class NBAGameListPage implements OnInit{
NBATeamMapData: any[];
NBAGameList: any[] = []; //to avoid error: Cannot read property 'push' of undefined, so we give it an initial value
gameCount: number;
gameDate: string;
ChangedDate: string;
selectedYear: string;
constructor(private NBAteammap: NBATeamMap, private NBAdataservices: NBADataServices,
private GamealertCtrl: AlertController, public navCtrl: NavController, public platform: Platform ) {
this.NBATeamMapData = NBAteammap.getNBATeamArrayData();
}
ngOnInit() {
let nowLocalTime: Date = new Date();
//we use America/Los_Angeles time zone because L.A. game start at last in one day.
let SpecificDateTimeArray: any[] = this.GetSpecificTimeZoneFormat(nowLocalTime, -7);
this.ChangedDate = SpecificDateTimeArray[0] + SpecificDateTimeArray[1] + SpecificDateTimeArray[2];
this.gameDate = SpecificDateTimeArray[0] + '-' + SpecificDateTimeArray[1] + '-' + SpecificDateTimeArray[2];
this.selectedYear = SpecificDateTimeArray[0];
this.GetNBAGameList(this.ChangedDate).then(() => null).catch(this.handleError);
}
doRefresh(refresher: Refresher) {
this.GetNBAGameList(this.ChangedDate).then(() => refresher.complete()).catch(this.handleError);
}
OpenDatePicker(): void {
let options: DatePickerOptions;
if (this.platform.is('android')) {
options = {
date: new Date(),
mode: 'datetime',
androidTheme : 3
}
}
else {
options = {
date: new Date(),
mode: 'datetime'
}
}
DatePicker.show(options).then(
date => {
let SpecificDateTimeArray: any[] = this.GetSpecificTimeZoneFormat(date, -7);
this.ChangedDate = SpecificDateTimeArray[0] + SpecificDateTimeArray[1] + SpecificDateTimeArray[2];
this.gameDate = SpecificDateTimeArray[0] + '-' + SpecificDateTimeArray[1] + '-' + SpecificDateTimeArray[2];
this.selectedYear = SpecificDateTimeArray[0];
this.GetNBAGameList(this.ChangedDate).then(() => null).catch(this.handleError);
},
error => console.log('Error occurred while getting date: ', error)
);
}
| (event, GameItem) {
this.navCtrl.push(NBAGameDetailPage, {
GameItem: GameItem,
SelectedYear: this.selectedYear
});
}
/**
* @Param: nowDateTime, UTC
* @Example: UTC => +8: "Asia/Shanghai", -4: "America/New_York", -7: "America/Los_Angeles"
* @Remark: "Asia/Shanghai" => GMT+8 (CST)
* "America/New_York" => GMT-4 (EDT)
* "America/Los_Angeles" => GMT-7 (PDT)
*/
private GetSpecificTimeZoneFormat(SpecificLocalDateTime: Date, UTC: number): any[] {
let YMDDateArray: any[] = []; //need to assign initial value or occur error : cannot access property push of undefined
//Return the number of milliseconds since 1970/01/01
let localTime: number = SpecificLocalDateTime.getTime();
//Return the timezone difference between UTC and Local Time
//By default, this method returns the time zone offset in minutes, so convert this value to milliseconds for easier manipulation:
//Note that a negative return value from getTimezoneOffset() indicates that the current location is ahead of UTC,
//while a positive value indicates that the location is behind UTC. (ex: Taiwan is UTC+8 => 8 * 60 = -480)
//Obtain local UTC offset and convert to msec
let localOffset: number = SpecificLocalDateTime.getTimezoneOffset() * 60000;
//Obtain UTC time in msec
let utc = localTime + localOffset;
// Obtain and add destination's UTC time offset. For example, America/Los_Angeles which is UTC - 7 hours
let offset = UTC;
//Note: In case you're wondering how I arrived at 3600000 as the multiplication factor, remember that
//1000 millseconds = 1 second, and 1 hour = 3600 seconds. Therefore, converting hours to milliseconds
//involves multiplying by 3600 * 1000 = 3600000.
let timezone = utc + (3600000 * offset);
let anotherTimeZoneTime = new Date(timezone);
let nowYear: string = anotherTimeZoneTime.getFullYear().toString();
//Months are zero based
let nowMonth: string = (anotherTimeZoneTime.getMonth() + 1) < 10 ? "0" + (anotherTimeZoneTime.getMonth() + 1).toString() : (anotherTimeZoneTime.getMonth() + 1).toString();
let nowDate: string = anotherTimeZoneTime.getDate() < 10 ? "0" + anotherTimeZoneTime.getDate().toString() : anotherTimeZoneTime.getDate().toString();
//let nowHour: string = anotherTimeZoneTime.getHours() < 10 ? "0" + anotherTimeZoneTime.getHours().toString() : anotherTimeZoneTime.getHours().toString();
//let nowMinute: string = anotherTimeZoneTime.getMinutes().toString();
//let nowSecond: string = anotherTimeZoneTime.getSeconds().toString();
YMDDateArray.push(nowYear, nowMonth, nowDate);
return YMDDateArray;
//[0]:year, [1]:month, [2]:day
}
private GetNBAGameList(GameDate: string): Promise<any> {
return this.NBAdataservices.GetGameGeneral(GameDate)
.then(gameGeneral => {
//before everytime we get data, we should clear the former game data
this.NBAGameList.length = 0;
let unstartArray: any[] = gameGeneral['unstart'];
let liveArray: any[] = gameGeneral['live'];
let overArray: any[] = gameGeneral['over'];
this.gameCount = unstartArray.length + liveArray.length + overArray.length;
if(this.gameCount !== 0) {
//The concat() method is used to join two or more arrays.
let totalArray: any[] = unstartArray.concat(liveArray, overArray);
totalArray.forEach(EachGameitem => {
let GameID: string = EachGameitem['gameID'];
let GameDate: string = EachGameitem['gameDate'];
let HomeTeamID: string = EachGameitem['home']['TeamID'];
let HomeAbbr: string = EachGameitem['home']['TeamAbbr'];
let VisitorTeamID: string = EachGameitem['visitor']['TeamID'];
let VisitorAbbr: string = EachGameitem['visitor']['TeamAbbr'];
let GameProcess: string;
let GameType: string = EachGameitem['GameProcessType'];
switch(GameType) {
case 'unstart':
GameProcess = EachGameitem['GameDate'].replace(/\s*ET\s*/, '');
break;
case 'live':
GameProcess = EachGameitem['process']['Quarter'] + ' ';
GameProcess += EachGameitem['process']['Time'].replace(/\s+/, '');
break;
case 'over':
GameProcess = 'Final';
break;
default:
return;
}
let HomeScore: number = EachGameitem['home']['Score'];
let VisitorScore: number = EachGameitem['visitor']['Score'];
//Get City, Team Name, Logo, Color from TeamMap services based on Abbreviation, so use Array.filter to get match result
//Only return an array that count is 1 because Abbreviation is unique.
let HomeMatchAbbrArray: any[] = this.NBATeamMapData.filter(data => data['Abbreviation'] === HomeAbbr);
let VisitorMatchAbbrArray: any[] = this.NBATeamMapData.filter(data => data['Abbreviation'] === VisitorAbbr);
let HomeTeamColor: string = HomeMatchAbbrArray[0]['Color'];
let HomeTeamCity: string = HomeMatchAbbrArray[0]['City'];
let VisitorTeamCity: string = VisitorMatchAbbrArray[0]['City'];
let HomeTeamName: string = HomeMatchAbbrArray[0]['Team'];
let VisitorTeamName: string = VisitorMatchAbbrArray[0]['Team'];
let HomeTeamLogo: string = HomeMatchAbbrArray[0]['imgGameUrl'];
let VisitorTeamLogo: string = VisitorMatchAbbrArray[0]['imgGameUrl'];
//Finally, we create an array contains all information we want to show in html, and use this array
//as the datasource of ion-card, which use *ngFor to produce game list | GameItemTapped | identifier_name |
nba-game-list.ts | DetailPage } from '../nba-game-detail/nba-game-detail';
@Component({
templateUrl: 'build/pages/nba-game-list/nba-game-list.html'
})
export class NBAGameListPage implements OnInit{
NBATeamMapData: any[];
NBAGameList: any[] = []; //to avoid error: Cannot read property 'push' of undefined, so we give it an initial value
gameCount: number;
gameDate: string;
ChangedDate: string;
selectedYear: string;
constructor(private NBAteammap: NBATeamMap, private NBAdataservices: NBADataServices,
private GamealertCtrl: AlertController, public navCtrl: NavController, public platform: Platform ) {
this.NBATeamMapData = NBAteammap.getNBATeamArrayData();
}
ngOnInit() |
doRefresh(refresher: Refresher) {
this.GetNBAGameList(this.ChangedDate).then(() => refresher.complete()).catch(this.handleError);
}
OpenDatePicker(): void {
let options: DatePickerOptions;
if (this.platform.is('android')) {
options = {
date: new Date(),
mode: 'datetime',
androidTheme : 3
}
}
else {
options = {
date: new Date(),
mode: 'datetime'
}
}
DatePicker.show(options).then(
date => {
let SpecificDateTimeArray: any[] = this.GetSpecificTimeZoneFormat(date, -7);
this.ChangedDate = SpecificDateTimeArray[0] + SpecificDateTimeArray[1] + SpecificDateTimeArray[2];
this.gameDate = SpecificDateTimeArray[0] + '-' + SpecificDateTimeArray[1] + '-' + SpecificDateTimeArray[2];
this.selectedYear = SpecificDateTimeArray[0];
this.GetNBAGameList(this.ChangedDate).then(() => null).catch(this.handleError);
},
error => console.log('Error occurred while getting date: ', error)
);
}
GameItemTapped(event, GameItem) {
this.navCtrl.push(NBAGameDetailPage, {
GameItem: GameItem,
SelectedYear: this.selectedYear
});
}
/**
* @Param: nowDateTime, UTC
* @Example: UTC => +8: "Asia/Shanghai", -4: "America/New_York", -7: "America/Los_Angeles"
* @Remark: "Asia/Shanghai" => GMT+8 (CST)
* "America/New_York" => GMT-4 (EDT)
* "America/Los_Angeles" => GMT-7 (PDT)
*/
private GetSpecificTimeZoneFormat(SpecificLocalDateTime: Date, UTC: number): any[] {
let YMDDateArray: any[] = []; //need to assign initial value or occur error : cannot access property push of undefined
//Return the number of milliseconds since 1970/01/01
let localTime: number = SpecificLocalDateTime.getTime();
//Return the timezone difference between UTC and Local Time
//By default, this method returns the time zone offset in minutes, so convert this value to milliseconds for easier manipulation:
//Note that a negative return value from getTimezoneOffset() indicates that the current location is ahead of UTC,
//while a positive value indicates that the location is behind UTC. (ex: Taiwan is UTC+8 => 8 * 60 = -480)
//Obtain local UTC offset and convert to msec
let localOffset: number = SpecificLocalDateTime.getTimezoneOffset() * 60000;
//Obtain UTC time in msec
let utc = localTime + localOffset;
// Obtain and add destination's UTC time offset. For example, America/Los_Angeles which is UTC - 7 hours
let offset = UTC;
//Note: In case you're wondering how I arrived at 3600000 as the multiplication factor, remember that
//1000 millseconds = 1 second, and 1 hour = 3600 seconds. Therefore, converting hours to milliseconds
//involves multiplying by 3600 * 1000 = 3600000.
let timezone = utc + (3600000 * offset);
let anotherTimeZoneTime = new Date(timezone);
let nowYear: string = anotherTimeZoneTime.getFullYear().toString();
//Months are zero based
let nowMonth: string = (anotherTimeZoneTime.getMonth() + 1) < 10 ? "0" + (anotherTimeZoneTime.getMonth() + 1).toString() : (anotherTimeZoneTime.getMonth() + 1).toString();
let nowDate: string = anotherTimeZoneTime.getDate() < 10 ? "0" + anotherTimeZoneTime.getDate().toString() : anotherTimeZoneTime.getDate().toString();
//let nowHour: string = anotherTimeZoneTime.getHours() < 10 ? "0" + anotherTimeZoneTime.getHours().toString() : anotherTimeZoneTime.getHours().toString();
//let nowMinute: string = anotherTimeZoneTime.getMinutes().toString();
//let nowSecond: string = anotherTimeZoneTime.getSeconds().toString();
YMDDateArray.push(nowYear, nowMonth, nowDate);
return YMDDateArray;
//[0]:year, [1]:month, [2]:day
}
private GetNBAGameList(GameDate: string): Promise<any> {
return this.NBAdataservices.GetGameGeneral(GameDate)
.then(gameGeneral => {
//before everytime we get data, we should clear the former game data
this.NBAGameList.length = 0;
let unstartArray: any[] = gameGeneral['unstart'];
let liveArray: any[] = gameGeneral['live'];
let overArray: any[] = gameGeneral['over'];
this.gameCount = unstartArray.length + liveArray.length + overArray.length;
if(this.gameCount !== 0) {
//The concat() method is used to join two or more arrays.
let totalArray: any[] = unstartArray.concat(liveArray, overArray);
totalArray.forEach(EachGameitem => {
let GameID: string = EachGameitem['gameID'];
let GameDate: string = EachGameitem['gameDate'];
let HomeTeamID: string = EachGameitem['home']['TeamID'];
let HomeAbbr: string = EachGameitem['home']['TeamAbbr'];
let VisitorTeamID: string = EachGameitem['visitor']['TeamID'];
let VisitorAbbr: string = EachGameitem['visitor']['TeamAbbr'];
let GameProcess: string;
let GameType: string = EachGameitem['GameProcessType'];
switch(GameType) {
case 'unstart':
GameProcess = EachGameitem['GameDate'].replace(/\s*ET\s*/, '');
break;
case 'live':
GameProcess = EachGameitem['process']['Quarter'] + ' ';
GameProcess += EachGameitem['process']['Time'].replace(/\s+/, '');
break;
case 'over':
GameProcess = 'Final';
break;
default:
return;
}
let HomeScore: number = EachGameitem['home']['Score'];
let VisitorScore: number = EachGameitem['visitor']['Score'];
//Get City, Team Name, Logo, Color from TeamMap services based on Abbreviation, so use Array.filter to get match result
//Only return an array that count is 1 because Abbreviation is unique.
let HomeMatchAbbrArray: any[] = this.NBATeamMapData.filter(data => data['Abbreviation'] === HomeAbbr);
let VisitorMatchAbbrArray: any[] = this.NBATeamMapData.filter(data => data['Abbreviation'] === VisitorAbbr);
let HomeTeamColor: string = HomeMatchAbbrArray[0]['Color'];
let HomeTeamCity: string = HomeMatchAbbrArray[0]['City'];
let VisitorTeamCity: string = VisitorMatchAbbrArray[0]['City'];
let HomeTeamName: string = HomeMatchAbbrArray[0]['Team'];
let VisitorTeamName: string = VisitorMatchAbbrArray[0]['Team'];
let HomeTeamLogo: string = HomeMatchAbbrArray[0]['imgGameUrl'];
let VisitorTeamLogo: string = VisitorMatchAbbrArray[0]['imgGameUrl'];
//Finally, we create an array contains all information we want to show in html, and use this array
//as the datasource of ion-card, which use *ngFor to produce game | {
let nowLocalTime: Date = new Date();
//we use America/Los_Angeles time zone because L.A. game start at last in one day.
let SpecificDateTimeArray: any[] = this.GetSpecificTimeZoneFormat(nowLocalTime, -7);
this.ChangedDate = SpecificDateTimeArray[0] + SpecificDateTimeArray[1] + SpecificDateTimeArray[2];
this.gameDate = SpecificDateTimeArray[0] + '-' + SpecificDateTimeArray[1] + '-' + SpecificDateTimeArray[2];
this.selectedYear = SpecificDateTimeArray[0];
this.GetNBAGameList(this.ChangedDate).then(() => null).catch(this.handleError);
} | identifier_body |
nba-game-list.ts | ameDetailPage } from '../nba-game-detail/nba-game-detail';
@Component({
templateUrl: 'build/pages/nba-game-list/nba-game-list.html'
})
export class NBAGameListPage implements OnInit{
NBATeamMapData: any[];
NBAGameList: any[] = []; //to avoid error: Cannot read property 'push' of undefined, so we give it an initial value
gameCount: number;
gameDate: string;
ChangedDate: string;
selectedYear: string;
constructor(private NBAteammap: NBATeamMap, private NBAdataservices: NBADataServices,
private GamealertCtrl: AlertController, public navCtrl: NavController, public platform: Platform ) {
this.NBATeamMapData = NBAteammap.getNBATeamArrayData();
}
ngOnInit() {
let nowLocalTime: Date = new Date();
//we use America/Los_Angeles time zone because L.A. game start at last in one day.
let SpecificDateTimeArray: any[] = this.GetSpecificTimeZoneFormat(nowLocalTime, -7);
this.ChangedDate = SpecificDateTimeArray[0] + SpecificDateTimeArray[1] + SpecificDateTimeArray[2];
this.gameDate = SpecificDateTimeArray[0] + '-' + SpecificDateTimeArray[1] + '-' + SpecificDateTimeArray[2];
this.selectedYear = SpecificDateTimeArray[0];
this.GetNBAGameList(this.ChangedDate).then(() => null).catch(this.handleError);
}
doRefresh(refresher: Refresher) {
this.GetNBAGameList(this.ChangedDate).then(() => refresher.complete()).catch(this.handleError);
}
OpenDatePicker(): void {
let options: DatePickerOptions;
if (this.platform.is('android')) {
options = {
date: new Date(),
mode: 'datetime',
androidTheme : 3
}
}
else {
options = {
date: new Date(),
mode: 'datetime'
}
}
DatePicker.show(options).then(
date => {
let SpecificDateTimeArray: any[] = this.GetSpecificTimeZoneFormat(date, -7);
this.ChangedDate = SpecificDateTimeArray[0] + SpecificDateTimeArray[1] + SpecificDateTimeArray[2];
this.gameDate = SpecificDateTimeArray[0] + '-' + SpecificDateTimeArray[1] + '-' + SpecificDateTimeArray[2];
this.selectedYear = SpecificDateTimeArray[0];
this.GetNBAGameList(this.ChangedDate).then(() => null).catch(this.handleError);
},
error => console.log('Error occurred while getting date: ', error)
);
}
GameItemTapped(event, GameItem) {
this.navCtrl.push(NBAGameDetailPage, {
GameItem: GameItem,
SelectedYear: this.selectedYear
});
}
/**
* @Param: nowDateTime, UTC
* @Example: UTC => +8: "Asia/Shanghai", -4: "America/New_York", -7: "America/Los_Angeles"
* @Remark: "Asia/Shanghai" => GMT+8 (CST)
* "America/New_York" => GMT-4 (EDT)
* "America/Los_Angeles" => GMT-7 (PDT)
*/
private GetSpecificTimeZoneFormat(SpecificLocalDateTime: Date, UTC: number): any[] {
let YMDDateArray: any[] = []; //need to assign initial value or occur error : cannot access property push of undefined
//Return the number of milliseconds since 1970/01/01
let localTime: number = SpecificLocalDateTime.getTime();
//Return the timezone difference between UTC and Local Time
//By default, this method returns the time zone offset in minutes, so convert this value to milliseconds for easier manipulation:
//Note that a negative return value from getTimezoneOffset() indicates that the current location is ahead of UTC,
//while a positive value indicates that the location is behind UTC. (ex: Taiwan is UTC+8 => 8 * 60 = -480)
//Obtain local UTC offset and convert to msec
let localOffset: number = SpecificLocalDateTime.getTimezoneOffset() * 60000;
//Obtain UTC time in msec
let utc = localTime + localOffset;
// Obtain and add destination's UTC time offset. For example, America/Los_Angeles which is UTC - 7 hours
let offset = UTC;
//Note: In case you're wondering how I arrived at 3600000 as the multiplication factor, remember that
//1000 millseconds = 1 second, and 1 hour = 3600 seconds. Therefore, converting hours to milliseconds
//involves multiplying by 3600 * 1000 = 3600000.
let timezone = utc + (3600000 * offset);
let anotherTimeZoneTime = new Date(timezone);
let nowYear: string = anotherTimeZoneTime.getFullYear().toString();
//Months are zero based
let nowMonth: string = (anotherTimeZoneTime.getMonth() + 1) < 10 ? "0" + (anotherTimeZoneTime.getMonth() + 1).toString() : (anotherTimeZoneTime.getMonth() + 1).toString();
let nowDate: string = anotherTimeZoneTime.getDate() < 10 ? "0" + anotherTimeZoneTime.getDate().toString() : anotherTimeZoneTime.getDate().toString();
//let nowHour: string = anotherTimeZoneTime.getHours() < 10 ? "0" + anotherTimeZoneTime.getHours().toString() : anotherTimeZoneTime.getHours().toString();
//let nowMinute: string = anotherTimeZoneTime.getMinutes().toString();
//let nowSecond: string = anotherTimeZoneTime.getSeconds().toString();
YMDDateArray.push(nowYear, nowMonth, nowDate);
return YMDDateArray;
//[0]:year, [1]:month, [2]:day
}
private GetNBAGameList(GameDate: string): Promise<any> {
return this.NBAdataservices.GetGameGeneral(GameDate)
.then(gameGeneral => {
//before everytime we get data, we should clear the former game data
this.NBAGameList.length = 0;
let unstartArray: any[] = gameGeneral['unstart'];
let liveArray: any[] = gameGeneral['live'];
let overArray: any[] = gameGeneral['over'];
this.gameCount = unstartArray.length + liveArray.length + overArray.length;
if(this.gameCount !== 0) {
//The concat() method is used to join two or more arrays.
let totalArray: any[] = unstartArray.concat(liveArray, overArray);
totalArray.forEach(EachGameitem => {
let GameID: string = EachGameitem['gameID'];
let GameDate: string = EachGameitem['gameDate'];
let HomeTeamID: string = EachGameitem['home']['TeamID'];
let HomeAbbr: string = EachGameitem['home']['TeamAbbr'];
let VisitorTeamID: string = EachGameitem['visitor']['TeamID'];
let VisitorAbbr: string = EachGameitem['visitor']['TeamAbbr'];
let GameProcess: string;
let GameType: string = EachGameitem['GameProcessType'];
switch(GameType) {
case 'unstart': | GameProcess = EachGameitem['GameDate'].replace(/\s*ET\s*/, '');
break;
case 'live':
GameProcess = EachGameitem['process']['Quarter'] + ' ';
GameProcess += EachGameitem['process']['Time'].replace(/\s+/, '');
break;
case 'over':
GameProcess = 'Final';
break;
default:
return;
}
let HomeScore: number = EachGameitem['home']['Score'];
let VisitorScore: number = EachGameitem['visitor']['Score'];
//Get City, Team Name, Logo, Color from TeamMap services based on Abbreviation, so use Array.filter to get match result
//Only return an array that count is 1 because Abbreviation is unique.
let HomeMatchAbbrArray: any[] = this.NBATeamMapData.filter(data => data['Abbreviation'] === HomeAbbr);
let VisitorMatchAbbrArray: any[] = this.NBATeamMapData.filter(data => data['Abbreviation'] === VisitorAbbr);
let HomeTeamColor: string = HomeMatchAbbrArray[0]['Color'];
let HomeTeamCity: string = HomeMatchAbbrArray[0]['City'];
let VisitorTeamCity: string = VisitorMatchAbbrArray[0]['City'];
let HomeTeamName: string = HomeMatchAbbrArray[0]['Team'];
let VisitorTeamName: string = VisitorMatchAbbrArray[0]['Team'];
let HomeTeamLogo: string = HomeMatchAbbrArray[0]['imgGameUrl'];
let VisitorTeamLogo: string = VisitorMatchAbbrArray[0]['imgGameUrl'];
//Finally, we create an array contains all information we want to show in html, and use this array
//as the datasource of ion-card, which use *ngFor to produce game list | random_line_split |
|
ddglCtrl.js | ', function() {
$("input[name='orderid-query']").val('');
$("input[name='sender-query']").val('');
$("input[name='receiver-query']").val('');
$("input[name='username-query']").val('');
$("select[name='state-query']").val('');
$("select[name='iscreditapply-query']").val('');
$('.setTableLength').val(10);
$("#date-start").val('');
$("#date-end").val('');
});
$(".setTableLength").change(function() {
reload = true;
showOrderByPage(1);
});
$(".searchBtn").click(function() {
reload = true;
showOrderByPage(1);
});
$('.clearBtn').on('click', function() {
$("input[name='loginname-query']").val('');
$("select[name='status-query']").val('');
$("#date-start").val('');
$("#date-end").val('');
$("input[name='vipname-nickName']").val('');
$("select[name='status-query']").val('');
$("select[name='viprole-query']").val('');
});
$.validator.addMethod("isPhone", function(value, element) {
var length = value.length;
return this.optional(element) || (length == 11 && /^(((13[0-9]{1})|(17[0-9]{1})|(15[0-9]{1})|(18[0-9]{1}))+\d{8})$/.test(value))||/^((0\d{2,3})-)?(\d{7,8})(-(\d{3,}))?$/.test(value);
}, "手机、电话格式不正确");
/** 添加验证start **/
$('#addStaffForm').validate({
rules: {
'nickname': {
required: true
},
'loginname': {
required: true
},
'phone':{
isPhone:true
}
},
messages: {
'nickname': {
required: "请填写姓名"
},
'loginname': {
required: "请填写登录帐号"
},
'phone': { }
},
errorClass: "error",
success: 'valid',
unhighlight: function(element, errorClass, validClass) { //验证通过
$(element).tooltip('destroy').removeClass(errorClass);
},
errorPlacement: function(label, element) {
$(element).tooltip('destroy'); /*必需*/
$(element).attr('title', $(label).text()).tooltip('show');
},
onclick: true
});
/** 添加验证end **/
/** 清空添加数据 start **/
$('.add').click(function(e) {
$(".errorMsg").html("");
$("#addStaffForm").find("input[name='nickname']").val('');
$("#addStaffForm").find("input[name='loginname']").val('');
$("#addStaffForm").find("select[name='role']").val('管理员');
$("#addStaffForm").find("input[name='phone']").val('');
})
/** 清空添加数据 end **/
/*添加start **/
$('#btn_addstaff').click(function(e) {
var isValid = addStaffForm.valid();
if (isValid) {
$.post(ctx+"/staff/savestaff", $('#addStaffForm').serialize(),function (data) {
var result = data.result;
if(result!=null && result=="noauthority") {
common.alert1('你无此权限!');
return;
}
if(data.result!=null && data.result=="againLogin") {
$('.commonConfirm').modal('hide');
common.alert1("登录过期,请重新登录");
setTimeout("location.href = ctx + '/login.html'",1000);
}else if(data.result=='success'){
$('.modal-staff').modal('hide');
$('.mySuccessModal1').modal({show:true,backdrop: 'static', keyboard: false});
common.mySuccessModal1('添加成功!','如需继续添加,请点击返回继续添加',
'返回继续添加','前往后台用户列表',
function () {
$('.mySuccessModal1').modal('hide');
$('.modal-order').modal('hide');
showOrderByPage(1);
},
function () {
$('.modal-staff').modal('show');
});
$("#addStaffForm").find("input[name='nickname']").val('');
$("#addStaffForm").find("input[name='loginname']").val('');
$("#addStaffForm").find("select[name='role']").val('管理员');
$("#addStaffForm").find("input[name='phone']").val('');
}else if(data.result=='fail'){
$(".errorMsg").html(data.message);
}else{
$(".errorMsg").html("信息提交错误");
}
},"json");
}
});
/** 添加end **/
/** 修改验证start **/
$('#updateStaffForm').validate({
rules: {
'nickname': {
required: true
},
'phone':{
isPhone:true
}
},
messages: {
'nickname': {
required: "请填写姓名"
},
'phone': {
}
},
errorClass: "error",
success: 'valid',
unhighlight: function(element, errorClass, validClass) { //验证通过
$(element).tooltip('destroy').removeClass(errorClass);
},
errorPlacement: function(label, element) {
$(element).tooltip('destroy'); /*必需*/
$(element).attr('title', $(label).text()).tooltip('show');
},
onclick: true
});
/** 修改验证end **/
/** 修改start **/
$("#btn_updatestaff").click(function() {
var isValid = updateStaffForm.valid();
if (isValid) {
$.post(ctx+"/staff/updatestaff", $('#updateStaffForm').serialize(),function (data) {
var result = data.result;
if(result!=null && result=="noauthority") {
common.alert1('你无此权限!');
return;
}
if(data.result!=null && data.result=="againLogin") {
$('.commonConfirm').modal('hide');
common.alert1("登录过期,请重新登录");;
setTimeout("location.href = ctx + '/login.html'",1000);
}else if(data.result=='success'){
$(".modal-view").modal('hide');
common.mySuccessModal2('修改成功!');
showOrderByPage(1);
}else{
$(".updateMsg").html(data.message);
}
},"json");
}
});
/** 修改end **/
function showOrderByPage(pageNum) {
if ($(".all").is(':checked')) {
$(".all").removeAttr('checked');
}
$("#dataBody").append(loadingHtml);
if($('.alert-warning').length>0){
$('.alert-warning').remove();
}
$.getJSON(ctx + '/adminorder/showorderbypage', {
pageNum : pageNum,
pageSize : $('.setTableLength').val() || 10,
startTime : $("#date-start").val(),
endTime : $("#date-end").val(),
orderid : $("input[name='orderid-query']").val(),
sender : $("input[name='sender-query']").val(),
receiver : $("input[name='receiver-query']").val(),
state : $("select[name='state-query").val(),
iscreditapply : $("select[name='iscreditapply-query").val(),
username : $("input[name='username-query']").val(),
},
function(data){
if($('.loading')){
$('.loading').remove();
}
var sumnum = 0;
var sumweight = 0.0;
var sumvalume = 0.0;
var sumorderfee = 0.0;
var sumagencyfund = 0.0;
var result = data.result;
if(result!=null && result=="noauthority") {
common.alert1('你无此权限!');
return;
}
if(data.result!=null && data.result=="againLogin") {
common.alert1("登录过期,请重新登录");
setTimeout("location.href = ctx + '/login.html'",1000);
} else if(data.result.length == 0) {
$('.table-responsive').append('<div class="alert alert-warning no-result text-center" role="alert">没有找到记录... </div>');
} else{
var datapage = data.result;
var datalist = datapage.result;
var totalCount = datapage.totalCount;
var pageNum = datapage.pageNo;
var pageSize = datapage.pageSize;
var totalPages = datapage.totalPages;
$(".totalCount").html(totalCount);
$(".totalPages").html(totalPages);
$(".pageNo").html(pageNum);
var tempHtml="";
var currentorderid = $('#orderid').val();
var currentUtype = $('#currentUtype').val(); | );
}
//placeholder兼容ie end
//清除按钮
$('.clearBtn').on('click | identifier_body |
|
ddglCtrl.js | ender : $("input[name='sender-query']").val(),
receiver : $("input[name='receiver-query']").val(),
state : $("select[name='state-query").val(),
iscreditapply : $("select[name='iscreditapply-query").val(),
username : $("input[name='username-query']").val(),
},
function(data){
if($('.loading')){
$('.loading').remove();
}
var sumnum = 0;
var sumweight = 0.0;
var sumvalume = 0.0;
var sumorderfee = 0.0;
var sumagencyfund = 0.0;
var result = data.result;
if(result!=null && result=="noauthority") {
common.alert1('你无此权限!');
return;
}
if(data.result!=null && data.result=="againLogin") {
common.alert1("登录过期,请重新登录");
setTimeout("location.href = ctx + '/login.html'",1000);
} else if(data.result.length == 0) {
$('.table-responsive').append('<div class="alert alert-warning no-result text-center" role="alert">没有找到记录... </div>');
} else{
var datapage = data.result;
var datalist = datapage.result;
var totalCount = datapage.totalCount;
var pageNum = datapage.pageNo;
var pageSize = datapage.pageSize;
var totalPages = datapage.totalPages;
$(".totalCount").html(totalCount);
$(".totalPages").html(totalPages);
$(".pageNo").html(pageNum);
var tempHtml="";
var currentorderid = $('#orderid').val();
var currentUtype = $('#currentUtype').val();
for(var i=0;i<datalist.length;i++){
var order = datalist[i];
var no = (pageNum-1)*pageSize+i+1;
tempHtml +="<tr data-id='"+order.id+"' data-pageno='" + pageNum + "'>";
// tempHtml +="<td><div class='checkbox'><label><input type='checkbox' data-orderid="+order.id+" /></label></div></td>";
tempHtml +="<td class='excel'>"+no+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.orderid)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.username)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.sender)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.senderphone)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.receiver)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.receiverphone)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.goodsname)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.goodsnum)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.weight)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.volume)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.transportway)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.sendfee)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.orgname)+"</td>";
tempHtml +="<td class='excel'>"+isnull(deliverytype[order.deliverytype])+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.pickupgoodstime)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.settletype)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.agencyfund)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.state)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.createtime)+"</td>";
tempHtml +="<td><button type='button' data-toggle='modal' data-target='.modal-edityd' onclick='showOrderInfo(" +order.id+")' class='btn btn-primary btn-xs'>查看</button> </td>";
tempHtml +="</tr>";
sumnum += order.goodsnum;
sumweight += order.weight;
sumvalume += order.volume;
sumorderfee += order.sendfee;
sumagencyfund += order.agencyfund;
if($('.alert-warning').length>0){
$('.alert-warning').remove();
}
}
var sl=parseFloat(sumvalume).toFixed(2);
var sf=parseFloat(sumorderfee).toFixed(2);
tempHtml +="<tr>" +
"<td style='padding: 6px 5px 3px;'>总计</td> <td></td> <td></td> <td></td> <td></td> <td></td> <td></td> <td></td> " +
"<td>"+sumnum+"</td>" +
"<td>"+sumweight+"</td>" +
"<td>"+sl+"</td>" +
"<td></td>" +
"<td>"+sf+"</td>" +
"<td></td> <td></td> <td></td> <td></td>" +
"<td>"+sumagencyfund+"</td>" +
"<td></td> <td></td> <td></td>" +
"</tr>";
if(datalist.length === 0&&$('.alert-warning').length<=0) {
$('.table-responsive').append('<div class="alert alert-warning no-result text-center" role="alert">没有找到记录 </div>');
}
$('#dataBody').html(tempHtml);
if (reload) {
reload = false;
first_load = true;
var initPagination = function() {
$("#Pagination").pagination(totalCount, {
num_edge_entries : 1, // 边缘页数
num_display_entries : 4, // 主体页数
callback: pageselectCallback,
prev_text : "«",
next_text : "»",
items_per_page : pageSize,
link_to : 'javascript:void(0);'
});
}();
/**分页e**/
}
}
}).fail(function (err) {
if($('.loading')){
$('.loading').remove();
}
common.alert('服务器错误!请刷新页面重试或联系管理员',"确定");
});
}
function pageselectCallback(page_index, jq) {
if (first_load) { //第一次初始化不需要请求
first_load = false;
return false;
}
pageNum = page_index + 1;
showOrderByPage(pageNum);
}
function isnull(value) {
if (value == null || value == '' || typeof (value) == "undefined" || value == '/')
return "--";
else
return value;
}
function isnumber(value) {
if (value == null || value == '')
return "0.00";
else
return value;
}
});
function showOrderInfo(id){
$.getJSON(ctx + '/adminorder/showorderinfo', {
id : id
}, function(data) {
if(data != null){
$(".modal-edityd").find("input[name='orderid']").val(data.orderid);
$(".modal-edityd").find("input[name='orgname']").val(data.orgname);
$(".modal-edityd").find("select[name='transportway']").val(data.transportway);
$(".modal-edityd").find("input[name='createtime']").val(data.createtime);
$(".modal-edityd").find("select[name='settletype']").val(data.settletype);
$(".modal-edityd").find("select[name='ifnotifydispatch']").val(data.ifnotifydispatch);
$(".modal-edityd").find("select[name='paytype']").val(data.paytype | );
$(".modal-edityd").find("select[name='dispatchtype']").val(data.dispatchtype);
$(".modal-edityd").find("input[name='agencyfund']").val(data.agencyfund);
$(".modal-edityd").find("input[name='receiptnum']").val(data.receiptnum);
$(".modal-edityd").find("input[name='receiptno']").val(data.receiptno);
$(".modal-edityd").find("input[name='receiver']").val(data.receiver);
$(".modal-edityd").find("input[name='receiverphone']").val(data.receiverphone);
$(".modal-edityd").find("input[name='sender']").val(data.sender);
$(".modal-edityd").find("input[name='senderphone']").val(data.senderphone);
$(".modal-edityd").find("input[name='destname']").val(data.destname);
$(".modal-edityd").find("input[name='receivercompany']").val(data.receivercompany);
$(".modal-edityd").find("input[name='sendvipname']").val(data.sendvipname);
$(".modal-edityd").find("input[name='sendercompany']").val(data.sendercompany);
$(".modal-edityd").find("input[name='receiverpcdt']").val(data.receiverpcdt);
$(".modal-edityd").find("input[name='senderpcdt']").val(data.senderpcdt);
| conditional_block |
|
ddglCtrl.js | ]{1}))+\d{8})$/.test(value))||/^((0\d{2,3})-)?(\d{7,8})(-(\d{3,}))?$/.test(value);
}, "手机、电话格式不正确");
/** 添加验证start **/
$('#addStaffForm').validate({
rules: {
'nickname': {
required: true
},
'loginname': {
required: true
},
'phone':{
isPhone:true
}
},
messages: {
'nickname': {
required: "请填写姓名"
},
'loginname': {
required: "请填写登录帐号"
},
'phone': { }
},
errorClass: "error",
success: 'valid',
unhighlight: function(element, errorClass, validClass) { //验证通过
$(element).tooltip('destroy').removeClass(errorClass);
},
errorPlacement: function(label, element) {
$(element).tooltip('destroy'); /*必需*/
$(element).attr('title', $(label).text()).tooltip('show');
},
onclick: true
});
/** 添加验证end **/
/** 清空添加数据 start **/
$('.add').click(function(e) {
$(".errorMsg").html("");
$("#addStaffForm").find("input[name='nickname']").val('');
$("#addStaffForm").find("input[name='loginname']").val('');
$("#addStaffForm").find("select[name='role']").val('管理员');
$("#addStaffForm").find("input[name='phone']").val('');
})
/** 清空添加数据 end **/
/*添加start **/
$('#btn_addstaff').click(function(e) {
var isValid = addStaffForm.valid();
if (isValid) {
$.post(ctx+"/staff/savestaff", $('#addStaffForm').serialize(),function (data) {
var result = data.result;
if(result!=null && result=="noauthority") {
common.alert1('你无此权限!');
return;
}
if(data.result!=null && data.result=="againLogin") {
$('.commonConfirm').modal('hide');
common.alert1("登录过期,请重新登录");
setTimeout("location.href = ctx + '/login.html'",1000);
}else if(data.result=='success'){
$('.modal-staff').modal('hide');
$('.mySuccessModal1').modal({show:true,backdrop: 'static', keyboard: false});
common.mySuccessModal1('添加成功!','如需继续添加,请点击返回继续添加',
'返回继续添加','前往后台用户列表',
function () {
$('.mySuccessModal1').modal('hide');
$('.modal-order').modal('hide');
showOrderByPage(1);
},
function () {
$('.modal-staff').modal('show');
});
$("#addStaffForm").find("input[name='nickname']").val('');
$("#addStaffForm").find("input[name='loginname']").val('');
$("#addStaffForm").find("select[name='role']").val('管理员');
$("#addStaffForm").find("input[name='phone']").val('');
}else if(data.result=='fail'){
$(".errorMsg").html(data.message);
}else{
$(".errorMsg").html("信息提交错误");
}
},"json");
}
});
/** 添加end **/
/** 修改验证start **/
$('#updateStaffForm').validate({
rules: {
'nickname': {
required: true
},
'phone':{
isPhone:true
}
},
messages: {
'nickname': {
required: "请填写姓名"
},
'phone': {
}
},
errorClass: "error",
success: 'valid',
unhighlight: function(element, errorClass, validClass) { //验证通过
$(element).tooltip('destroy').removeClass(errorClass);
},
errorPlacement: function(label, element) {
$(element).tooltip('destroy'); /*必需*/
$(element).attr('title', $(label).text()).tooltip('show');
},
onclick: true
});
/** 修改验证end **/
/** 修改start **/
$("#btn_updatestaff").click(function() {
var isValid = updateStaffForm.valid();
if (isValid) {
$.post(ctx+"/staff/updatestaff", $('#updateStaffForm').serialize(),function (data) {
var result = data.result;
if(result!=null && result=="noauthority") {
common.alert1('你无此权限!');
return;
}
if(data.result!=null && data.result=="againLogin") {
$('.commonConfirm').modal('hide');
common.alert1("登录过期,请重新登录");;
setTimeout("location.href = ctx + '/login.html'",1000);
}else if(data.result=='success'){
$(".modal-view").modal('hide');
common.mySuccessModal2('修改成功!');
showOrderByPage(1);
}else{
$(".updateMsg").html(data.message);
}
},"json");
}
});
/** 修改end **/
function showOrderByPage(pageNum) {
if ($(".all").is(':checked')) {
$(".all").removeAttr('checked');
}
$("#dataBody").append(loadingHtml);
if($('.alert-warning').length>0){
$('.alert-warning').remove();
}
$.getJSON(ctx + '/adminorder/showorderbypage', {
pageNum : pageNum,
pageSize : $('.setTableLength').val() || 10,
startTime : $("#date-start").val(),
endTime : $("#date-end").val(),
orderid : $("input[name='orderid-query']").val(),
sender : $("input[name='sender-query']").val(),
receiver : $("input[name='receiver-query']").val(),
state : $("select[name='state-query").val(),
iscreditapply : $("select[name='iscreditapply-query").val(),
username : $("input[name='username-query']").val(),
},
function(data){
if($('.loading')){ | }
var sumnum = 0;
var sumweight = 0.0;
var sumvalume = 0.0;
var sumorderfee = 0.0;
var sumagencyfund = 0.0;
var result = data.result;
if(result!=null && result=="noauthority") {
common.alert1('你无此权限!');
return;
}
if(data.result!=null && data.result=="againLogin") {
common.alert1("登录过期,请重新登录");
setTimeout("location.href = ctx + '/login.html'",1000);
} else if(data.result.length == 0) {
$('.table-responsive').append('<div class="alert alert-warning no-result text-center" role="alert">没有找到记录... </div>');
} else{
var datapage = data.result;
var datalist = datapage.result;
var totalCount = datapage.totalCount;
var pageNum = datapage.pageNo;
var pageSize = datapage.pageSize;
var totalPages = datapage.totalPages;
$(".totalCount").html(totalCount);
$(".totalPages").html(totalPages);
$(".pageNo").html(pageNum);
var tempHtml="";
var currentorderid = $('#orderid').val();
var currentUtype = $('#currentUtype').val();
for(var i=0;i<datalist.length;i++){
var order = datalist[i];
var no = (pageNum-1)*pageSize+i+1;
tempHtml +="<tr data-id='"+order.id+"' data-pageno='" + pageNum + "'>";
// tempHtml +="<td><div class='checkbox'><label><input type='checkbox' data-orderid="+order.id+" /></label></div></td>";
tempHtml +="<td class='excel'>"+no+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.orderid)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.username)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.sender)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.senderphone)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.receiver)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.receiverphone)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.goodsname)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.goodsnum)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.weight)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.volume)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.transportway)+"</td>";
tempHtml +="<td class='excel'>"+isnull(order.sendfee)+"</td>";
tempHtml +="<td class='excel'>"+is | $('.loading').remove(); | random_line_split |
ddglCtrl.js | ut');
}
//placeholder兼容ie end
//清除按钮
$('.clearBtn').on('click', function() {
$("input[name='orderid-query']").val('');
$("input[name='sender-query']").val('');
$("input[name='receiver-query']").val('');
$("input[name='username-query']").val('');
$("select[name='state-query']").val('');
$("select[name='iscreditapply-query']").val('');
$('.setTableLength').val(10);
$("#date-start").val('');
$("#date-end").val('');
});
$(".setTableLength").change(function() {
reload = true;
showOrderByPage(1);
});
$(".searchBtn").click(function() {
reload = true;
showOrderByPage(1);
});
$('.clearBtn').on('click', function() {
$("input[name='loginname-query']").val('');
$("select[name='status-query']").val('');
$("#date-start").val('');
$("#date-end").val('');
$("input[name='vipname-nickName']").val('');
$("select[name='status-query']").val('');
$("select[name='viprole-query']").val('');
});
$.validator.addMethod("isPhone", function(value, element) {
var length = value.length;
return this.optional(element) || (length == 11 && /^(((13[0-9]{1})|(17[0-9]{1})|(15[0-9]{1})|(18[0-9]{1}))+\d{8})$/.test(value))||/^((0\d{2,3})-)?(\d{7,8})(-(\d{3,}))?$/.test(value);
}, "手机、电话格式不正确");
/** 添加验证start **/
$('#addStaffForm').validate({
rules: {
'nickname': {
required: true
},
'loginname': {
required: true
},
'phone':{
isPhone:true
}
},
messages: {
'nickname': {
required: "请填写姓名"
},
'loginname': {
required: "请填写登录帐号"
},
'phone': { }
},
errorClass: "error",
success: 'valid',
unhighlight: function(element, errorClass, validClass) { //验证通过
$(element).tooltip('destroy').removeClass(errorClass);
},
errorPlacement: function(label, element) {
$(element).tooltip('destroy'); /*必需*/
$(element).attr('title', $(label).text()).tooltip('show');
},
onclick: true
});
/** 添加验证end **/
/** 清空添加数据 start **/
$('.add').click(function(e) {
$(".errorMsg").html("");
$("#addStaffForm").find("input[name='nickname']").val('');
$("#addStaffForm").find("input[name='loginname']").val('');
$("#addStaffForm").find("select[name='role']").val('管理员');
$("#addStaffForm").find("input[name='phone']").val('');
})
/** 清空添加数据 end **/
/*添加start **/
$('#btn_addstaff').click(function(e) {
var isValid = addStaffForm.valid();
if (isValid) {
$.post(ctx+"/staff/savestaff", $('#addStaffForm').serialize(),function (data) {
var result = data.result;
if(result!=null && result=="noauthority") {
common.alert1('你无此权限!');
return;
}
if(data.result!=null && data.result=="againLogin") {
$('.commonConfirm').modal('hide');
common.alert1("登录过期,请重新登录");
setTimeout("location.href = ctx + '/login.html'",1000);
}else if(data.result=='success'){
$('.modal-staff').modal('hide');
$('.mySuccessModal1').modal({show:true,backdrop: 'static', keyboard: false});
common.mySuccessModal1('添加成功!','如需继续添加,请点击返回继续添加',
'返回继续添加','前往后台用户列表',
function () {
$('.mySuccessModal1').modal('hide');
$('.modal-order').modal('hide');
showOrderByPage(1);
},
function () {
$('.modal-staff').modal('show');
});
$("#addStaffForm").find("input[name='nickname']").val('');
$("#addStaffForm").find("input[name='loginname']").val('');
$("#addStaffForm").find("select[name='role']").val('管理员');
$("#addStaffForm").find("input[name='phone']").val('');
}else if(data.result=='fail'){
$(".errorMsg").html(data.message);
}else{
$(".errorMsg").html("信息提交错误");
}
},"json");
}
});
/** 添加end **/
/** 修改验证start **/
$('#updateStaffForm').validate({
rules: {
'nickname': {
required: true
},
'phone':{
isPhone:true
}
},
messages: {
'nickname': {
required: "请填写姓名"
},
'phone': {
}
},
errorClass: "error",
success: 'valid',
unhighlight: function(element, errorClass, validClass) { //验证通过
$(element).tooltip('destroy').removeClass(errorClass);
},
errorPlacement: function(label, element) {
$(element).tooltip('destroy'); /*必需*/
$(element).attr('title', $(label).text()).tooltip('show');
},
onclick: true
});
/** 修改验证end **/
/** 修改start **/
$("#btn_updatestaff").click(function() {
var isValid = updateStaffForm.valid();
if (isValid) {
$.post(ctx+"/staff/updatestaff", $('#updateStaffForm').serialize(),function (data) {
var result = data.result;
if(result!=null && result=="noauthority") {
common.alert1('你无此权限!');
return;
}
if(data.result!=null && data.result=="againLogin") {
$('.commonConfirm').modal('hide');
common.alert1("登录过期,请重新登录");;
setTimeout("location.href = ctx + '/login.html'",1000);
}else if(data.result=='success'){
$(".modal-view").modal('hide');
common.mySuccessModal2('修改成功!');
showOrderByPage(1);
}else{
$(".updateMsg").html(data.message);
}
},"json");
}
});
/** 修改end **/
function showOrderByPage(pageNum) {
if ($(".all").is(':checked')) {
$(".all").removeAttr('checked');
}
$("#dataBody").append(loadingHtml);
if($('.alert-warning').length>0){
$('.alert-warning').remove();
}
$.getJSON(ctx + '/adminorder/showorderbypage', {
pageNum : pageNum,
pageSize : $('.setTableLength').val() || 10,
startTime : $("#date-start").val(),
endTime : $("#date-end").val(),
orderid : $("input[name='orderid-query']").val(),
sender : $("input[name='sender-query']").val(),
receiver : $("input[name='receiver-query']").val(),
state : $("select[name='state-query").val(),
iscreditapply : $("select[name='iscreditapply-query").val(),
username : $("input[name='username-query']").val(),
},
function(data){
if($('.loading')){
$('.loading').remove();
}
var sumnum = 0;
var sumweight = 0.0;
var sumvalume = 0.0;
var sumorderfee = 0.0;
var sumagencyfund = 0.0;
var result = data.result;
if(result!=null && result=="noauthority") {
common.alert1('你无此权限!');
return;
}
if(data.result!=null && data.result=="againLogin") {
common.alert1("登录过期,请重新登录");
setTimeout("location.href = ctx + '/login.html'",1000);
} else if(data.result.length == 0) {
$('.table-responsive').append('<div class="alert alert-warning no-result text-center" role="alert">没有找到记录... </div>');
} else{
var datapage = data.result;
var datalist = datapage.result;
var totalCount = datapage.totalCount;
var pageNum = datapage.pageNo;
var pageSize = datapage.pageSize;
var totalPages = datapage.totalPages;
$(".totalCount").html(totalCount);
$(".totalPages").html(totalPages);
$(".pageNo").html(pageNum);
var tempHtml="";
var currentorderid = $('#orderid').val();
var currentUtype = $('#currentU | createElement('inp | identifier_name |
|
SiteInfoDialog.py | , Qt
from PyQt5.QtGui import QPixmap, QImage, QPainter, QColor, QBrush
from PyQt5.QtNetwork import QNetworkRequest, QNetworkReply
from PyQt5.QtWidgets import (
QDialog, QTreeWidgetItem, QGraphicsScene, QMenu, QApplication,
QGraphicsPixmapItem
)
from E5Gui import E5MessageBox, E5FileDialog
from .Ui_SiteInfoDialog import Ui_SiteInfoDialog
from ..Tools import Scripts, WebBrowserTools
from ..WebBrowserPage import WebBrowserPage
import UI.PixmapCache
class SiteInfoDialog(QDialog, Ui_SiteInfoDialog):
"""
Class implementing a dialog to show some information about a site.
"""
okStyle = "QLabel { color : white; background-color : green; }"
nokStyle = "QLabel { color : white; background-color : red; }"
def __init__(self, browser, parent=None):
"""
Constructor
@param browser reference to the browser window (HelpBrowser)
@param parent reference to the parent widget (QWidget)
"""
super(SiteInfoDialog, self).__init__(parent)
self.setupUi(self)
self.setWindowFlags(Qt.Window)
# put icons
self.tabWidget.setTabIcon(
0, UI.PixmapCache.getIcon("siteinfo-general.png"))
self.tabWidget.setTabIcon(
1, UI.PixmapCache.getIcon("siteinfo-media.png"))
self.__imageReply = None
self.__baseUrl = browser.url()
title = browser.title()
#prepare background of image preview
self.__imagePreviewStandardBackground = (
self.imagePreview.backgroundBrush()
)
color1 = QColor(220, 220, 220)
color2 = QColor(160, 160, 160)
self.__tilePixmap = QPixmap(8, 8)
self.__tilePixmap.fill(color1)
tilePainter = QPainter(self.__tilePixmap)
tilePainter.fillRect(0, 0, 4, 4, color2)
tilePainter.fillRect(4, 4, 4, 4, color2)
tilePainter.end()
# populate General tab
self.heading.setText("<b>{0}</b>".format(title))
self.siteAddressLabel.setText(self.__baseUrl.toString())
if self.__baseUrl.scheme() in ["https"]:
self.securityLabel.setStyleSheet(SiteInfoDialog.okStyle)
self.securityLabel.setText(
self.tr('<b>Connection is encrypted.</b>'))
else:
self.securityLabel.setStyleSheet(SiteInfoDialog.nokStyle)
self.securityLabel.setText(
self.tr('<b>Connection is not encrypted.</b>'))
browser.page().runJavaScript(
"document.charset", WebBrowserPage.SafeJsWorld,
lambda res: self.encodingLabel.setText(res))
# populate Meta tags
browser.page().runJavaScript(Scripts.getAllMetaAttributes(),
WebBrowserPage.SafeJsWorld,
self.__processMetaAttributes)
# populate Media tab
browser.page().runJavaScript(Scripts.getAllImages(),
WebBrowserPage.SafeJsWorld,
self.__processImageTags)
def __processImageTags(self, res):
"""
Private method to process the image tags.
@param res result of the JavaScript script
@type list of dict
"""
for img in res:
src = img["src"]
alt = img["alt"]
if not alt:
if src.find("/") == -1:
alt = src
else:
pos = src.rfind("/")
alt = src[pos + 1:]
if not src or not alt:
continue
QTreeWidgetItem(self.imagesTree, [alt, src])
for col in range(self.imagesTree.columnCount()):
self.imagesTree.resizeColumnToContents(col)
if self.imagesTree.columnWidth(0) > 300:
self.imagesTree.setColumnWidth(0, 300)
self.imagesTree.setCurrentItem(self.imagesTree.topLevelItem(0))
self.imagesTree.setContextMenuPolicy(Qt.CustomContextMenu)
self.imagesTree.customContextMenuRequested.connect(
self.__imagesTreeContextMenuRequested)
def __processMetaAttributes(self, res):
"""
Private method to process the meta attributes.
@param res result of the JavaScript script
@type list of dict
"""
for meta in res:
content = meta["content"]
name = meta["name"]
if not name:
name = meta["httpequiv"]
if not name or not content:
continue
if meta["charset"]:
self.encodingLabel.setText(meta["charset"])
if "charset=" in content:
self.encodingLabel.setText(
content[content.index("charset=") + 8:])
QTreeWidgetItem(self.tagsTree, [name, content])
for col in range(self.tagsTree.columnCount()):
|
@pyqtSlot(QTreeWidgetItem, QTreeWidgetItem)
def on_imagesTree_currentItemChanged(self, current, previous):
"""
Private slot to show a preview of the selected image.
@param current current image entry (QTreeWidgetItem)
@param previous old current entry (QTreeWidgetItem)
"""
if current is None:
return
imageUrl = QUrl(current.text(1))
if imageUrl.isRelative():
imageUrl = self.__baseUrl.resolved(imageUrl)
pixmap = QPixmap()
loading = False
if imageUrl.scheme() == "data":
encodedUrl = current.text(1).encode("utf-8")
imageData = encodedUrl[encodedUrl.find(b",") + 1:]
pixmap = WebBrowserTools.pixmapFromByteArray(imageData)
elif imageUrl.scheme() == "file":
pixmap = QPixmap(imageUrl.toLocalFile())
elif imageUrl.scheme() == "qrc":
pixmap = QPixmap(imageUrl.toString()[3:])
else:
if self.__imageReply is not None:
self.__imageReply.deleteLater()
self.__imageReply = None
from WebBrowser.WebBrowserWindow import WebBrowserWindow
self.__imageReply = WebBrowserWindow.networkManager().get(
QNetworkRequest(imageUrl))
self.__imageReply.finished.connect(self.__imageReplyFinished)
loading = True
self.__showLoadingText()
if not loading:
self.__showPixmap(pixmap)
@pyqtSlot()
def __imageReplyFinished(self):
"""
Private slot handling the loading of an image.
"""
if self.__imageReply.error() != QNetworkReply.NoError:
return
data = self.__imageReply.readAll()
self.__showPixmap(QPixmap.fromImage(QImage.fromData(data)))
def __showPixmap(self, pixmap):
"""
Private method to show a pixmap in the preview pane.
@param pixmap pixmap to be shown
@type QPixmap
"""
scene = QGraphicsScene(self.imagePreview)
if pixmap.isNull():
self.imagePreview.setBackgroundBrush(
self.__imagePreviewStandardBackground)
scene.addText(self.tr("Preview not available."))
else:
self.imagePreview.setBackgroundBrush(QBrush(self.__tilePixmap))
scene.addPixmap(pixmap)
self.imagePreview.setScene(scene)
def __showLoadingText(self):
"""
Private method to show some text while loading an image.
"""
self.imagePreview.setBackgroundBrush(
self.__imagePreviewStandardBackground)
scene = QGraphicsScene(self.imagePreview)
scene.addText(self.tr("Loading..."))
self.imagePreview.setScene(scene)
def __imagesTreeContextMenuRequested(self, pos):
"""
Private slot to show a context menu for the images list.
@param pos position for the menu (QPoint)
"""
itm = self.imagesTree.itemAt(pos)
if itm is None:
return
menu = QMenu()
act = menu.addAction(self.tr("Copy Image Location to Clipboard"))
act.setData(itm.text(1))
act.triggered.connect(lambda: self.__copyAction(act))
act = menu.addAction(self.tr("Copy Image Name to Clipboard"))
act.setData(itm.text(0))
act.triggered.connect(lambda: self.__copyAction(act))
menu.addSeparator()
act = menu.addAction(self.tr("Save Image"))
act.setData(self.imagesTree.indexOfTopLevelItem(itm))
act.triggered.connect(lambda: self.__saveImage(act))
menu.exec_(self.imagesTree.viewport().mapToGlobal(pos))
def __copyAction(self, act):
"""
Private slot to copy the image URL or the image name to the clipboard.
@param act reference to the action that triggered
@type QAction
"""
QApplication.clipboard().setText(act.data())
def __saveImage(self, act):
"""
Private slot to save the selected image to disk.
@param act reference to the action that triggered
@type QAction
"""
index = act.data()
itm = self.imagesTree.topLevelItem(index)
if itm is None:
return
if (
not self.imagePreview.scene() or
len(self.imagePreview.scene().items()) == 0
):
return
pixmapItem = self.imagePreview.scene().items()[0]
if not isinstance(pixmapItem, QGraphicsPixmapItem):
return
if pixmapItem.pixmap().isNull():
E5MessageBox.warning(
self,
self.tr("Save Image"),
self.tr(
"""<p>This preview is not available.</p>"""))
return
imageFileName = WebBrowserTools.getFileName | self.tagsTree.resizeColumnToContents(col) | conditional_block |
SiteInfoDialog.py | , Qt
from PyQt5.QtGui import QPixmap, QImage, QPainter, QColor, QBrush
from PyQt5.QtNetwork import QNetworkRequest, QNetworkReply
from PyQt5.QtWidgets import (
QDialog, QTreeWidgetItem, QGraphicsScene, QMenu, QApplication,
QGraphicsPixmapItem
)
from E5Gui import E5MessageBox, E5FileDialog
from .Ui_SiteInfoDialog import Ui_SiteInfoDialog
from ..Tools import Scripts, WebBrowserTools
from ..WebBrowserPage import WebBrowserPage
import UI.PixmapCache
class SiteInfoDialog(QDialog, Ui_SiteInfoDialog):
"""
Class implementing a dialog to show some information about a site.
"""
okStyle = "QLabel { color : white; background-color : green; }"
nokStyle = "QLabel { color : white; background-color : red; }"
def __init__(self, browser, parent=None):
"""
Constructor
@param browser reference to the browser window (HelpBrowser)
@param parent reference to the parent widget (QWidget)
"""
super(SiteInfoDialog, self).__init__(parent)
self.setupUi(self)
self.setWindowFlags(Qt.Window)
# put icons
self.tabWidget.setTabIcon(
0, UI.PixmapCache.getIcon("siteinfo-general.png"))
self.tabWidget.setTabIcon(
1, UI.PixmapCache.getIcon("siteinfo-media.png"))
self.__imageReply = None
self.__baseUrl = browser.url()
title = browser.title()
#prepare background of image preview
self.__imagePreviewStandardBackground = (
self.imagePreview.backgroundBrush()
)
color1 = QColor(220, 220, 220)
color2 = QColor(160, 160, 160)
self.__tilePixmap = QPixmap(8, 8)
self.__tilePixmap.fill(color1)
tilePainter = QPainter(self.__tilePixmap)
tilePainter.fillRect(0, 0, 4, 4, color2)
tilePainter.fillRect(4, 4, 4, 4, color2)
tilePainter.end()
# populate General tab
self.heading.setText("<b>{0}</b>".format(title))
self.siteAddressLabel.setText(self.__baseUrl.toString())
if self.__baseUrl.scheme() in ["https"]:
self.securityLabel.setStyleSheet(SiteInfoDialog.okStyle)
self.securityLabel.setText(
self.tr('<b>Connection is encrypted.</b>'))
else:
self.securityLabel.setStyleSheet(SiteInfoDialog.nokStyle)
self.securityLabel.setText(
self.tr('<b>Connection is not encrypted.</b>'))
browser.page().runJavaScript(
"document.charset", WebBrowserPage.SafeJsWorld,
lambda res: self.encodingLabel.setText(res))
# populate Meta tags
browser.page().runJavaScript(Scripts.getAllMetaAttributes(),
WebBrowserPage.SafeJsWorld,
self.__processMetaAttributes)
# populate Media tab
browser.page().runJavaScript(Scripts.getAllImages(),
WebBrowserPage.SafeJsWorld,
self.__processImageTags)
def __processImageTags(self, res):
"""
Private method to process the image tags.
@param res result of the JavaScript script
@type list of dict
"""
for img in res:
src = img["src"]
alt = img["alt"]
if not alt:
if src.find("/") == -1:
alt = src
else:
pos = src.rfind("/")
alt = src[pos + 1:]
if not src or not alt:
continue
QTreeWidgetItem(self.imagesTree, [alt, src])
for col in range(self.imagesTree.columnCount()):
self.imagesTree.resizeColumnToContents(col)
if self.imagesTree.columnWidth(0) > 300:
self.imagesTree.setColumnWidth(0, 300)
self.imagesTree.setCurrentItem(self.imagesTree.topLevelItem(0))
self.imagesTree.setContextMenuPolicy(Qt.CustomContextMenu)
self.imagesTree.customContextMenuRequested.connect(
self.__imagesTreeContextMenuRequested)
def __processMetaAttributes(self, res):
"""
Private method to process the meta attributes.
@param res result of the JavaScript script
@type list of dict
"""
for meta in res:
content = meta["content"]
name = meta["name"]
if not name:
name = meta["httpequiv"]
if not name or not content:
continue
if meta["charset"]:
self.encodingLabel.setText(meta["charset"])
if "charset=" in content:
self.encodingLabel.setText(
content[content.index("charset=") + 8:])
QTreeWidgetItem(self.tagsTree, [name, content])
for col in range(self.tagsTree.columnCount()):
self.tagsTree.resizeColumnToContents(col)
@pyqtSlot(QTreeWidgetItem, QTreeWidgetItem)
def on_imagesTree_currentItemChanged(self, current, previous):
| elif imageUrl.scheme() == "file":
pixmap = QPixmap(imageUrl.toLocalFile())
elif imageUrl.scheme() == "qrc":
pixmap = QPixmap(imageUrl.toString()[3:])
else:
if self.__imageReply is not None:
self.__imageReply.deleteLater()
self.__imageReply = None
from WebBrowser.WebBrowserWindow import WebBrowserWindow
self.__imageReply = WebBrowserWindow.networkManager().get(
QNetworkRequest(imageUrl))
self.__imageReply.finished.connect(self.__imageReplyFinished)
loading = True
self.__showLoadingText()
if not loading:
self.__showPixmap(pixmap)
@pyqtSlot()
def __imageReplyFinished(self):
"""
Private slot handling the loading of an image.
"""
if self.__imageReply.error() != QNetworkReply.NoError:
return
data = self.__imageReply.readAll()
self.__showPixmap(QPixmap.fromImage(QImage.fromData(data)))
def __showPixmap(self, pixmap):
"""
Private method to show a pixmap in the preview pane.
@param pixmap pixmap to be shown
@type QPixmap
"""
scene = QGraphicsScene(self.imagePreview)
if pixmap.isNull():
self.imagePreview.setBackgroundBrush(
self.__imagePreviewStandardBackground)
scene.addText(self.tr("Preview not available."))
else:
self.imagePreview.setBackgroundBrush(QBrush(self.__tilePixmap))
scene.addPixmap(pixmap)
self.imagePreview.setScene(scene)
def __showLoadingText(self):
"""
Private method to show some text while loading an image.
"""
self.imagePreview.setBackgroundBrush(
self.__imagePreviewStandardBackground)
scene = QGraphicsScene(self.imagePreview)
scene.addText(self.tr("Loading..."))
self.imagePreview.setScene(scene)
def __imagesTreeContextMenuRequested(self, pos):
"""
Private slot to show a context menu for the images list.
@param pos position for the menu (QPoint)
"""
itm = self.imagesTree.itemAt(pos)
if itm is None:
return
menu = QMenu()
act = menu.addAction(self.tr("Copy Image Location to Clipboard"))
act.setData(itm.text(1))
act.triggered.connect(lambda: self.__copyAction(act))
act = menu.addAction(self.tr("Copy Image Name to Clipboard"))
act.setData(itm.text(0))
act.triggered.connect(lambda: self.__copyAction(act))
menu.addSeparator()
act = menu.addAction(self.tr("Save Image"))
act.setData(self.imagesTree.indexOfTopLevelItem(itm))
act.triggered.connect(lambda: self.__saveImage(act))
menu.exec_(self.imagesTree.viewport().mapToGlobal(pos))
def __copyAction(self, act):
"""
Private slot to copy the image URL or the image name to the clipboard.
@param act reference to the action that triggered
@type QAction
"""
QApplication.clipboard().setText(act.data())
def __saveImage(self, act):
"""
Private slot to save the selected image to disk.
@param act reference to the action that triggered
@type QAction
"""
index = act.data()
itm = self.imagesTree.topLevelItem(index)
if itm is None:
return
if (
not self.imagePreview.scene() or
len(self.imagePreview.scene().items()) == 0
):
return
pixmapItem = self.imagePreview.scene().items()[0]
if not isinstance(pixmapItem, QGraphicsPixmapItem):
return
if pixmapItem.pixmap().isNull():
E5MessageBox.warning(
self,
self.tr("Save Image"),
self.tr(
"""<p>This preview is not available.</p>"""))
return
imageFileName = WebBrowserTools.getFileName | """
Private slot to show a preview of the selected image.
@param current current image entry (QTreeWidgetItem)
@param previous old current entry (QTreeWidgetItem)
"""
if current is None:
return
imageUrl = QUrl(current.text(1))
if imageUrl.isRelative():
imageUrl = self.__baseUrl.resolved(imageUrl)
pixmap = QPixmap()
loading = False
if imageUrl.scheme() == "data":
encodedUrl = current.text(1).encode("utf-8")
imageData = encodedUrl[encodedUrl.find(b",") + 1:]
pixmap = WebBrowserTools.pixmapFromByteArray(imageData) | identifier_body |
SiteInfoDialog.py | , Qt
from PyQt5.QtGui import QPixmap, QImage, QPainter, QColor, QBrush
from PyQt5.QtNetwork import QNetworkRequest, QNetworkReply
from PyQt5.QtWidgets import (
QDialog, QTreeWidgetItem, QGraphicsScene, QMenu, QApplication,
QGraphicsPixmapItem
)
from E5Gui import E5MessageBox, E5FileDialog
from .Ui_SiteInfoDialog import Ui_SiteInfoDialog
from ..Tools import Scripts, WebBrowserTools
from ..WebBrowserPage import WebBrowserPage
import UI.PixmapCache
class SiteInfoDialog(QDialog, Ui_SiteInfoDialog):
"""
Class implementing a dialog to show some information about a site.
"""
okStyle = "QLabel { color : white; background-color : green; }"
nokStyle = "QLabel { color : white; background-color : red; }"
def __init__(self, browser, parent=None):
"""
Constructor
@param browser reference to the browser window (HelpBrowser)
@param parent reference to the parent widget (QWidget)
"""
super(SiteInfoDialog, self).__init__(parent)
self.setupUi(self)
self.setWindowFlags(Qt.Window)
# put icons
self.tabWidget.setTabIcon(
0, UI.PixmapCache.getIcon("siteinfo-general.png"))
self.tabWidget.setTabIcon(
1, UI.PixmapCache.getIcon("siteinfo-media.png"))
self.__imageReply = None
self.__baseUrl = browser.url()
title = browser.title()
#prepare background of image preview
self.__imagePreviewStandardBackground = (
self.imagePreview.backgroundBrush()
)
color1 = QColor(220, 220, 220)
color2 = QColor(160, 160, 160)
self.__tilePixmap = QPixmap(8, 8)
self.__tilePixmap.fill(color1)
tilePainter = QPainter(self.__tilePixmap)
tilePainter.fillRect(0, 0, 4, 4, color2)
tilePainter.fillRect(4, 4, 4, 4, color2)
tilePainter.end()
# populate General tab
self.heading.setText("<b>{0}</b>".format(title))
self.siteAddressLabel.setText(self.__baseUrl.toString())
if self.__baseUrl.scheme() in ["https"]:
self.securityLabel.setStyleSheet(SiteInfoDialog.okStyle)
self.securityLabel.setText(
self.tr('<b>Connection is encrypted.</b>'))
else:
self.securityLabel.setStyleSheet(SiteInfoDialog.nokStyle)
self.securityLabel.setText(
self.tr('<b>Connection is not encrypted.</b>'))
browser.page().runJavaScript(
"document.charset", WebBrowserPage.SafeJsWorld,
lambda res: self.encodingLabel.setText(res))
# populate Meta tags
browser.page().runJavaScript(Scripts.getAllMetaAttributes(),
WebBrowserPage.SafeJsWorld,
self.__processMetaAttributes)
# populate Media tab
browser.page().runJavaScript(Scripts.getAllImages(),
WebBrowserPage.SafeJsWorld,
self.__processImageTags)
def __processImageTags(self, res):
"""
Private method to process the image tags.
@param res result of the JavaScript script
@type list of dict
"""
for img in res:
src = img["src"]
alt = img["alt"]
if not alt:
if src.find("/") == -1:
alt = src
else:
pos = src.rfind("/")
alt = src[pos + 1:]
if not src or not alt:
continue
QTreeWidgetItem(self.imagesTree, [alt, src])
for col in range(self.imagesTree.columnCount()):
self.imagesTree.resizeColumnToContents(col)
if self.imagesTree.columnWidth(0) > 300:
self.imagesTree.setColumnWidth(0, 300)
self.imagesTree.setCurrentItem(self.imagesTree.topLevelItem(0))
self.imagesTree.setContextMenuPolicy(Qt.CustomContextMenu)
self.imagesTree.customContextMenuRequested.connect(
self.__imagesTreeContextMenuRequested)
def __processMetaAttributes(self, res):
"""
Private method to process the meta attributes.
@param res result of the JavaScript script
@type list of dict
"""
for meta in res:
content = meta["content"]
name = meta["name"]
if not name:
name = meta["httpequiv"]
if not name or not content:
continue
if meta["charset"]:
self.encodingLabel.setText(meta["charset"])
if "charset=" in content:
self.encodingLabel.setText(
content[content.index("charset=") + 8:])
QTreeWidgetItem(self.tagsTree, [name, content])
for col in range(self.tagsTree.columnCount()):
self.tagsTree.resizeColumnToContents(col)
@pyqtSlot(QTreeWidgetItem, QTreeWidgetItem)
def on_imagesTree_currentItemChanged(self, current, previous):
"""
Private slot to show a preview of the selected image.
@param current current image entry (QTreeWidgetItem)
@param previous old current entry (QTreeWidgetItem)
"""
if current is None:
return
imageUrl = QUrl(current.text(1))
if imageUrl.isRelative():
imageUrl = self.__baseUrl.resolved(imageUrl)
pixmap = QPixmap()
loading = False
if imageUrl.scheme() == "data":
encodedUrl = current.text(1).encode("utf-8")
imageData = encodedUrl[encodedUrl.find(b",") + 1:]
pixmap = WebBrowserTools.pixmapFromByteArray(imageData)
elif imageUrl.scheme() == "file":
pixmap = QPixmap(imageUrl.toLocalFile())
elif imageUrl.scheme() == "qrc":
pixmap = QPixmap(imageUrl.toString()[3:])
else:
if self.__imageReply is not None:
self.__imageReply.deleteLater()
self.__imageReply = None
from WebBrowser.WebBrowserWindow import WebBrowserWindow
self.__imageReply = WebBrowserWindow.networkManager().get(
QNetworkRequest(imageUrl))
self.__imageReply.finished.connect(self.__imageReplyFinished)
loading = True
self.__showLoadingText()
if not loading:
self.__showPixmap(pixmap)
@pyqtSlot()
def __imageReplyFinished(self):
"""
Private slot handling the loading of an image.
"""
if self.__imageReply.error() != QNetworkReply.NoError:
return
data = self.__imageReply.readAll()
self.__showPixmap(QPixmap.fromImage(QImage.fromData(data)))
def __showPixmap(self, pixmap):
"""
Private method to show a pixmap in the preview pane.
@param pixmap pixmap to be shown
@type QPixmap
"""
scene = QGraphicsScene(self.imagePreview)
if pixmap.isNull():
self.imagePreview.setBackgroundBrush(
self.__imagePreviewStandardBackground)
scene.addText(self.tr("Preview not available."))
else:
self.imagePreview.setBackgroundBrush(QBrush(self.__tilePixmap))
scene.addPixmap(pixmap)
self.imagePreview.setScene(scene)
def | (self):
"""
Private method to show some text while loading an image.
"""
self.imagePreview.setBackgroundBrush(
self.__imagePreviewStandardBackground)
scene = QGraphicsScene(self.imagePreview)
scene.addText(self.tr("Loading..."))
self.imagePreview.setScene(scene)
def __imagesTreeContextMenuRequested(self, pos):
"""
Private slot to show a context menu for the images list.
@param pos position for the menu (QPoint)
"""
itm = self.imagesTree.itemAt(pos)
if itm is None:
return
menu = QMenu()
act = menu.addAction(self.tr("Copy Image Location to Clipboard"))
act.setData(itm.text(1))
act.triggered.connect(lambda: self.__copyAction(act))
act = menu.addAction(self.tr("Copy Image Name to Clipboard"))
act.setData(itm.text(0))
act.triggered.connect(lambda: self.__copyAction(act))
menu.addSeparator()
act = menu.addAction(self.tr("Save Image"))
act.setData(self.imagesTree.indexOfTopLevelItem(itm))
act.triggered.connect(lambda: self.__saveImage(act))
menu.exec_(self.imagesTree.viewport().mapToGlobal(pos))
def __copyAction(self, act):
"""
Private slot to copy the image URL or the image name to the clipboard.
@param act reference to the action that triggered
@type QAction
"""
QApplication.clipboard().setText(act.data())
def __saveImage(self, act):
"""
Private slot to save the selected image to disk.
@param act reference to the action that triggered
@type QAction
"""
index = act.data()
itm = self.imagesTree.topLevelItem(index)
if itm is None:
return
if (
not self.imagePreview.scene() or
len(self.imagePreview.scene().items()) == 0
):
return
pixmapItem = self.imagePreview.scene().items()[0]
if not isinstance(pixmapItem, QGraphicsPixmapItem):
return
if pixmapItem.pixmap().isNull():
E5MessageBox.warning(
self,
self.tr("Save Image"),
self.tr(
"""<p>This preview is not available.</p>"""))
return
imageFileName = WebBrowserTools.get | __showLoadingText | identifier_name |
SiteInfoDialog.py | Url, Qt
from PyQt5.QtGui import QPixmap, QImage, QPainter, QColor, QBrush
from PyQt5.QtNetwork import QNetworkRequest, QNetworkReply
from PyQt5.QtWidgets import (
QDialog, QTreeWidgetItem, QGraphicsScene, QMenu, QApplication,
QGraphicsPixmapItem
)
from E5Gui import E5MessageBox, E5FileDialog
from .Ui_SiteInfoDialog import Ui_SiteInfoDialog
from ..Tools import Scripts, WebBrowserTools
from ..WebBrowserPage import WebBrowserPage
import UI.PixmapCache
class SiteInfoDialog(QDialog, Ui_SiteInfoDialog):
"""
Class implementing a dialog to show some information about a site.
"""
okStyle = "QLabel { color : white; background-color : green; }"
nokStyle = "QLabel { color : white; background-color : red; }"
def __init__(self, browser, parent=None):
"""
Constructor
@param browser reference to the browser window (HelpBrowser)
@param parent reference to the parent widget (QWidget)
"""
super(SiteInfoDialog, self).__init__(parent)
self.setupUi(self)
self.setWindowFlags(Qt.Window)
# put icons
self.tabWidget.setTabIcon(
0, UI.PixmapCache.getIcon("siteinfo-general.png"))
self.tabWidget.setTabIcon(
1, UI.PixmapCache.getIcon("siteinfo-media.png"))
self.__imageReply = None
self.__baseUrl = browser.url()
title = browser.title()
#prepare background of image preview
self.__imagePreviewStandardBackground = (
self.imagePreview.backgroundBrush()
)
color1 = QColor(220, 220, 220)
color2 = QColor(160, 160, 160)
self.__tilePixmap = QPixmap(8, 8)
self.__tilePixmap.fill(color1)
tilePainter = QPainter(self.__tilePixmap)
tilePainter.fillRect(0, 0, 4, 4, color2)
tilePainter.fillRect(4, 4, 4, 4, color2)
tilePainter.end()
# populate General tab
self.heading.setText("<b>{0}</b>".format(title))
self.siteAddressLabel.setText(self.__baseUrl.toString())
if self.__baseUrl.scheme() in ["https"]:
self.securityLabel.setStyleSheet(SiteInfoDialog.okStyle)
self.securityLabel.setText(
self.tr('<b>Connection is encrypted.</b>'))
else:
self.securityLabel.setStyleSheet(SiteInfoDialog.nokStyle)
self.securityLabel.setText(
self.tr('<b>Connection is not encrypted.</b>'))
browser.page().runJavaScript(
"document.charset", WebBrowserPage.SafeJsWorld,
lambda res: self.encodingLabel.setText(res))
# populate Meta tags
browser.page().runJavaScript(Scripts.getAllMetaAttributes(),
WebBrowserPage.SafeJsWorld,
self.__processMetaAttributes)
# populate Media tab
browser.page().runJavaScript(Scripts.getAllImages(),
WebBrowserPage.SafeJsWorld,
self.__processImageTags)
def __processImageTags(self, res):
"""
Private method to process the image tags.
@param res result of the JavaScript script
@type list of dict
"""
for img in res:
src = img["src"]
alt = img["alt"]
if not alt:
if src.find("/") == -1:
alt = src
else:
pos = src.rfind("/")
alt = src[pos + 1:]
if not src or not alt:
continue
QTreeWidgetItem(self.imagesTree, [alt, src])
for col in range(self.imagesTree.columnCount()):
self.imagesTree.resizeColumnToContents(col)
if self.imagesTree.columnWidth(0) > 300:
self.imagesTree.setColumnWidth(0, 300)
self.imagesTree.setCurrentItem(self.imagesTree.topLevelItem(0))
self.imagesTree.setContextMenuPolicy(Qt.CustomContextMenu)
self.imagesTree.customContextMenuRequested.connect(
self.__imagesTreeContextMenuRequested)
def __processMetaAttributes(self, res):
"""
Private method to process the meta attributes.
@param res result of the JavaScript script
@type list of dict
"""
for meta in res:
content = meta["content"]
name = meta["name"]
if not name:
name = meta["httpequiv"]
if not name or not content:
continue
if meta["charset"]:
self.encodingLabel.setText(meta["charset"])
if "charset=" in content:
self.encodingLabel.setText(
content[content.index("charset=") + 8:])
QTreeWidgetItem(self.tagsTree, [name, content])
for col in range(self.tagsTree.columnCount()):
self.tagsTree.resizeColumnToContents(col)
@pyqtSlot(QTreeWidgetItem, QTreeWidgetItem)
def on_imagesTree_currentItemChanged(self, current, previous):
"""
Private slot to show a preview of the selected image.
@param current current image entry (QTreeWidgetItem)
@param previous old current entry (QTreeWidgetItem)
"""
if current is None:
return
imageUrl = QUrl(current.text(1))
if imageUrl.isRelative():
imageUrl = self.__baseUrl.resolved(imageUrl)
pixmap = QPixmap()
loading = False
if imageUrl.scheme() == "data":
encodedUrl = current.text(1).encode("utf-8")
imageData = encodedUrl[encodedUrl.find(b",") + 1:]
pixmap = WebBrowserTools.pixmapFromByteArray(imageData)
elif imageUrl.scheme() == "file":
pixmap = QPixmap(imageUrl.toLocalFile())
elif imageUrl.scheme() == "qrc":
pixmap = QPixmap(imageUrl.toString()[3:])
else:
if self.__imageReply is not None:
self.__imageReply.deleteLater()
self.__imageReply = None
from WebBrowser.WebBrowserWindow import WebBrowserWindow
self.__imageReply = WebBrowserWindow.networkManager().get(
QNetworkRequest(imageUrl))
self.__imageReply.finished.connect(self.__imageReplyFinished)
loading = True
self.__showLoadingText()
if not loading:
self.__showPixmap(pixmap)
@pyqtSlot()
def __imageReplyFinished(self):
"""
Private slot handling the loading of an image.
"""
if self.__imageReply.error() != QNetworkReply.NoError:
return
data = self.__imageReply.readAll()
self.__showPixmap(QPixmap.fromImage(QImage.fromData(data)))
def __showPixmap(self, pixmap):
"""
Private method to show a pixmap in the preview pane.
@param pixmap pixmap to be shown
@type QPixmap
"""
scene = QGraphicsScene(self.imagePreview)
if pixmap.isNull():
self.imagePreview.setBackgroundBrush(
self.__imagePreviewStandardBackground)
scene.addText(self.tr("Preview not available."))
else:
self.imagePreview.setBackgroundBrush(QBrush(self.__tilePixmap))
scene.addPixmap(pixmap)
self.imagePreview.setScene(scene)
def __showLoadingText(self):
"""
Private method to show some text while loading an image.
"""
self.imagePreview.setBackgroundBrush(
self.__imagePreviewStandardBackground)
scene = QGraphicsScene(self.imagePreview)
scene.addText(self.tr("Loading..."))
self.imagePreview.setScene(scene)
def __imagesTreeContextMenuRequested(self, pos):
"""
Private slot to show a context menu for the images list.
@param pos position for the menu (QPoint)
"""
itm = self.imagesTree.itemAt(pos)
if itm is None:
return
menu = QMenu()
act = menu.addAction(self.tr("Copy Image Location to Clipboard"))
act.setData(itm.text(1))
act.triggered.connect(lambda: self.__copyAction(act))
act = menu.addAction(self.tr("Copy Image Name to Clipboard"))
act.setData(itm.text(0))
act.triggered.connect(lambda: self.__copyAction(act))
menu.addSeparator()
act = menu.addAction(self.tr("Save Image"))
act.setData(self.imagesTree.indexOfTopLevelItem(itm))
act.triggered.connect(lambda: self.__saveImage(act))
menu.exec_(self.imagesTree.viewport().mapToGlobal(pos))
def __copyAction(self, act):
"""
Private slot to copy the image URL or the image name to the clipboard.
@param act reference to the action that triggered
@type QAction
"""
QApplication.clipboard().setText(act.data())
def __saveImage(self, act):
"""
Private slot to save the selected image to disk.
@param act reference to the action that triggered
@type QAction
"""
index = act.data()
itm = self.imagesTree.topLevelItem(index)
if itm is None:
return
if (
not self.imagePreview.scene() or
len(self.imagePreview.scene().items()) == 0
): |
pixmapItem = self.imagePreview.scene().items()[0]
if not isinstance(pixmapItem, QGraphicsPixmapItem):
return
if pixmapItem.pixmap().isNull():
E5MessageBox.warning(
self,
self.tr("Save Image"),
self.tr(
"""<p>This preview is not available.</p>"""))
return
imageFileName = WebBrowserTools.get | return | random_line_split |
kNN.py | normMat
def file2matrix(filename):
fr = open(filename)
lines = fr.readlines()
numberOfLines = len(lines)
dataMat = np.zeros((numberOfLines, 3))
classLabelVec = []
index = 0
for line in lines:
listFromLine = line.strip().split('\t')
dataMat[index, :] = listFromLine[0:3]
classLabelVec.append(int(listFromLine[-1]))
index += 1
return dataMat, classLabelVec
def datingClassTest():
hoRatio = 0.50
datingDataMat, datingLabelVec = file2matrix('datingTestSet2.txt')
# 归一化
normMat = autoNorm(datingDataMat)
m = normMat.shape[0]
# 抽样测试样本
numTestVecs = int(m * hoRatio)
errorCount = 0.0
for i in range(numTestVecs):
classifierResult = classify0(normMat[i,:], normMat[numTestVecs:m,:], datingLabelVec[numTestVecs:m], 3)
print "the classifier came back with: %d, the real answer is: %d" % (classifierResult, datingLabelVec[i])
if (classifierResult != datingLabelVec[i]):
errorCount += 1.0
print "the total error rate is:%f" % (errorCount/float(numTestVecs))
print errorCount
def img2vector(filename):
returnVec = np.zeros((1, 1024))
fr = open(filename)
for i in range(32):
lineStr = fr.readline()
for j in range(32):
returnVec[0, i * 32 + j] = int(l |
return returnVec
def handwritingClassTest():
from os import listdir
from sklearn.neighbors import KNeighborsClassifier as kNN
hwLabels = []
# 目录下文件名
trainingFileList = listdir('trainingDigits')
m = len(trainingFileList)
trainingMat = np.zeros((m, 1024))
# 获取训练数据和标签
for i in range(m):
fileNameStr = trainingFileList[i]
classNumStr = int(fileNameStr.split('_')[0])
hwLabels.append(classNumStr)
trainingMat[i, :] = img2vector('trainingDigits/%s' % fileNameStr)
#构建kNN分类器
neigh = kNN(n_neighbors = 10, algorithm = 'auto')
#拟合模型, trainingMat为测试矩阵,hwLabels为对应的标签
neigh.fit(trainingMat, hwLabels)
testFileList = listdir('testDigits')
errorCount = 0.0
mTest = len(testFileList)
for i in range(mTest):
fileNameStr = testFileList[i]
classNumStr = int(fileNameStr.split('_')[0])
vectorUnderTest = img2vector('testDigits/%s' % fileNameStr)
#classifierResult = classify0(vectorUnderTest, trainingMat, hwLabels, 3)
classifierResult = neigh.predict(vectorUnderTest);
print "the classifier came back with: %d, the real answer is: %d" % (classifierResult, classNumStr)
if (classifierResult != classNumStr):
errorCount += 1.0
print "\nthe total number of errors is: %d" % errorCount
print "\nthe total error rate is: %f" % (errorCount / float(mTest))
return trainingFileList
"""
函数说明:可视化数据
Parameters:
datingDataMat - 特征矩阵
datingLabels - 分类Label
Returns:
无
Modify:
2017-03-24
"""
def showdatas(datingDataMat, datingLabels):
import matplotlib.pyplot as plt
import matplotlib.lines as mlines
#设置汉字格式
#font = FontProperties(fname=r"c:\windows\fonts\simsun.ttc", size=14)
#将fig画布分隔成1行1列,不共享x轴和y轴,fig画布的大小为(13,8)
#当nrow=2,nclos=2时,代表fig画布被分为四个区域,axs[0][0]表示第一行第一个区域
fig, axs = plt.subplots(nrows=2, ncols=2,sharex=False, sharey=False, figsize=(13,8))
numberOfLabels = len(datingLabels)
LabelsColors = []
for i in datingLabels:
if i == 1:
LabelsColors.append('black')
if i == 2:
LabelsColors.append('orange')
if i == 3:
LabelsColors.append('red')
#画出散点图,以datingDataMat矩阵的第一(飞行常客例程)、第二列(玩游戏)数据画散点数据,散点大小为15,透明度为0.5
axs[0][0].scatter(x=datingDataMat[:,0], y=datingDataMat[:,1], color=LabelsColors,s=15, alpha=.5)
#设置标题,x轴label,y轴label
axs0_title_text = axs[0][0].set_title(u'每年获得的飞行常客里程数与玩视频游戏所消耗时间占比')
axs0_xlabel_text = axs[0][0].set_xlabel(u'每年获得的飞行常客里程数')
axs0_ylabel_text = axs[0][0].set_ylabel(u'玩视频游戏所消耗时间占')
plt.setp(axs0_title_text, size=9, weight='bold', color='red')
plt.setp(axs0_xlabel_text, size=7, weight='bold', color='black')
plt.setp(axs0_ylabel_text, size=7, weight='bold', color='black')
#画出散点图,以datingDataMat矩阵的第一(飞行常客例程)、第三列(冰激凌)数据画散点数据,散点大小为15,透明度为0.5
axs[0][1].scatter(x=datingDataMat[:,0], y=datingDataMat[:,2], color=LabelsColors,s=15, alpha=.5)
#设置标题,x轴label,y轴label
axs1_title_text = axs[0][1].set_title(u'每年获得的飞行常客里程数与每周消费的冰激淋公升数')
axs1_xlabel_text = axs[0][1].set_xlabel(u'每年获得的飞行常客里程数')
axs1_ylabel_text = axs[0][1].set_ylabel(u'每周消费的冰激淋公升数')
plt.setp(axs1_title_text, size=9, weight='bold', color='red')
plt.setp(axs1_xlabel_text, size=7, weight='bold', color='black')
plt.setp(axs1_ylabel_text, size=7, weight='bold', color='black')
#画出散点图,以datingDataMat矩阵的第二(玩游戏)、第三列(冰激凌)数据画散点数据,散点大小为15,透明度为0.5
axs[1][0].scatter(x=datingDataMat[:,1], y=datingDataMat[:,2], color=LabelsColors,s=15, alpha=.5)
#设置标题,x轴label,y轴label
axs2_title_text = axs[1][0].set_title(u'玩视频游戏所消耗时间占比与每周消费的冰激淋公升数')
axs2_xlabel_text = axs[1][0].set_xlabel(u'玩视频游戏所消耗时间占比')
axs2_ylabel_text = axs[1][0].set_ylabel(u'每周消费的冰激淋公升数')
plt.setp(axs2_title_text, size=9, weight='bold', color='red')
plt.setp(axs2_xlabel_text, size=7, weight='bold', color='black')
plt.setp(axs2_ylabel_text, size=7, weight='bold', color='black')
#设置图例
didntLike = mlines.Line2D([], [], color='black', marker='.',
markersize=6, label='didntLike')
smallDoses = mlines.Line2D([], [], color='orange', marker='.',
markersize=6, label='smallDoses')
largeDoses = mlines.Line2D([], [], color='red', marker='.',
markersize=6, label='largeDoses')
#添加图例
axs[0][0].legend(handles=[didntLike,smallDoses,largeDoses])
axs[0][1].legend(handles=[didntLike,smallDoses,largeDoses])
axs[1][0].legend(handles=[didntLike,smallDoses,largeDoses])
#显示图片
plt.show()
def showDataPlot():
import matplotlib.pyplot as plt
plt.plot([1, 2, 3, 4], [1, 4, 9, 16], 'ro')
plt.axis([0, 6, 0, 20])
plt.ylabel('some numbers')
plt.show | ineStr[i]) | identifier_name |
kNN.py | Labels = len(datingLabels)
LabelsColors = []
for i in datingLabels:
if i == 1:
LabelsColors.append('black')
if i == 2:
LabelsColors.append('orange')
if i == 3:
LabelsColors.append('red')
#画出散点图,以datingDataMat矩阵的第一(飞行常客例程)、第二列(玩游戏)数据画散点数据,散点大小为15,透明度为0.5
axs[0][0].scatter(x=datingDataMat[:,0], y=datingDataMat[:,1], color=LabelsColors,s=15, alpha=.5)
#设置标题,x轴label,y轴label
axs0_title_text = axs[0][0].set_title(u'每年获得的飞行常客里程数与玩视频游戏所消耗时间占比')
axs0_xlabel_text = axs[0][0].set_xlabel(u'每年获得的飞行常客里程数')
axs0_ylabel_text = axs[0][0].set_ylabel(u'玩视频游戏所消耗时间占')
plt.setp(axs0_title_text, size=9, weight='bold', color='red')
plt.setp(axs0_xlabel_text, size=7, weight='bold', color='black')
plt.setp(axs0_ylabel_text, size=7, weight='bold', color='black')
#画出散点图,以datingDataMat矩阵的第一(飞行常客例程)、第三列(冰激凌)数据画散点数据,散点大小为15,透明度为0.5
axs[0][1].scatter(x=datingDataMat[:,0], y=datingDataMat[:,2], color=LabelsColors,s=15, alpha=.5)
#设置标题,x轴label,y轴label
axs1_title_text = axs[0][1].set_title(u'每年获得的飞行常客里程数与每周消费的冰激淋公升数')
axs1_xlabel_text = axs[0][1].set_xlabel(u'每年获得的飞行常客里程数')
axs1_ylabel_text = axs[0][1].set_ylabel(u'每周消费的冰激淋公升数')
plt.setp(axs1_title_text, size=9, weight='bold', color='red')
plt.setp(axs1_xlabel_text, size=7, weight='bold', color='black')
plt.setp(axs1_ylabel_text, size=7, weight='bold', color='black')
#画出散点图,以datingDataMat矩阵的第二(玩游戏)、第三列(冰激凌)数据画散点数据,散点大小为15,透明度为0.5
axs[1][0].scatter(x=datingDataMat[:,1], y=datingDataMat[:,2], color=LabelsColors,s=15, alpha=.5)
#设置标题,x轴label,y轴label
axs2_title_text = axs[1][0].set_title(u'玩视频游戏所消耗时间占比与每周消费的冰激淋公升数')
axs2_xlabel_text = axs[1][0].set_xlabel(u'玩视频游戏所消耗时间占比')
axs2_ylabel_text = axs[1][0].set_ylabel(u'每周消费的冰激淋公升数')
plt.setp(axs2_title_text, size=9, weight='bold', color='red')
plt.setp(axs2_xlabel_text, size=7, weight='bold', color='black')
plt.setp(axs2_ylabel_text, size=7, weight='bold', color='black')
#设置图例
didntLike = mlines.Line2D([], [], color='black', marker='.',
markersize=6, label='didntLike')
smallDoses = mlines.Line2D([], [], color='orange', marker='.',
markersize=6, label='smallDoses')
largeDoses = mlines.Line2D([], [], color='red', marker='.',
markersize=6, label='largeDoses')
#添加图例
axs[0][0].legend(handles=[didntLike,smallDoses,largeDoses])
axs[0][1].legend(handles=[didntLike,smallDoses,largeDoses])
axs[1][0].legend(handles=[didntLike,smallDoses,largeDoses])
#显示图片
plt.show()
def showDataPlot():
import matplotlib.pyplot as plt
plt.plot([1, 2, 3, 4], [1, 4, 9, 16], 'ro')
plt.axis([0, 6, 0, 20])
plt.ylabel('some numbers')
plt.show()
def showDifPlot():
import numpy as np
import matplotlib.pyplot as plt
# evenly sampled time at 200ms intervals
t = np.arange(0., 5., 0.2)
# red dashes, blue squares and green triangles
plt.plot(t, t, 'r--', t, t**2, 'bs', t, t**3, 'g^')
plt.show()
def showPic():
import matplotlib.pyplot as plt
import matplotlib.image as mpimg
import numpy as np
img=mpimg.imread('stinkbug.png')
print img
imgplot = plt.imshow(img)
plt.show()
# RGB单色图
lum_img = img[:, :, 0]
plt.imshow(lum_img)
plt.show()
# 热力图
plt.imshow(lum_img, cmap="hot")
plt.show()
#
imgplot = plt.imshow(lum_img)
imgplot.set_cmap('nipy_spectral')
plt.colorbar()
plt.show()
def showHistogram():
import matplotlib.pyplot as plt
import matplotlib.image as mpimg
import numpy as np
img=mpimg.imread('stinkbug.png')
lum_img = img[:, :, 0]
plt.hist(lum_img.ravel(), bins=256, range=(0.0, 1.0), fc='k', ec='k')
plt.show()
def showComplex():
import matplotlib.pyplot as plt
import matplotlib.image as mpimg
import numpy as np
img=mpimg.imread('stinkbug.png')
lum_img = img[:, :, 0]
imgplot = plt.imshow(lum_img, clim=(0.0, 0.7))
plt.show()
def showImages():
import matplotlib.pyplot as plt
import matplotlib.image as mpimg
import numpy as np
from PIL import Image
img = Image.open('stinkbug.png')
img.thumbnail((64, 64), Image.ANTIALIAS) # resizes image in-place
imgplot = plt.imshow(img, interpolation="bicubic")
plt.show()
def showSubplot():
import matplotlib.pyplot as plt
ax1 = plt.subplot2grid((3, 3), (0, 0), colspan=3)
ax2 = plt.subplot2grid((3, 3), (1, 0), colspan=2)
ax3 = plt.subplot2grid((3, 3), (1, 2), rowspan=2)
ax4 = plt.subplot2grid((3, 3), (2, 0))
ax5 = plt.subplot2grid((3, 3), (2, 1))
plt.show()
def showGredSpec():
import matplotlib.gridspec as gridspec
gs = gridspec.GridSpec(3, 3)
ax1 = plt.subplot(gs[0, :])
ax2 = plt.subplot(gs[1, :-1])
ax3 = plt.subplot(gs[1:, -1])
ax4 = plt.subplot(gs[-1, 0])
ax5 = plt.subplot(gs[-1, -2])
plt.show()
def showDemoGridSpec():
import matplotlib.pyplot as plt
import matplotlib.gridspec as gridspec
import numpy as np
try:
from itertools import product
except ImportError:
# product is new in v 2.6
def product(*args, **kwds):
pools = map(tuple, args) * kwds.get('repeat', 1)
result = [[]]
for pool in pools:
result = [x+[y] for x in result for y in pool]
for prod in result:
yield tuple(prod)
def squiggle_xy(a, b, c, d, i=np.arange(0.0, 2*np.pi, 0.05)):
return np.sin(i*a)*np.cos(i*b), np.sin(i*c)*np.cos(i*d)
fig = plt.figure(figsize=(8, 8))
# gridspec inside gridspec
outer_grid = gridspec.Gr | idSpec(4, 4, wspace=0.0, hspace=0.0)
for i in range(16):
inner_grid = gridspec.GridSpecFromSubplotSpec(3, 3,
subplot_spec=outer_grid[i], wspace=0.0, hspace=0.0)
a, b = int(i/4)+1,i%4+1
for j, (c, d) in enumer | identifier_body |
|
kNN.py | 14)
#将fig画布分隔成1行1列,不共享x轴和y轴,fig画布的大小为(13,8)
#当nrow=2,nclos=2时,代表fig画布被分为四个区域,axs[0][0]表示第一行第一个区域
fig, axs = plt.subplots(nrows=2, ncols=2,sharex=False, sharey=False, figsize=(13,8))
numberOfLabels = len(datingLabels)
LabelsColors = []
for i in datingLabels:
if i == 1:
LabelsColors.append('black')
if i == 2:
LabelsColors.append('orange')
if i == 3:
LabelsColors.append('red')
#画出散点图,以datingDataMat矩阵的第一(飞行常客例程)、第二列(玩游戏)数据画散点数据,散点大小为15,透明度为0.5
axs[0][0].scatter(x=datingDataMat[:,0], y=datingDataMat[:,1], color=LabelsColors,s=15, alpha=.5)
#设置标题,x轴label,y轴label
axs0_title_text = axs[0][0].set_title(u'每年获得的飞行常客里程数与玩视频游戏所消耗时间占比')
axs0_xlabel_text = axs[0][0].set_xlabel(u'每年获得的飞行常客里程数')
axs0_ylabel_text = axs[0][0].set_ylabel(u'玩视频游戏所消耗时间占')
plt.setp(axs0_title_text, size=9, weight='bold', color='red')
plt.setp(axs0_xlabel_text, size=7, weight='bold', color='black')
plt.setp(axs0_ylabel_text, size=7, weight='bold', color='black')
#画出散点图,以datingDataMat矩阵的第一(飞行常客例程)、第三列(冰激凌)数据画散点数据,散点大小为15,透明度为0.5
axs[0][1].scatter(x=datingDataMat[:,0], y=datingDataMat[:,2], color=LabelsColors,s=15, alpha=.5)
#设置标题,x轴label,y轴label
axs1_title_text = axs[0][1].set_title(u'每年获得的飞行常客里程数与每周消费的冰激淋公升数')
axs1_xlabel_text = axs[0][1].set_xlabel(u'每年获得的飞行常客里程数')
axs1_ylabel_text = axs[0][1].set_ylabel(u'每周消费的冰激淋公升数')
plt.setp(axs1_title_text, size=9, weight='bold', color='red')
plt.setp(axs1_xlabel_text, size=7, weight='bold', color='black')
plt.setp(axs1_ylabel_text, size=7, weight='bold', color='black')
#画出散点图,以datingDataMat矩阵的第二(玩游戏)、第三列(冰激凌)数据画散点数据,散点大小为15,透明度为0.5
axs[1][0].scatter(x=datingDataMat[:,1], y=datingDataMat[:,2], color=LabelsColors,s=15, alpha=.5)
#设置标题,x轴label,y轴label
axs2_title_text = axs[1][0].set_title(u'玩视频游戏所消耗时间占比与每周消费的冰激淋公升数')
axs2_xlabel_text = axs[1][0].set_xlabel(u'玩视频游戏所消耗时间占比')
axs2_ylabel_text = axs[1][0].set_ylabel(u'每周消费的冰激淋公升数')
plt.setp(axs2_title_text, size=9, weight='bold', color='red')
plt.setp(axs2_xlabel_text, size=7, weight='bold', color='black')
plt.setp(axs2_ylabel_text, size=7, weight='bold', color='black')
#设置图例
didntLike = mlines.Line2D([], [], color='black', marker='.',
markersize=6, label='didntLike')
smallDoses = mlines.Line2D([], [], color='orange', marker='.',
markersize=6, label='smallDoses')
largeDoses = mlines.Line2D([], [], color='red', marker='.',
markersize=6, label='largeDoses')
#添加图例
axs[0][0].legend(handles=[didntLike,smallDoses,largeDoses])
axs[0][1].legend(handles=[didntLike,smallDoses,largeDoses])
axs[1][0].legend(handles=[didntLike,smallDoses,largeDoses])
#显示图片
plt.show()
def showDataPlot():
import matplotlib.pyplot as plt
plt.plot([1, 2, 3, 4], [1, 4, 9, 16], 'ro')
plt.axis([0, 6, 0, 20])
plt.ylabel('some numbers')
plt.show()
def showDifPlot():
import numpy as np
import matplotlib.pyplot as plt
# evenly sampled time at 200ms intervals
t = np.arange(0., 5., 0.2)
# red dashes, blue squares and green triangles
plt.plot(t, t, 'r--', t, t**2, 'bs', t, t**3, 'g^')
plt.show()
def showPic():
import matplotlib.pyplot as plt
import matplotlib.image as mpimg
import numpy as np
img=mpimg.imread('stinkbug.png')
print img
imgplot = plt.imshow(img)
plt.show()
# RGB单色图
lum_img = img[:, :, 0]
plt.imshow(lum_img)
plt.show()
# 热力图
plt.imshow(lum_img, cmap="hot")
plt.show()
#
imgplot = plt.imshow(lum_img)
imgplot.set_cmap('nipy_spectral')
plt.colorbar()
plt.show()
def showHistogram():
import matplotlib.pyplot as plt
import matplotlib.image as mpimg
import numpy as np
img=mpimg.imread('stinkbug.png')
lum_img = img[:, :, 0]
plt.hist(lum_img.ravel(), bins=256, range=(0.0, 1.0), fc='k', ec='k')
plt.show()
def showComplex():
import matplotlib.pyplot as plt
import matplotlib.image as mpimg
import numpy as np
img=mpimg.imread('stinkbug.png')
lum_img = img[:, :, 0]
imgplot = plt.imshow(lum_img, clim=(0.0, 0.7))
plt.show()
def showImages():
import matplotlib.pyplot as plt
import matplotlib.image as mpimg
import numpy as np
from PIL import Image
img = Image.open('stinkbug.png')
img.thumbnail((64, 64), Image.ANTIALIAS) # resizes image in-place
imgplot = plt.imshow(img, interpolation="bicubic")
plt.show()
def showSubplot():
import matplotlib.pyplot as plt
ax1 = plt.subplot2grid((3, 3), (0, 0), colspan=3)
ax2 = plt.subplot2grid((3, 3), (1, 0), colspan=2)
ax3 = plt.subplot2grid((3, 3), (1, 2), rowspan=2)
ax4 = plt.subplot2grid((3, 3), (2, 0))
ax5 = plt.subplot2grid((3, 3), (2, 1))
plt.show()
def showGredSpec():
import matplotlib.gridspec as gridspec
gs = gridspec.GridSpec(3, 3)
ax1 = plt.subplot(gs[0, :])
ax2 = plt.subplot(gs[1, :-1])
ax3 = plt.subplot(gs[1:, -1])
ax4 = plt.subplot(gs[-1, 0])
ax5 = plt.subplot(gs[-1, -2])
plt.show()
def showDemoGridSpec():
import matplotlib.pyplot as plt
import matplotlib.gridspec as gridspec
import numpy as np
try:
from itertools import product
except ImportError:
# product is new in v 2.6
def product(*args, **kwds):
pools = map(tuple, args) * kwds.get('repeat', 1)
result = [[]]
for pool in pools:
result = [x+[y] for x in result for y in pool]
for prod in result:
yield tuple(prod)
def squiggle_xy(a, b, c, d, i=np.arange(0.0, 2*np.pi, 0.05)):
return np.sin(i*a)*np.cos(i*b), np.sin(i*c)*np.cos(i*d) |
fig = plt.figure(figsize=(8, 8))
# gridspec inside gridspec | random_line_split |
|
kNN.py | normMat
def file2matrix(filename):
fr = open(filename)
lines = fr.readlines()
numberOfLines = len(lines)
dataMat = np.zeros((numberOfLines, 3))
classLabelVec = []
index = 0
for line in lines:
listFromLine = line.strip().split('\t')
dataMat[index, :] = listFromLine[0:3]
classLabelVec.append(int(listFromLine[-1]))
index += 1
return dataMat, classLabelVec
def datingClassTest():
hoRatio = 0.50
datingDataMat, datingLabelVec = file2matrix('datingTestSet2.txt')
# 归一化
normMat = autoNorm(datingDataMat)
m = normMat.shape[0]
# 抽样测试样本
numTestVecs = int(m * hoRatio)
errorCount = 0.0
for i in range(numTestVecs):
classifierResult = classify0(normMat[i,:], normMat[numTestVecs:m,:], datingLabelVec[numTestVecs:m], 3)
print "the classifier came back with: %d, the real answer is: %d" % (classifierResult, datingLabelV | lineStr = fr.readline()
for j in range(32):
returnVec[0, i * 32 + j] = int(lineStr[i])
return returnVec
def handwritingClassTest():
from os import listdir
from sklearn.neighbors import KNeighborsClassifier as kNN
hwLabels = []
# 目录下文件名
trainingFileList = listdir('trainingDigits')
m = len(trainingFileList)
trainingMat = np.zeros((m, 1024))
# 获取训练数据和标签
for i in range(m):
fileNameStr = trainingFileList[i]
classNumStr = int(fileNameStr.split('_')[0])
hwLabels.append(classNumStr)
trainingMat[i, :] = img2vector('trainingDigits/%s' % fileNameStr)
#构建kNN分类器
neigh = kNN(n_neighbors = 10, algorithm = 'auto')
#拟合模型, trainingMat为测试矩阵,hwLabels为对应的标签
neigh.fit(trainingMat, hwLabels)
testFileList = listdir('testDigits')
errorCount = 0.0
mTest = len(testFileList)
for i in range(mTest):
fileNameStr = testFileList[i]
classNumStr = int(fileNameStr.split('_')[0])
vectorUnderTest = img2vector('testDigits/%s' % fileNameStr)
#classifierResult = classify0(vectorUnderTest, trainingMat, hwLabels, 3)
classifierResult = neigh.predict(vectorUnderTest);
print "the classifier came back with: %d, the real answer is: %d" % (classifierResult, classNumStr)
if (classifierResult != classNumStr):
errorCount += 1.0
print "\nthe total number of errors is: %d" % errorCount
print "\nthe total error rate is: %f" % (errorCount / float(mTest))
return trainingFileList
"""
函数说明:可视化数据
Parameters:
datingDataMat - 特征矩阵
datingLabels - 分类Label
Returns:
无
Modify:
2017-03-24
"""
def showdatas(datingDataMat, datingLabels):
import matplotlib.pyplot as plt
import matplotlib.lines as mlines
#设置汉字格式
#font = FontProperties(fname=r"c:\windows\fonts\simsun.ttc", size=14)
#将fig画布分隔成1行1列,不共享x轴和y轴,fig画布的大小为(13,8)
#当nrow=2,nclos=2时,代表fig画布被分为四个区域,axs[0][0]表示第一行第一个区域
fig, axs = plt.subplots(nrows=2, ncols=2,sharex=False, sharey=False, figsize=(13,8))
numberOfLabels = len(datingLabels)
LabelsColors = []
for i in datingLabels:
if i == 1:
LabelsColors.append('black')
if i == 2:
LabelsColors.append('orange')
if i == 3:
LabelsColors.append('red')
#画出散点图,以datingDataMat矩阵的第一(飞行常客例程)、第二列(玩游戏)数据画散点数据,散点大小为15,透明度为0.5
axs[0][0].scatter(x=datingDataMat[:,0], y=datingDataMat[:,1], color=LabelsColors,s=15, alpha=.5)
#设置标题,x轴label,y轴label
axs0_title_text = axs[0][0].set_title(u'每年获得的飞行常客里程数与玩视频游戏所消耗时间占比')
axs0_xlabel_text = axs[0][0].set_xlabel(u'每年获得的飞行常客里程数')
axs0_ylabel_text = axs[0][0].set_ylabel(u'玩视频游戏所消耗时间占')
plt.setp(axs0_title_text, size=9, weight='bold', color='red')
plt.setp(axs0_xlabel_text, size=7, weight='bold', color='black')
plt.setp(axs0_ylabel_text, size=7, weight='bold', color='black')
#画出散点图,以datingDataMat矩阵的第一(飞行常客例程)、第三列(冰激凌)数据画散点数据,散点大小为15,透明度为0.5
axs[0][1].scatter(x=datingDataMat[:,0], y=datingDataMat[:,2], color=LabelsColors,s=15, alpha=.5)
#设置标题,x轴label,y轴label
axs1_title_text = axs[0][1].set_title(u'每年获得的飞行常客里程数与每周消费的冰激淋公升数')
axs1_xlabel_text = axs[0][1].set_xlabel(u'每年获得的飞行常客里程数')
axs1_ylabel_text = axs[0][1].set_ylabel(u'每周消费的冰激淋公升数')
plt.setp(axs1_title_text, size=9, weight='bold', color='red')
plt.setp(axs1_xlabel_text, size=7, weight='bold', color='black')
plt.setp(axs1_ylabel_text, size=7, weight='bold', color='black')
#画出散点图,以datingDataMat矩阵的第二(玩游戏)、第三列(冰激凌)数据画散点数据,散点大小为15,透明度为0.5
axs[1][0].scatter(x=datingDataMat[:,1], y=datingDataMat[:,2], color=LabelsColors,s=15, alpha=.5)
#设置标题,x轴label,y轴label
axs2_title_text = axs[1][0].set_title(u'玩视频游戏所消耗时间占比与每周消费的冰激淋公升数')
axs2_xlabel_text = axs[1][0].set_xlabel(u'玩视频游戏所消耗时间占比')
axs2_ylabel_text = axs[1][0].set_ylabel(u'每周消费的冰激淋公升数')
plt.setp(axs2_title_text, size=9, weight='bold', color='red')
plt.setp(axs2_xlabel_text, size=7, weight='bold', color='black')
plt.setp(axs2_ylabel_text, size=7, weight='bold', color='black')
#设置图例
didntLike = mlines.Line2D([], [], color='black', marker='.',
markersize=6, label='didntLike')
smallDoses = mlines.Line2D([], [], color='orange', marker='.',
markersize=6, label='smallDoses')
largeDoses = mlines.Line2D([], [], color='red', marker='.',
markersize=6, label='largeDoses')
#添加图例
axs[0][0].legend(handles=[didntLike,smallDoses,largeDoses])
axs[0][1].legend(handles=[didntLike,smallDoses,largeDoses])
axs[1][0].legend(handles=[didntLike,smallDoses,largeDoses])
#显示图片
plt.show()
def showDataPlot():
import matplotlib.pyplot as plt
plt.plot([1, 2, 3, 4], [1, 4, 9, 16], 'ro')
plt.axis([0, 6, 0, 20])
plt.ylabel('some numbers')
plt.show | ec[i])
if (classifierResult != datingLabelVec[i]):
errorCount += 1.0
print "the total error rate is:%f" % (errorCount/float(numTestVecs))
print errorCount
def img2vector(filename):
returnVec = np.zeros((1, 1024))
fr = open(filename)
for i in range(32):
| conditional_block |
chat-ui.js | Name);
var $roomItem = $(template(room));
$roomItem.children('a').bind('click', selectRoomListItem);
count++;
this.$roomList.append($roomItem.toggle(true));
}
_sortListLexicographically(self.$roomList);
cb();
};
NKChatUI.prototype.UIPaintUserList = function (users, cb) {
var template = function (obj) { obj || (obj = {}); var __t, __p = '', __e = _.escape, __j = Array.prototype.join; function print() { __p += __j.call(arguments, '') } with (obj) {
__p += '<li class=\'list-group-item\' data-user-id=\'' + __e(id) + '\'><b>' + __e(nameTrimmed) + '</b>';
__p += '<span class=\'pull-right\'><img style=\'height: 25px; width: auto\' src=\'img\/avatar-peer' + avatar + '.svg\' alt=\'User profile picture\'></span>';
__p += '\n</li>'; } return __p };
this.$userList.empty();
for (var username in users) {
var user = users[username];
user.disableActions = (!this._user || user.id === this._user.id);
if (user.name.substring(0,1) == '@')
{
user.avatar = user.name.substring(1,2);
user.name = user.name.substring(3);
} else
{
var s = "0" + _hashCode(user.name);
user.avatar = s.substr(s.length-2);
}
user.nameTrimmed = _trimWithEllipsis(user.name, this.maxLengthUsernameDisplay);
user.isMuted = (this._user && this._user.muted && this._user.muted[user.id]);
this.$userList.append(template(user));
}
_sortListLexicographically(this.$userList);
cb();
};
NKChatUI.prototype.UIPaintPrimaryUser = function () {
var self = this;
if (!self._user.avatar)
{
var s = "0" + _hashCode(self._user.name);
self._user.avatar = s.substr(s.length-2);
}
this.$currentUserDivs.each(function()
{
$( this ).html(self._user.name);
})
this.$currentUserAvatars.each(function()
{
$( this ).attr('src', 'img\/avatar-peer' + self._user.avatar + '.svg');
})
};
NKChatUI.prototype.UIClearMessages = function () {
$('.' + this.$$chatItemClass).not('.' + this.$$chatItemHiddenClass).remove();
};
var _scrollTime = (new Date()).getTime();
NKChatUI.prototype.UIPaintChatMessage = function(message) {
var self = this;
var $chatBox = $('.' + this.$$chatItemPrefix + message.origin).first().clone();
$chatBox.find('p').html('<div>' + message.message + '<div class=\'' + this.$$chatItemTitleClass + '\'>' + message.name + '</div>' + ' </div>')
$chatBox.attr('data-message-id', message.messageId);
if (message.avatar)
$chatBox.find("img").eq(0).attr("src", "img/avatar-peer" + message.avatar + ".svg");
$chatBox.insertBefore(this.$loading);
setTimeout(function() {
$chatBox.removeClass(self.$$chatItemHiddenClass);
}, 100);
var newScrollTime = (new Date()).getTime();
if ((newScrollTime - _scrollTime) > 500)
this.UIScrollChatToBottom();
_scrollTime = newScrollTime;
if (!message.messageId)
this.$loading.hide();
};
NKChatUI.prototype.UIRemoveChatMessage = function (messageId) {
$('.' + this.$$chatItemClass + '[data-message-id="' + messageId + '"]').remove()
};
// BRIDGE METHODS BETWEEN CHAT API AND UI METHODS ABOVE
NKChatUI.prototype.refreshRooms = function() {
var self = this;
this._chat.getRoomList(function(rooms) {
self.UIPaintRoomList(rooms, function() {
self._chat.enterRoom(self._roomId);
});
});
};
NKChatUI.prototype.refreshUsers = function () {
var self = this;
this._chat.getUsersByRoom(self._roomId, function(users){
self.UIPaintUserList(users, function(){});
});
};
NKChatUI.prototype.sendMessage = function (msg, cb) {
this._chat.sendMessage(this._roomId, msg, 'default', cb);
};
NKChatUI.prototype.createRoom = function (roomName, cb) {
this._chat.createRoom('NK' + roomName, 'public', cb);
};
NKChatUI.prototype._onEnterRoom = function(room) {
var self = this;
if (room.name == "MyNewRoom")
room.name = "Public";
this.$roomTitle.html(room.name);
this.$roomTitle2.html(room.name);
this._roomId = room.id;
this.UIClearMessages();
this.refreshRooms();
this.refreshUsers();
setTimeout(function() {
var element = self.$chatPane;
element.animate({
scrollTop: element[0].scrollHeight
}, 420);
}, 500);
};
NKChatUI.prototype._onNewMessage = function(roomId, rawMessage) {
if (roomId == this._digitalAssistantRoomId) {
if (rawMessage.message.userid != this._user.id)
return;
rawMessage.isDigitalAssistant = true;
} else
rawMessage.isDigitalAssistant = false;
var userId = rawMessage.userId;
if (!this._user || !this._user.muted || !this._user.muted[userId]) {
var self = this;
var origin;
if (rawMessage.isDigitalAssistant)
origin = "ASSISTANT"
else
origin = (this._user && rawMessage.userId == this._user.id) ? "YOU" : "PEER";
// Setup defaults
var message = {
id: rawMessage.id,
localtime: _formatTime(rawMessage.timestamp),
message: (rawMessage.isDigitalAssistant) ? rawMessage.message.body : rawMessage.message || '',
userId: rawMessage.userId,
name: rawMessage.name,
origin: origin,
type: rawMessage.type || 'default',
disableActions: (!self._user || rawMessage.userId == self._user.id)
};
if (!rawMessage.isDigitalAssistant) {
if (message.name.substring(0, 1) == '@') {
message.avatar = message.name.substring(1, 2);
message.name = message.name.substring(3);
} else {
var s = "0" + _hashCode(message.name);
message.avatar = s.substr(s.length - 2);
}
}
if (!rawMessage.isDigitalAssistant) {
message.message = _.map(message.message.split(' '), function(token) {
if (self.urlPattern.test(token) || self.pseudoUrlPattern.test(token)) {
return _linkify(encodeURI(token));
} else {
return _.escape(token);
}
}).join(' ');
message.message = _trimWithEllipsis(message.message, self.maxLengthMessage);
}
this.UIPaintChatMessage(message);
}
};
NKChatUI.prototype._onRemoveMessage = function (roomId, messageId) {
this.UIRemoveChatMessage(messageId);
};
// private helper functions
function _trimWithEllipsis(str, length) {
str = str.replace(/^\s\s*/, '').replace(/\s\s*$/, '');
return (length && str.length <= length) ? str : str.substring(0, length) + '...';
};
function _sortListLexicographically(selector) {
$(selector).children("li").sort(function (a, b) {
var upA = $(a).text().toUpperCase();
var upB = $(b).text().toUpperCase();
return (upA < upB) ? -1 : (upA > upB) ? 1 : 0;
}).appendTo(selector);
};
function _formatTime(timestamp) {
var date = (timestamp) ? new Date(timestamp) : new Date(),
hours = date.getHours() || 12,
minutes = '' + date.getMinutes(),
ampm = (date.getHours() >= 12) ? 'pm' : 'am';
hours = (hours > 12) ? hours - 12 : hours;
minutes = (minutes.length < 2) ? '0' + minutes : minutes;
return '' + hours + ':' + minutes + ampm;
};
function _linkify(str) {
return str
.replace(self.urlPattern, '<a target="_blank" href="$&">$&</a>')
.replace(self.pseudoUrlPattern, '$1<a target="_blank" href="http://$2">$2</a>');
};
function _hashCode(str) | {
var hash = 0;
if (str.length == 0) return hash;
for (var i = 0; i < str.length; i++) {
var char = str.charCodeAt(i);
hash = ((hash<<5)-hash)+char;
hash = hash & hash; // Convert to 32bit integer
}
return ((hash + 2147483647 + 1) % 6) + 1;
} | identifier_body |
|
chat-ui.js | this.maxUserSearchResults = 100;
this.urlPattern = /\b(?:https?|ftp):\/\/[a-z0-9-+&@#\/%?=~_|!:,.;]*[a-z0-9-+&@#\/%=~_|]/gim;
this.pseudoUrlPattern = /(^|[^\/])(www\.[\S]+(\b|$))/gim;
var self = this;
// Initialize the UI
this.UIbindElements();
this.UIScrollToInput();
// Initialize the Chat
this._chat = new NKChatChannel('Noddy KitKat', function(user){
self._user = user;
self.refreshRooms();
self.UIPaintPrimaryUser();
self._chat.on('room-enter', self._onEnterRoom.bind(self));
self._chat.on('message-add', self._onNewMessage.bind(self));
self._chat.on('message-remove', self._onRemoveMessage.bind(self));
});
};
Object.defineProperty(NKChatUI.prototype, "userid", {get: function() { return this._user.id; } });
Object.defineProperty(NKChatUI.prototype, "roomid", {get: function() { return this._roomId; } });
NKChatUI.prototype.UIbindElements = function () {
var self = this;
var _converse = function(userText) {
self.$loading.show();
self.sendMessage(userText, function() {
self.$loading.hide();
self.UIClearInput();
});
}
this.$chatInput.keyup(function (event) {
if (event.keyCode === 13) {
_converse($(this).val());
}
});
this.$chatInputButton.bind('click', { self: this }, function(event) {
_converse(self.$chatInput.val());
});
var _createRoom = function(roomName) {
self.$loading.show();
self.createRoom(roomName, function() {
self.$loading.hide();
self.$roomCreate.val('');
});
}
this.$roomCreateButton.bind('click', { self: this }, function(event) {
_createRoom($roomCreate.val());
});
this.$roomCreate.keyup(function (event) {
if (event.keyCode === 13) {
_createRoom($(this).val());
}
});
};
NKChatUI.prototype.UIScrollChatToBottom = function() {
var element = this.$chatPane;
element.animate({
scrollTop: element[0].scrollHeight
}, 420);
};
NKChatUI.prototype.UIScrollToInput = function() {
var element = this.$chatInput;
$('body, html').animate({
scrollTop: (element.offset().top - window.innerHeight + element[0].offsetHeight) + 20 + 'px'
});
};
NKChatUI.prototype.UIClearInput = function() {
this.$chatInput.val('');
};
NKChatUI.prototype.UIPaintRoomList = function(rooms, cb) {
var self = this;
var template = function(obj) { obj || (obj = {}); var __t, __p = '', __e = _.escape, __j = Array.prototype.join; function print() { __p += __j.call(arguments, '') } with (obj) { __p += '<li data-room-type=\'' + __e(type) + '\' data-room-id=\'' + __e(id) + '\' data-room-name=\'' + __e(name) + '\'>\n<a href=\'#!\' class=\'clearfix '; if (isRoomOpen) { ; __p += ' highlight '; }; __p += '\'>\n<i class=\'fa fa-hashtag\'></i><span class=\'left\' title=\'' + __e(name) + '\'>' + __e(nameTrimmed) + '</span><small class=\'label pull-right bg-green\'>public</small>\n</a>\n</li>'; } return __p };
var selectRoomListItem = function(e) {
var parent = $(this).parent(),
roomId = parent.data('room-id'),
roomName = parent.data('room-name');
self._chat.leaveRoom(self._roomId);
self._roomId = roomId;
self._chat.enterRoom(roomId, roomName);
return false;
};
var count = 0;
this.$roomList.empty();
var keys = Object.keys(rooms);
for (var i = keys.length - 1; i >= 0; i--) {
var roomId = keys[i];
var room = rooms[roomId];
if (room.name == "MyNewRoom")
room.name = "Public"
else if (room.name.substr(0, 2) == "NK" && room.name.length > 2)
room.name = room.name.substr(2);
else
continue;
if (count >= 4 && room.name != "Public") continue;
if (self._roomId == null && room.name == "Public") self._roomId = roomId;
room.isRoomOpen = (roomId == self._roomId);
room.nameTrimmed = _trimWithEllipsis(room.name, self.maxLengthRoomName);
var $roomItem = $(template(room));
$roomItem.children('a').bind('click', selectRoomListItem);
count++;
this.$roomList.append($roomItem.toggle(true));
}
_sortListLexicographically(self.$roomList);
cb();
};
NKChatUI.prototype.UIPaintUserList = function (users, cb) {
var template = function (obj) { obj || (obj = {}); var __t, __p = '', __e = _.escape, __j = Array.prototype.join; function print() { __p += __j.call(arguments, '') } with (obj) {
__p += '<li class=\'list-group-item\' data-user-id=\'' + __e(id) + '\'><b>' + __e(nameTrimmed) + '</b>';
__p += '<span class=\'pull-right\'><img style=\'height: 25px; width: auto\' src=\'img\/avatar-peer' + avatar + '.svg\' alt=\'User profile picture\'></span>';
__p += '\n</li>'; } return __p };
this.$userList.empty();
for (var username in users) {
var user = users[username];
user.disableActions = (!this._user || user.id === this._user.id);
if (user.name.substring(0,1) == '@')
{
user.avatar = user.name.substring(1,2);
user.name = user.name.substring(3);
} else
{
var s = "0" + _hashCode(user.name);
user.avatar = s.substr(s.length-2);
}
user.nameTrimmed = _trimWithEllipsis(user.name, this.maxLengthUsernameDisplay);
user.isMuted = (this._user && this._user.muted && this._user.muted[user.id]);
this.$userList.append(template(user));
}
_sortListLexicographically(this.$userList);
cb();
};
NKChatUI.prototype.UIPaintPrimaryUser = function () {
var self = this;
if (!self._user.avatar)
{
var s = "0" + _hashCode(self._user.name);
self._user.avatar = s.substr(s.length-2);
}
this.$currentUserDivs.each(function()
{
$( this ).html(self._user.name);
})
this.$currentUserAvatars.each(function()
{
$( this ).attr('src', 'img\/avatar-peer' + self._user.avatar + '.svg');
})
};
NKChatUI.prototype.UIClearMessages = function () {
$('.' + this.$$chatItemClass).not('.' + this.$$chatItemHiddenClass).remove();
};
var _scrollTime = (new Date()).getTime();
NKChatUI.prototype.UIPaintChatMessage = function(message) {
var self = this;
var $chatBox = $('.' + this.$$chatItemPrefix + message.origin).first().clone();
$chatBox.find('p').html('<div>' + message.message + '<div class=\'' + this.$$chatItemTitleClass + '\'>' + message.name + '</div>' + ' </div>')
$chatBox.attr('data-message-id', message.messageId);
if (message.avatar)
$chatBox.find("img").eq(0).attr("src", "img/avatar-peer" + message.avatar + ".svg");
$chatBox.insertBefore(this.$loading);
setTimeout(function() {
$chatBox.removeClass(self.$$chatItemHiddenClass);
}, 100);
var newScrollTime = (new Date()).getTime();
if ((newScrollTime - _scrollTime) > 500)
this.UIScrollChatToBottom();
_scrollTime = newScrollTime;
if (!message.messageId)
this.$loading.hide();
};
NKChatUI.prototype.UIRemoveChatMessage = function (messageId) {
$('.' + this.$$chatItemClass + '[data-message-id="' + messageId + '"]').remove()
};
// BRIDGE METHODS BETWEEN CHAT API AND UI METHODS ABOVE
NKChatUI.prototype.refreshRooms = function() {
var self = this;
this._chat.getRoomList(function(rooms) {
self.UIPaint | this.maxLengthUsername = 15;
this.maxLengthUsernameDisplay = 13;
this.maxLengthRoomName = 15;
this.maxLengthMessage = 120; | random_line_split |
|
chat-ui.js | };
NKChatUI.prototype.UIScrollToInput = function() {
var element = this.$chatInput;
$('body, html').animate({
scrollTop: (element.offset().top - window.innerHeight + element[0].offsetHeight) + 20 + 'px'
});
};
NKChatUI.prototype.UIClearInput = function() {
this.$chatInput.val('');
};
NKChatUI.prototype.UIPaintRoomList = function(rooms, cb) {
var self = this;
var template = function(obj) { obj || (obj = {}); var __t, __p = '', __e = _.escape, __j = Array.prototype.join; function print() { __p += __j.call(arguments, '') } with (obj) { __p += '<li data-room-type=\'' + __e(type) + '\' data-room-id=\'' + __e(id) + '\' data-room-name=\'' + __e(name) + '\'>\n<a href=\'#!\' class=\'clearfix '; if (isRoomOpen) { ; __p += ' highlight '; }; __p += '\'>\n<i class=\'fa fa-hashtag\'></i><span class=\'left\' title=\'' + __e(name) + '\'>' + __e(nameTrimmed) + '</span><small class=\'label pull-right bg-green\'>public</small>\n</a>\n</li>'; } return __p };
var selectRoomListItem = function(e) {
var parent = $(this).parent(),
roomId = parent.data('room-id'),
roomName = parent.data('room-name');
self._chat.leaveRoom(self._roomId);
self._roomId = roomId;
self._chat.enterRoom(roomId, roomName);
return false;
};
var count = 0;
this.$roomList.empty();
var keys = Object.keys(rooms);
for (var i = keys.length - 1; i >= 0; i--) {
var roomId = keys[i];
var room = rooms[roomId];
if (room.name == "MyNewRoom")
room.name = "Public"
else if (room.name.substr(0, 2) == "NK" && room.name.length > 2)
room.name = room.name.substr(2);
else
continue;
if (count >= 4 && room.name != "Public") continue;
if (self._roomId == null && room.name == "Public") self._roomId = roomId;
room.isRoomOpen = (roomId == self._roomId);
room.nameTrimmed = _trimWithEllipsis(room.name, self.maxLengthRoomName);
var $roomItem = $(template(room));
$roomItem.children('a').bind('click', selectRoomListItem);
count++;
this.$roomList.append($roomItem.toggle(true));
}
_sortListLexicographically(self.$roomList);
cb();
};
NKChatUI.prototype.UIPaintUserList = function (users, cb) {
var template = function (obj) { obj || (obj = {}); var __t, __p = '', __e = _.escape, __j = Array.prototype.join; function print() { __p += __j.call(arguments, '') } with (obj) {
__p += '<li class=\'list-group-item\' data-user-id=\'' + __e(id) + '\'><b>' + __e(nameTrimmed) + '</b>';
__p += '<span class=\'pull-right\'><img style=\'height: 25px; width: auto\' src=\'img\/avatar-peer' + avatar + '.svg\' alt=\'User profile picture\'></span>';
__p += '\n</li>'; } return __p };
this.$userList.empty();
for (var username in users) {
var user = users[username];
user.disableActions = (!this._user || user.id === this._user.id);
if (user.name.substring(0,1) == '@')
{
user.avatar = user.name.substring(1,2);
user.name = user.name.substring(3);
} else
{
var s = "0" + _hashCode(user.name);
user.avatar = s.substr(s.length-2);
}
user.nameTrimmed = _trimWithEllipsis(user.name, this.maxLengthUsernameDisplay);
user.isMuted = (this._user && this._user.muted && this._user.muted[user.id]);
this.$userList.append(template(user));
}
_sortListLexicographically(this.$userList);
cb();
};
NKChatUI.prototype.UIPaintPrimaryUser = function () {
var self = this;
if (!self._user.avatar)
{
var s = "0" + _hashCode(self._user.name);
self._user.avatar = s.substr(s.length-2);
}
this.$currentUserDivs.each(function()
{
$( this ).html(self._user.name);
})
this.$currentUserAvatars.each(function()
{
$( this ).attr('src', 'img\/avatar-peer' + self._user.avatar + '.svg');
})
};
NKChatUI.prototype.UIClearMessages = function () {
$('.' + this.$$chatItemClass).not('.' + this.$$chatItemHiddenClass).remove();
};
var _scrollTime = (new Date()).getTime();
NKChatUI.prototype.UIPaintChatMessage = function(message) {
var self = this;
var $chatBox = $('.' + this.$$chatItemPrefix + message.origin).first().clone();
$chatBox.find('p').html('<div>' + message.message + '<div class=\'' + this.$$chatItemTitleClass + '\'>' + message.name + '</div>' + ' </div>')
$chatBox.attr('data-message-id', message.messageId);
if (message.avatar)
$chatBox.find("img").eq(0).attr("src", "img/avatar-peer" + message.avatar + ".svg");
$chatBox.insertBefore(this.$loading);
setTimeout(function() {
$chatBox.removeClass(self.$$chatItemHiddenClass);
}, 100);
var newScrollTime = (new Date()).getTime();
if ((newScrollTime - _scrollTime) > 500)
this.UIScrollChatToBottom();
_scrollTime = newScrollTime;
if (!message.messageId)
this.$loading.hide();
};
NKChatUI.prototype.UIRemoveChatMessage = function (messageId) {
$('.' + this.$$chatItemClass + '[data-message-id="' + messageId + '"]').remove()
};
// BRIDGE METHODS BETWEEN CHAT API AND UI METHODS ABOVE
NKChatUI.prototype.refreshRooms = function() {
var self = this;
this._chat.getRoomList(function(rooms) {
self.UIPaintRoomList(rooms, function() {
self._chat.enterRoom(self._roomId);
});
});
};
NKChatUI.prototype.refreshUsers = function () {
var self = this;
this._chat.getUsersByRoom(self._roomId, function(users){
self.UIPaintUserList(users, function(){});
});
};
NKChatUI.prototype.sendMessage = function (msg, cb) {
this._chat.sendMessage(this._roomId, msg, 'default', cb);
};
NKChatUI.prototype.createRoom = function (roomName, cb) {
this._chat.createRoom('NK' + roomName, 'public', cb);
};
NKChatUI.prototype._onEnterRoom = function(room) {
var self = this;
if (room.name == "MyNewRoom")
room.name = "Public";
this.$roomTitle.html(room.name);
this.$roomTitle2.html(room.name);
this._roomId = room.id;
this.UIClearMessages();
this.refreshRooms();
this.refreshUsers();
setTimeout(function() {
var element = self.$chatPane;
element.animate({
scrollTop: element[0].scrollHeight
}, 420);
}, 500);
};
NKChatUI.prototype._onNewMessage = function(roomId, rawMessage) {
if (roomId == this._digitalAssistantRoomId) {
if (rawMessage.message.userid != this._user.id)
return;
rawMessage.isDigitalAssistant = true;
} else
rawMessage.isDigitalAssistant = false;
var userId = rawMessage.userId;
if (!this._user || !this._user.muted || !this._user.muted[userId]) {
var self = this;
var origin;
if (rawMessage.isDigitalAssistant)
origin = "ASSISTANT"
else
origin = (this._user && rawMessage.userId == this._user.id) ? "YOU" : "PEER";
// Setup defaults
var message = {
id: rawMessage.id,
localtime: _formatTime(rawMessage.timestamp),
message: (rawMessage.isDigitalAssistant) ? rawMessage.message.body : rawMessage.message || '',
userId: rawMessage.userId,
name: rawMessage.name,
origin: origin,
type: rawMessage.type || 'default',
disableActions: (!self._user || rawMessage.userId == self._user.id)
};
if (!rawMessage.isDigitalAssistant) {
if (message.name.substring(0, 1) == '@') {
message.avatar = message.name.substring(1, 2);
message.name = message.name.substring(3);
} else | {
var s = "0" + _hashCode(message.name);
message.avatar = s.substr(s.length - 2);
} | conditional_block |
|
chat-ui.js | = Object.keys(rooms);
for (var i = keys.length - 1; i >= 0; i--) {
var roomId = keys[i];
var room = rooms[roomId];
if (room.name == "MyNewRoom")
room.name = "Public"
else if (room.name.substr(0, 2) == "NK" && room.name.length > 2)
room.name = room.name.substr(2);
else
continue;
if (count >= 4 && room.name != "Public") continue;
if (self._roomId == null && room.name == "Public") self._roomId = roomId;
room.isRoomOpen = (roomId == self._roomId);
room.nameTrimmed = _trimWithEllipsis(room.name, self.maxLengthRoomName);
var $roomItem = $(template(room));
$roomItem.children('a').bind('click', selectRoomListItem);
count++;
this.$roomList.append($roomItem.toggle(true));
}
_sortListLexicographically(self.$roomList);
cb();
};
NKChatUI.prototype.UIPaintUserList = function (users, cb) {
var template = function (obj) { obj || (obj = {}); var __t, __p = '', __e = _.escape, __j = Array.prototype.join; function print() { __p += __j.call(arguments, '') } with (obj) {
__p += '<li class=\'list-group-item\' data-user-id=\'' + __e(id) + '\'><b>' + __e(nameTrimmed) + '</b>';
__p += '<span class=\'pull-right\'><img style=\'height: 25px; width: auto\' src=\'img\/avatar-peer' + avatar + '.svg\' alt=\'User profile picture\'></span>';
__p += '\n</li>'; } return __p };
this.$userList.empty();
for (var username in users) {
var user = users[username];
user.disableActions = (!this._user || user.id === this._user.id);
if (user.name.substring(0,1) == '@')
{
user.avatar = user.name.substring(1,2);
user.name = user.name.substring(3);
} else
{
var s = "0" + _hashCode(user.name);
user.avatar = s.substr(s.length-2);
}
user.nameTrimmed = _trimWithEllipsis(user.name, this.maxLengthUsernameDisplay);
user.isMuted = (this._user && this._user.muted && this._user.muted[user.id]);
this.$userList.append(template(user));
}
_sortListLexicographically(this.$userList);
cb();
};
NKChatUI.prototype.UIPaintPrimaryUser = function () {
var self = this;
if (!self._user.avatar)
{
var s = "0" + _hashCode(self._user.name);
self._user.avatar = s.substr(s.length-2);
}
this.$currentUserDivs.each(function()
{
$( this ).html(self._user.name);
})
this.$currentUserAvatars.each(function()
{
$( this ).attr('src', 'img\/avatar-peer' + self._user.avatar + '.svg');
})
};
NKChatUI.prototype.UIClearMessages = function () {
$('.' + this.$$chatItemClass).not('.' + this.$$chatItemHiddenClass).remove();
};
var _scrollTime = (new Date()).getTime();
NKChatUI.prototype.UIPaintChatMessage = function(message) {
var self = this;
var $chatBox = $('.' + this.$$chatItemPrefix + message.origin).first().clone();
$chatBox.find('p').html('<div>' + message.message + '<div class=\'' + this.$$chatItemTitleClass + '\'>' + message.name + '</div>' + ' </div>')
$chatBox.attr('data-message-id', message.messageId);
if (message.avatar)
$chatBox.find("img").eq(0).attr("src", "img/avatar-peer" + message.avatar + ".svg");
$chatBox.insertBefore(this.$loading);
setTimeout(function() {
$chatBox.removeClass(self.$$chatItemHiddenClass);
}, 100);
var newScrollTime = (new Date()).getTime();
if ((newScrollTime - _scrollTime) > 500)
this.UIScrollChatToBottom();
_scrollTime = newScrollTime;
if (!message.messageId)
this.$loading.hide();
};
NKChatUI.prototype.UIRemoveChatMessage = function (messageId) {
$('.' + this.$$chatItemClass + '[data-message-id="' + messageId + '"]').remove()
};
// BRIDGE METHODS BETWEEN CHAT API AND UI METHODS ABOVE
NKChatUI.prototype.refreshRooms = function() {
var self = this;
this._chat.getRoomList(function(rooms) {
self.UIPaintRoomList(rooms, function() {
self._chat.enterRoom(self._roomId);
});
});
};
NKChatUI.prototype.refreshUsers = function () {
var self = this;
this._chat.getUsersByRoom(self._roomId, function(users){
self.UIPaintUserList(users, function(){});
});
};
NKChatUI.prototype.sendMessage = function (msg, cb) {
this._chat.sendMessage(this._roomId, msg, 'default', cb);
};
NKChatUI.prototype.createRoom = function (roomName, cb) {
this._chat.createRoom('NK' + roomName, 'public', cb);
};
NKChatUI.prototype._onEnterRoom = function(room) {
var self = this;
if (room.name == "MyNewRoom")
room.name = "Public";
this.$roomTitle.html(room.name);
this.$roomTitle2.html(room.name);
this._roomId = room.id;
this.UIClearMessages();
this.refreshRooms();
this.refreshUsers();
setTimeout(function() {
var element = self.$chatPane;
element.animate({
scrollTop: element[0].scrollHeight
}, 420);
}, 500);
};
NKChatUI.prototype._onNewMessage = function(roomId, rawMessage) {
if (roomId == this._digitalAssistantRoomId) {
if (rawMessage.message.userid != this._user.id)
return;
rawMessage.isDigitalAssistant = true;
} else
rawMessage.isDigitalAssistant = false;
var userId = rawMessage.userId;
if (!this._user || !this._user.muted || !this._user.muted[userId]) {
var self = this;
var origin;
if (rawMessage.isDigitalAssistant)
origin = "ASSISTANT"
else
origin = (this._user && rawMessage.userId == this._user.id) ? "YOU" : "PEER";
// Setup defaults
var message = {
id: rawMessage.id,
localtime: _formatTime(rawMessage.timestamp),
message: (rawMessage.isDigitalAssistant) ? rawMessage.message.body : rawMessage.message || '',
userId: rawMessage.userId,
name: rawMessage.name,
origin: origin,
type: rawMessage.type || 'default',
disableActions: (!self._user || rawMessage.userId == self._user.id)
};
if (!rawMessage.isDigitalAssistant) {
if (message.name.substring(0, 1) == '@') {
message.avatar = message.name.substring(1, 2);
message.name = message.name.substring(3);
} else {
var s = "0" + _hashCode(message.name);
message.avatar = s.substr(s.length - 2);
}
}
if (!rawMessage.isDigitalAssistant) {
message.message = _.map(message.message.split(' '), function(token) {
if (self.urlPattern.test(token) || self.pseudoUrlPattern.test(token)) {
return _linkify(encodeURI(token));
} else {
return _.escape(token);
}
}).join(' ');
message.message = _trimWithEllipsis(message.message, self.maxLengthMessage);
}
this.UIPaintChatMessage(message);
}
};
NKChatUI.prototype._onRemoveMessage = function (roomId, messageId) {
this.UIRemoveChatMessage(messageId);
};
// private helper functions
function _trimWithEllipsis(str, length) {
str = str.replace(/^\s\s*/, '').replace(/\s\s*$/, '');
return (length && str.length <= length) ? str : str.substring(0, length) + '...';
};
function _sortListLexicographically(selector) {
$(selector).children("li").sort(function (a, b) {
var upA = $(a).text().toUpperCase();
var upB = $(b).text().toUpperCase();
return (upA < upB) ? -1 : (upA > upB) ? 1 : 0;
}).appendTo(selector);
};
function _formatTime(timestamp) {
var date = (timestamp) ? new Date(timestamp) : new Date(),
hours = date.getHours() || 12,
minutes = '' + date.getMinutes(),
ampm = (date.getHours() >= 12) ? 'pm' : 'am';
hours = (hours > 12) ? hours - 12 : hours;
minutes = (minutes.length < 2) ? '0' + minutes : minutes;
return '' + hours + ':' + minutes + ampm;
};
function | _linkify | identifier_name |
|
main.py | 0, 0))
text_line_2 = font_start.render("\\", 1, (0, 0, 0))
level_now = "00"
money_now = "00000"
live_now = "20"
sell_price = 0
# location = [window_size[0] * 0.5, window_size[1] * 0.5]
# dx_dy = [0,0,0,0]
clock = pygame.time.Clock()
playing = False
drop_it = False
tower_type_num = 0
tower_type = [leading_role, \
myst_array, \
tower_wall, \
tower_arrow]
tower_type_size = [leading_role_size, \
myst_array_size, \
tower_size, \
tower_size]
builded_towers = [leading_role, \
myst_array, \
myst_array]
builded_towers_pos = [[(250,500), 0], \
[(250,0), 1], \
[(300,0), 1]]
stop = False
monster_init_pos_0 = [265,15]
monster_init_pos_1 = [315,15]
monsters_pos = {}
monster_path_corner_list = translateMonsterWalkPath(monster_walk_path)
monster_path_point_list = createMonsterWalkPath(monster_path_corner_list, monster_init_pos_1)
crashed = False
while not crashed:
mouse_pos = pygame.mouse.get_pos()
mouse_pos_re = (mouse_pos[0]-25,mouse_pos[1]-25)
text_live = font_live.render(live_now, 1, white, black)
for event in pygame.event.get():
if event.type == pygame.QUIT:
crashed = True
break
# print event
if event.type == pygame.KEYDOWN:
if playing:
if event.key == pygame.K_ESCAPE:
if stop:
stop = False
elif drop_it:
drop_it = False
elif not stop and not drop_it:
stop = True
if stop:
pass
else:
if event.type == pygame.MOUSEBUTTONUP:
if playing:
#-- Right frame:
# Tools
if pygame.Rect((676, 135), sell_0_size).collidepoint(mouse_pos):
money_temp = int(money_now) + sell_price
if money_temp > 99999:
money_temp = 99999
money_now = money_now[0:5-len(str(money_temp))] + str(money_temp)
# Towers
if pygame.Rect((612, 192), tower_size).collidepoint(mouse_pos):
drop_it = True
tower_type_num = 2
if pygame.Rect((675, 192), tower_size).collidepoint(mouse_pos):
drop_it = True
tower_type_num = 3
#--------------------
#-- Left frame:
# Build tower
if drop_it:
build_pos = ((mouse_pos[0]//50)*50,(mouse_pos[1]//50)*50)
no_prob = True
if pygame.Rect(600, 0, 200, 600).collidepoint(build_pos):
continue
for i in builded_towers_pos:
if pygame.Rect(i[0], tower_type_size[i[1]]).collidepoint(build_pos):
no_prob = False
break
if no_prob:
builded_towers.append(tower_type[tower_type_num])
builded_towers_pos.append([build_pos, tower_type_num])
#--------------------
else:
if pygame.Rect((300, 350), play_size).collidepoint(mouse_pos):
playing = True
# end event
if stop:
pass
else:
gameDisplay.fill(white) # background
if playing:
#-- background:
showMonsterWalkPath(monster_path_corner_list)
gameDisplay.blit(shading, (0, 0))
#--------------------
#-- Right frame:
gameDisplay.blit(right_table, (600, 0))
# Infomations
# Level
gameDisplay.blit(level, (613, 13))
gameDisplay.blit(number[int(str(level_now)[0])], (733, 13))
gameDisplay.blit(number[int(str(level_now)[1])], (758, 13))
# Money
gameDisplay.blit(money, (613, 74))
gameDisplay.blit(number[int(str(money_now)[0])], (655, 74))
gameDisplay.blit(number[int(str(money_now)[1])], (680, 74))
gameDisplay.blit(number[int(str(money_now)[2])], (705, 74))
gameDisplay.blit(number[int(str(money_now)[3])], (730, 74))
gameDisplay.blit(number[int(str(money_now)[4])], (755, 74))
# Tools
if not pygame.Rect((739, 135), upgrade_0_size).collidepoint(mouse_pos):
gameDisplay.blit(upgrade_0, (739, 135))
else:
gameDisplay.blit(upgrade_1, (739, 135))
if not pygame.Rect((613, 135), repair_0_size).collidepoint(mouse_pos):
gameDisplay.blit(repair_0, (613, 135))
else:
gameDisplay.blit(repair_1, (613, 135))
if not pygame.Rect((676, 135), sell_0_size).collidepoint(mouse_pos):
gameDisplay.blit(sell_0, (676, 135))
else:
gameDisplay.blit(sell_1, (676, 135))
# Towers
gameDisplay.blit(tower_wall, (612, 192))
if pygame.Rect((612, 192), tower_size).collidepoint(mouse_pos):
gameDisplay.blit(tower_select, (612, 192))
gameDisplay.blit(tower_arrow, (675, 192))
if pygame.Rect((675, 192), tower_size).collidepoint(mouse_pos):
gameDisplay.blit(tower_select, (675, 192))
#--------------------
#-- object
# Towers on map
for i in range(len(builded_towers_pos)):
gameDisplay.blit(builded_towers[i], builded_towers_pos[i][0])
# Live
gameDisplay.blit(text_live, (280, 550))
# Show the block mouse on it
if pygame.Rect((0, 0), (600, 600)).collidepoint(mouse_pos):
gameDisplay.blit(tower_select, (mouse_pos[0]//50*50, mouse_pos[1]//50*50))
# Drop the tower that you want to build
if drop_it:
gameDisplay.blit(tower_type[tower_type_num], mouse_pos_re)
# Monsters
if 0 not in monsters_pos:
monsters_pos[0] = {}
monsters_pos[0]["move_times"] = 0
monsters_pos[0]["pos"] = monster_init_pos_1
showMonster(monster, monsters_pos[0]["pos"])
cannot_move = False
for i in builded_towers_pos:
if pygame.Rect(i[0], tower_type_size[i[1]]).colliderect(pygame.Rect(monsters_pos[0]["pos"],monster_size)) and not i[1] == 1:
cannot_move = True
break
if not cannot_move:
monsters_pos[0]["move_times"] += 1
monsters_pos[0]["pos"] = monster_path_point_list[monsters_pos[0]["move_times"]]
elif pygame.Rect(builded_towers_pos[0][0], leading_role_size).colliderect(pygame.Rect(monsters_pos[0]["pos"],monster_size)):
live_now = str(int(live_now)-1)
del monsters_pos[0]
#--------------------
else:
# Menu page
gameDisplay.blit(text_1, (81, 121))
gameDisplay.blit(text_1, (80, 120))
gameDisplay.blit(text_2, (251, 191))
gameDisplay.blit(text_2, (250, 190))
gameDisplay.blit(text_line, (290, 260))
gameDisplay.blit(text_line, (290, 360)) | gameDisplay.blit(text_line_1, (240, 340)) | random_line_split |
|
main.py | move = [0,50]
if move != [0,0]:
rang = int(data[i+1])
if i+2 < len(data):
if data[i+2] in "0123456789":
rang = int(data[i+1:i+3])
for t in range(rang):
pos[0] += move[0]
pos[1] += move[1]
path.append(copy.copy(pos))
return path
def showMonsterWalkPath(data_list):
"""
showMonsterWalkPath(data_list)
The data_list should be a 2D list.
"""
pos = [0,0]
for i in range(len(data_list)):
if i == 0:
pos = copy.copy(data_list[i])
gameDisplay.blit(monster_walk, pos)
else:
monster_move = False
num_cal = [1,0,0]
dx = (data_list[i][0] - pos[0])/50
dy = (data_list[i][1] - pos[1])/50
if dx < 0 or dy < 0:
num_cal[0] = -1
if dx != 0:
monster_move = True
num_cal[1] = 1
elif dy != 0:
monster_move = True
num_cal[2] = 1
if monster_move:
for t in range(abs(dx+dy)):
pos[0] += num_cal[0]*num_cal[1]*50
pos[1] += num_cal[0]*num_cal[2]*50
gameDisplay.blit(monster_walk, pos)
def createMonsterWalkPath(data_list, init_pos):
"""
createMonsterWalkPath(data_list, init_pos)
"""
path = []
pos = copy.copy(init_pos)
path.append(copy.copy(pos))
monster_size = 20
side_d = (50-monster_size)/2
for i in data_list:
pos_temp = [0,0]
pos_temp[0] = pos[0]-side_d
pos_temp[1] = pos[1]-side_d
dx = i[0] - pos_temp[0]
dy = i[1] - pos_temp[1]
move_info = [1,0,0]
if dx < 0 or dy < 0:
move_info[0] = -1
if dx != 0:
move_info[1] = 1
elif dy != 0:
move_info[2] = 1
for t in range(abs(dx+dy)):
pos[0] += move_info[0]*move_info[1]
pos[1] += move_info[0]*move_info[2]
path.append(copy.copy(pos))
dx = (250+side_d) - pos[0]
dy = (500+side_d) - pos[1]
move_info = [1,0,0]
if dx < 0 or dy < 0:
move_info[0] = -1
if dx != 0:
move_info[1] = 1
elif dy != 0:
move_info[2] = 1
for t in range(abs(dx+dy)-side_d):
pos[0] += move_info[0]*move_info[1]
pos[1] += move_info[0]*move_info[2]
path.append(copy.copy(pos))
return path
def showMonster(monster, pos):
gameDisplay.blit(monster, pos)
def run():
font = pygame.font.SysFont("colonna", 40)
font_live = pygame.font.SysFont("colonna", 40, True)
font_start = pygame.font.SysFont("colonna", 110)
text_1 = font_start.render("-+-[Tower", 1, (0, 0, 0))
text_2 = font_start.render("Defend]-+-", 1, (0, 0, 0))
text_line = font_start.render("____", 1, (0, 0, 0))
text_line_1 = font_start.render("/", 1, (0, 0, 0))
text_line_2 = font_start.render("\\", 1, (0, 0, 0))
level_now = "00"
money_now = "00000"
live_now = "20"
sell_price = 0
# location = [window_size[0] * 0.5, window_size[1] * 0.5]
# dx_dy = [0,0,0,0]
clock = pygame.time.Clock()
playing = False
drop_it = False
tower_type_num = 0
tower_type = [leading_role, \
myst_array, \
tower_wall, \
tower_arrow]
tower_type_size = [leading_role_size, \
myst_array_size, \
tower_size, \
tower_size]
builded_towers = [leading_role, \
myst_array, \
myst_array]
builded_towers_pos = [[(250,500), 0], \
[(250,0), 1], \
[(300,0), 1]]
stop = False
monster_init_pos_0 = [265,15]
monster_init_pos_1 = [315,15]
monsters_pos = {}
monster_path_corner_list = translateMonsterWalkPath(monster_walk_path)
monster_path_point_list = createMonsterWalkPath(monster_path_corner_list, monster_init_pos_1)
crashed = False
while not crashed:
mouse_pos = pygame.mouse.get_pos()
mouse_pos_re = (mouse_pos[0]-25,mouse_pos[1]-25)
text_live = font_live.render(live_now, 1, white, black)
for event in pygame.event.get():
if event.type == pygame.QUIT:
crashed = True
break
# print event
if event.type == pygame.KEYDOWN:
if playing:
if event.key == pygame.K_ESCAPE:
if stop:
stop = False
elif drop_it:
drop_it = False
elif not stop and not drop_it:
stop = True
if stop:
pass
else:
if event.type == pygame.MOUSEBUTTONUP:
if playing:
#-- Right frame:
# Tools
if pygame.Rect((676, 135), sell_0_size).collidepoint(mouse_pos):
money_temp = int(money_now) + sell_price
if money_temp > 99999:
money_temp = 99999
money_now = money_now[0:5-len(str(money_temp))] + str(money_temp)
# Towers
if pygame.Rect((612, 192), tower_size).collidepoint(mouse_pos):
drop_it = True
tower_type_num = 2
if pygame.Rect((675, 192), tower_size).collidepoint(mouse_pos):
drop_it = True
tower_type_num = 3
#--------------------
#-- Left frame:
# Build tower
if drop_it:
build_pos = ((mouse_pos[0]//50)*50,(mouse_pos[1]//50)*50)
no_prob = True
if pygame.Rect(600, 0, 200, 600).collidepoint(build_pos):
continue
for i in builded_towers_pos:
if pygame.Rect(i[0], tower_type_size[i[1]]).collidepoint(build_pos):
no_prob = False
break
if no_prob:
builded_towers.append(tower_type[tower_type_num])
builded_towers_pos.append([build_pos, tower_type_num])
#--------------------
else:
if pygame.Rect((300, 350), play_size).collidepoint(mouse_pos):
playing = True
# end event
if stop:
pass
else:
gameDisplay.fill(white) # background
if playing:
#-- background:
showMonsterWalkPath(monster_path_corner_list)
gameDisplay.blit(shading, (0, 0))
#--------------------
#-- Right frame:
gameDisplay.blit(right_table, (600, 0))
# Infomations
| """
Translate the monster_walk_path to a list and return it.
"""
path = []
pos = [0,0]
for i in range(len(data)):
if i == 0:
pos[0] = 50*(int(data[i])-1)
elif i == 1:
pos[1] = 50*(int(data[i])-1)
path.append(copy.copy(pos))
else:
move = [0,0]
if data[i] == "l":
move = [-50,0]
elif data[i] == "r":
move = [50,0]
elif data[i] == "u":
move = [0,-50]
elif data[i] == "d": | identifier_body |
|
main.py | ,0]
for i in range(len(data_list)):
if i == 0:
pos = copy.copy(data_list[i])
gameDisplay.blit(monster_walk, pos)
else:
monster_move = False
num_cal = [1,0,0]
dx = (data_list[i][0] - pos[0])/50
dy = (data_list[i][1] - pos[1])/50
if dx < 0 or dy < 0:
num_cal[0] = -1
if dx != 0:
monster_move = True
num_cal[1] = 1
elif dy != 0:
monster_move = True
num_cal[2] = 1
if monster_move:
for t in range(abs(dx+dy)):
pos[0] += num_cal[0]*num_cal[1]*50
pos[1] += num_cal[0]*num_cal[2]*50
gameDisplay.blit(monster_walk, pos)
def createMonsterWalkPath(data_list, init_pos):
"""
createMonsterWalkPath(data_list, init_pos)
"""
path = []
pos = copy.copy(init_pos)
path.append(copy.copy(pos))
monster_size = 20
side_d = (50-monster_size)/2
for i in data_list:
pos_temp = [0,0]
pos_temp[0] = pos[0]-side_d
pos_temp[1] = pos[1]-side_d
dx = i[0] - pos_temp[0]
dy = i[1] - pos_temp[1]
move_info = [1,0,0]
if dx < 0 or dy < 0:
move_info[0] = -1
if dx != 0:
move_info[1] = 1
elif dy != 0:
move_info[2] = 1
for t in range(abs(dx+dy)):
pos[0] += move_info[0]*move_info[1]
pos[1] += move_info[0]*move_info[2]
path.append(copy.copy(pos))
dx = (250+side_d) - pos[0]
dy = (500+side_d) - pos[1]
move_info = [1,0,0]
if dx < 0 or dy < 0:
move_info[0] = -1
if dx != 0:
move_info[1] = 1
elif dy != 0:
move_info[2] = 1
for t in range(abs(dx+dy)-side_d):
pos[0] += move_info[0]*move_info[1]
pos[1] += move_info[0]*move_info[2]
path.append(copy.copy(pos))
return path
def showMonster(monster, pos):
gameDisplay.blit(monster, pos)
def run():
font = pygame.font.SysFont("colonna", 40)
font_live = pygame.font.SysFont("colonna", 40, True)
font_start = pygame.font.SysFont("colonna", 110)
text_1 = font_start.render("-+-[Tower", 1, (0, 0, 0))
text_2 = font_start.render("Defend]-+-", 1, (0, 0, 0))
text_line = font_start.render("____", 1, (0, 0, 0))
text_line_1 = font_start.render("/", 1, (0, 0, 0))
text_line_2 = font_start.render("\\", 1, (0, 0, 0))
level_now = "00"
money_now = "00000"
live_now = "20"
sell_price = 0
# location = [window_size[0] * 0.5, window_size[1] * 0.5]
# dx_dy = [0,0,0,0]
clock = pygame.time.Clock()
playing = False
drop_it = False
tower_type_num = 0
tower_type = [leading_role, \
myst_array, \
tower_wall, \
tower_arrow]
tower_type_size = [leading_role_size, \
myst_array_size, \
tower_size, \
tower_size]
builded_towers = [leading_role, \
myst_array, \
myst_array]
builded_towers_pos = [[(250,500), 0], \
[(250,0), 1], \
[(300,0), 1]]
stop = False
monster_init_pos_0 = [265,15]
monster_init_pos_1 = [315,15]
monsters_pos = {}
monster_path_corner_list = translateMonsterWalkPath(monster_walk_path)
monster_path_point_list = createMonsterWalkPath(monster_path_corner_list, monster_init_pos_1)
crashed = False
while not crashed:
mouse_pos = pygame.mouse.get_pos()
mouse_pos_re = (mouse_pos[0]-25,mouse_pos[1]-25)
text_live = font_live.render(live_now, 1, white, black)
for event in pygame.event.get():
if event.type == pygame.QUIT:
crashed = True
break
# print event
if event.type == pygame.KEYDOWN:
if playing:
if event.key == pygame.K_ESCAPE:
if stop:
stop = False
elif drop_it:
|
elif not stop and not drop_it:
stop = True
if stop:
pass
else:
if event.type == pygame.MOUSEBUTTONUP:
if playing:
#-- Right frame:
# Tools
if pygame.Rect((676, 135), sell_0_size).collidepoint(mouse_pos):
money_temp = int(money_now) + sell_price
if money_temp > 99999:
money_temp = 99999
money_now = money_now[0:5-len(str(money_temp))] + str(money_temp)
# Towers
if pygame.Rect((612, 192), tower_size).collidepoint(mouse_pos):
drop_it = True
tower_type_num = 2
if pygame.Rect((675, 192), tower_size).collidepoint(mouse_pos):
drop_it = True
tower_type_num = 3
#--------------------
#-- Left frame:
# Build tower
if drop_it:
build_pos = ((mouse_pos[0]//50)*50,(mouse_pos[1]//50)*50)
no_prob = True
if pygame.Rect(600, 0, 200, 600).collidepoint(build_pos):
continue
for i in builded_towers_pos:
if pygame.Rect(i[0], tower_type_size[i[1]]).collidepoint(build_pos):
no_prob = False
break
if no_prob:
builded_towers.append(tower_type[tower_type_num])
builded_towers_pos.append([build_pos, tower_type_num])
#--------------------
else:
if pygame.Rect((300, 350), play_size).collidepoint(mouse_pos):
playing = True
# end event
if stop:
pass
else:
gameDisplay.fill(white) # background
if playing:
#-- background:
showMonsterWalkPath(monster_path_corner_list)
gameDisplay.blit(shading, (0, 0))
#--------------------
#-- Right frame:
gameDisplay.blit(right_table, (600, 0))
# Infomations
# Level
gameDisplay.blit(level, (613, 13))
gameDisplay.blit(number[int(str(level_now)[0])], (733, 13))
gameDisplay.blit(number[int(str(level_now)[1])], (758, 13))
# Money
gameDisplay.blit(money, (613, 74))
gameDisplay.blit(number[int(str(money_now)[0])], (655, 74))
gameDisplay.blit(number[int(str(money_now)[1])], (680, 74))
gameDisplay.blit(number[int(str(money_now)[2])], (705, 74))
gameDisplay.blit(number[int(str(money_now)[3])], (730, 74))
gameDisplay.blit(number[int(str(money_now)[4])], (755, 74))
# Tools
if not pygame.Rect((739, 135), upgrade_0_size).collidepoint(mouse_pos):
gameDisplay.blit(upgrade_0, (739, 135))
else:
gameDisplay.blit(upgrade_1, (739, 135))
if not pygame.Rect((613, 135), repair_0_size).collidepoint(mouse_pos):
gameDisplay.blit(repair_0 | drop_it = False | conditional_block |
main.py | ,0]
for i in range(len(data_list)):
if i == 0:
pos = copy.copy(data_list[i])
gameDisplay.blit(monster_walk, pos)
else:
monster_move = False
num_cal = [1,0,0]
dx = (data_list[i][0] - pos[0])/50
dy = (data_list[i][1] - pos[1])/50
if dx < 0 or dy < 0:
num_cal[0] = -1
if dx != 0:
monster_move = True
num_cal[1] = 1
elif dy != 0:
monster_move = True
num_cal[2] = 1
if monster_move:
for t in range(abs(dx+dy)):
pos[0] += num_cal[0]*num_cal[1]*50
pos[1] += num_cal[0]*num_cal[2]*50
gameDisplay.blit(monster_walk, pos)
def | (data_list, init_pos):
"""
createMonsterWalkPath(data_list, init_pos)
"""
path = []
pos = copy.copy(init_pos)
path.append(copy.copy(pos))
monster_size = 20
side_d = (50-monster_size)/2
for i in data_list:
pos_temp = [0,0]
pos_temp[0] = pos[0]-side_d
pos_temp[1] = pos[1]-side_d
dx = i[0] - pos_temp[0]
dy = i[1] - pos_temp[1]
move_info = [1,0,0]
if dx < 0 or dy < 0:
move_info[0] = -1
if dx != 0:
move_info[1] = 1
elif dy != 0:
move_info[2] = 1
for t in range(abs(dx+dy)):
pos[0] += move_info[0]*move_info[1]
pos[1] += move_info[0]*move_info[2]
path.append(copy.copy(pos))
dx = (250+side_d) - pos[0]
dy = (500+side_d) - pos[1]
move_info = [1,0,0]
if dx < 0 or dy < 0:
move_info[0] = -1
if dx != 0:
move_info[1] = 1
elif dy != 0:
move_info[2] = 1
for t in range(abs(dx+dy)-side_d):
pos[0] += move_info[0]*move_info[1]
pos[1] += move_info[0]*move_info[2]
path.append(copy.copy(pos))
return path
def showMonster(monster, pos):
gameDisplay.blit(monster, pos)
def run():
font = pygame.font.SysFont("colonna", 40)
font_live = pygame.font.SysFont("colonna", 40, True)
font_start = pygame.font.SysFont("colonna", 110)
text_1 = font_start.render("-+-[Tower", 1, (0, 0, 0))
text_2 = font_start.render("Defend]-+-", 1, (0, 0, 0))
text_line = font_start.render("____", 1, (0, 0, 0))
text_line_1 = font_start.render("/", 1, (0, 0, 0))
text_line_2 = font_start.render("\\", 1, (0, 0, 0))
level_now = "00"
money_now = "00000"
live_now = "20"
sell_price = 0
# location = [window_size[0] * 0.5, window_size[1] * 0.5]
# dx_dy = [0,0,0,0]
clock = pygame.time.Clock()
playing = False
drop_it = False
tower_type_num = 0
tower_type = [leading_role, \
myst_array, \
tower_wall, \
tower_arrow]
tower_type_size = [leading_role_size, \
myst_array_size, \
tower_size, \
tower_size]
builded_towers = [leading_role, \
myst_array, \
myst_array]
builded_towers_pos = [[(250,500), 0], \
[(250,0), 1], \
[(300,0), 1]]
stop = False
monster_init_pos_0 = [265,15]
monster_init_pos_1 = [315,15]
monsters_pos = {}
monster_path_corner_list = translateMonsterWalkPath(monster_walk_path)
monster_path_point_list = createMonsterWalkPath(monster_path_corner_list, monster_init_pos_1)
crashed = False
while not crashed:
mouse_pos = pygame.mouse.get_pos()
mouse_pos_re = (mouse_pos[0]-25,mouse_pos[1]-25)
text_live = font_live.render(live_now, 1, white, black)
for event in pygame.event.get():
if event.type == pygame.QUIT:
crashed = True
break
# print event
if event.type == pygame.KEYDOWN:
if playing:
if event.key == pygame.K_ESCAPE:
if stop:
stop = False
elif drop_it:
drop_it = False
elif not stop and not drop_it:
stop = True
if stop:
pass
else:
if event.type == pygame.MOUSEBUTTONUP:
if playing:
#-- Right frame:
# Tools
if pygame.Rect((676, 135), sell_0_size).collidepoint(mouse_pos):
money_temp = int(money_now) + sell_price
if money_temp > 99999:
money_temp = 99999
money_now = money_now[0:5-len(str(money_temp))] + str(money_temp)
# Towers
if pygame.Rect((612, 192), tower_size).collidepoint(mouse_pos):
drop_it = True
tower_type_num = 2
if pygame.Rect((675, 192), tower_size).collidepoint(mouse_pos):
drop_it = True
tower_type_num = 3
#--------------------
#-- Left frame:
# Build tower
if drop_it:
build_pos = ((mouse_pos[0]//50)*50,(mouse_pos[1]//50)*50)
no_prob = True
if pygame.Rect(600, 0, 200, 600).collidepoint(build_pos):
continue
for i in builded_towers_pos:
if pygame.Rect(i[0], tower_type_size[i[1]]).collidepoint(build_pos):
no_prob = False
break
if no_prob:
builded_towers.append(tower_type[tower_type_num])
builded_towers_pos.append([build_pos, tower_type_num])
#--------------------
else:
if pygame.Rect((300, 350), play_size).collidepoint(mouse_pos):
playing = True
# end event
if stop:
pass
else:
gameDisplay.fill(white) # background
if playing:
#-- background:
showMonsterWalkPath(monster_path_corner_list)
gameDisplay.blit(shading, (0, 0))
#--------------------
#-- Right frame:
gameDisplay.blit(right_table, (600, 0))
# Infomations
# Level
gameDisplay.blit(level, (613, 13))
gameDisplay.blit(number[int(str(level_now)[0])], (733, 13))
gameDisplay.blit(number[int(str(level_now)[1])], (758, 13))
# Money
gameDisplay.blit(money, (613, 74))
gameDisplay.blit(number[int(str(money_now)[0])], (655, 74))
gameDisplay.blit(number[int(str(money_now)[1])], (680, 74))
gameDisplay.blit(number[int(str(money_now)[2])], (705, 74))
gameDisplay.blit(number[int(str(money_now)[3])], (730, 74))
gameDisplay.blit(number[int(str(money_now)[4])], (755, 74))
# Tools
if not pygame.Rect((739, 135), upgrade_0_size).collidepoint(mouse_pos):
gameDisplay.blit(upgrade_0, (739, 135))
else:
gameDisplay.blit(upgrade_1, (739, 135))
if not pygame.Rect((613, 135), repair_0_size).collidepoint(mouse_pos):
gameDisplay.blit(repair | createMonsterWalkPath | identifier_name |
Clout.js | property {number} CONFIG 5
* @property {number} MIDDLEWARE 10
* @property {number} MODEL 15
* @property {number} API 20
* @property {number} CONTROLLER 25
* @const
*/
const CORE_PRIORITY = {
CONFIG: 5,
MIDDLEWARE: 10,
MODEL: 15,
API: 20,
CONTROLLER: 25,
};
const CLOUT_MODULE_PATH = path.join(__dirname, '..');
/**
* Clout application
* @class
*/
class Clout extends EventEmitter {
/**
* @constructor
* @param {path} rootDirectory application directory
*/
constructor(rootDirectory) {
super();
this.process = process;
this.handleProcess();
this.rootDirectory = null;
this.package = {};
this.applicationPackage = {};
this.config = {};
this.logger = { debug };
this.app = null;
this.server = {};
this.modules = [];
this.moduleCache = [];
// expose core libraries
this.utils = utils;
this.async = async;
this._ = _;
this.fs = fs;
// allow application hooks (Synchronous)
this.CORE_PRIORITY = CORE_PRIORITY;
this.hooks = {
start: [],
stop: [],
reload: [],
};
// Load clout configuration
this.config = new Config();
this.config.loadFromDir(path.join(__dirname, '../resources/conf'));
this.applicationPackage = {};
// load clout package.json
this.package = require(path.join(__dirname, '../package.json'));
// load clout modules
if (this.package.modules) {
this.addModules(this.package.modules);
}
if (rootDirectory) {
// set application root directory
this.rootDirectory = path.resolve(rootDirectory);
// load application manifest
['package.json', 'clout.json'].forEach((fileName) => {
const filePath = path.resolve(this.rootDirectory, fileName);
if (!fs.existsSync(filePath)) {
return debug(`${fileName} not found`);
}
return _.merge(this.applicationPackage, require(filePath));
});
this.process.title = `[clout-js v${this.package.version}] ${this.applicationPackage.name}`;
// add rootdir to node_modules
module.paths.unshift(path.join(this.rootDirectory, 'node_modules'));
// load modules from application manifest
if (this.applicationPackage.modules) {
this.addModules(this.applicationPackage.modules);
}
}
// append module configuration
this.modules.forEach(module => this.config.loadFromDir(path.join(module.path, 'conf')));
// append application configuration (Overrides module conf)
this.config.loadFromDir(path.join(this.rootDirectory, 'conf'));
// initialize logger
this.logger = new Logger(this);
// 1) load core hooks
// 2) load application hooks
// 3) load module hooks
this.loadHooksFromDir(CLOUT_MODULE_PATH)
.then(this.loadHooksFromDir(this.rootDirectory))
.then(async () => this.modules.map(async module => this.loadHooksFromDir(module.path)))
.then((moduleHooks) => {
this.initialized = true;
return moduleHooks;
})
.catch(err => console.error(err));
}
/**
* hook into clout runtime
* @param {string} event event name
* @param {Function} fn function to execute
* @param {String} fn._name hook name
* @param {String} fn.group hook group
* @param {Number} priority function priority
* @param {Boolean} override override existing
* @example
* // register a function to the hook
* clout.registerHook('start', function (next) {
* next();
* });
* // invoking an error in clout runtime
* clout.registerHook('start', function (next) {
* next(new Error('Error executing function'));
* });
*/
registerHook(event, fn, priority, override) {
const hasPriority = typeof priority !== 'undefined';
const hasEvent = this.hooks[event];
debug('registerHook:event=%s:fn:priority=%s', event, hasEvent, priority);
if (!hasEvent) {
throw new Error('Invalid Hook Event');
}
if (hasPriority) {
fn.priority = priority;
}
// find existing, override
if (override === true) {
debug('override');
for (let i = 0, l = this.hooks[event].length; i < l; i += 1) {
const hook = this.hooks[event][i];
if (hook._name !== null && hook._name === fn._name && hook.group === fn.group) {
debug('match found, overriden');
this.hooks[event][i] = fn;
return;
}
}
}
// push is no priority
if (!fn.priority) {
debug('push hook (no priority)');
this.hooks[event].push(fn);
return;
}
// find the correct place to register hook
for (let i = 0, l = this.hooks[event].length; i < l; i += 1) {
const tmpPriority = this.hooks[event][i].priority || 99999;
if (fn.priority < tmpPriority) |
}
debug('push hook (lowest priority yet)');
this.hooks[event].push(fn);
}
/**
* Loads hooks from directory
* @param {Path} dir directory
* @return {Promise} promise
*/
loadHooksFromDir(dir) {
const glob = path.join(dir, '/hooks/**/*.js');
const files = utils.getGlobbedFiles(glob);
debug('loadHooksFromDir: %s', dir);
return new Promise((resolve, reject) => {
async.each(files, (file, next) => {
debug('loading hooks from file: %s', String(file));
const hooks = require(file);
const keys = Object.keys(hooks);
keys.forEach((key) => {
const hook = hooks[key];
const args = [];
debug('Loading hook: %s', key);
// create args
if (!hook.event || !hook.fn) {
throw new Error('Hook missing attributes');
}
hook.fn.group = file.split('hooks/')[1].replace('.js', '');
hook.fn._name = key;
args.push(hook.event);
args.push(hook.fn);
if (typeof hook.priority !== 'undefined') {
if (typeof hook.priority === 'string') {
if (!this.CORE_PRIORITY[hook.priority]) {
throw new Error('Invalid priority type');
}
hook.priority = this.CORE_PRIORITY[hook.priority];
}
args.push(hook.priority);
} else {
args.push(null);
}
if (hook.override) {
args.push(true);
}
this.registerHook(...args);
});
next();
}, (err) => {
if (err) {
debug(err);
return reject(err);
}
debug('all hooks loaded from %s', dir);
resolve();
});
});
}
addModules(modules) {
debug('loading modules', JSON.stringify(modules));
modules.forEach(moduleName => this.addModule(moduleName));
}
/**
* Load clout-js node module
* @param {string} moduleName clout node module name
*/
addModule(moduleName) {
if (this.moduleCache.includes(moduleName)) {
debug('module: %s already loaded', moduleName);
return;
}
this.logger.debug('loading module: %s', moduleName);
this.moduleCache.push(moduleName);
const cloutModule = {
name: moduleName,
path: path.dirname(require.resolve(moduleName)),
manifest: {},
};
this.modules.push(cloutModule);
debug(cloutModule);
// load module manifest
['package.json', 'clout.json'].forEach((fileName) => {
const filePath = path.resolve(cloutModule.path, fileName);
if (!fs.existsSync(filePath)) {
return debug(`${fileName} not found`);
}
_.merge(cloutModule.manifest, require(filePath));
});
// load module modules
if (cloutModule.manifest.modules) {
debug('%s loading modules %s', moduleName, cloutModule.manifest.modules);
this.addModules(cloutModule.manifest.modules);
}
}
/**
* Start clout
* @return {Promise} returns a promise
*/
start() {
this.emit('initialized');
if (!this.initialized) {
return new Promise((resolve) => {
setTimeout(() => resolve(this.start()), 100);
});
}
this.emit('start');
return new Promise((resolve, reject) => {
process.nextTick(() => {
async.eachLimit(this.hooks.start, 1, (hook, next) => {
debug('executing', hook.name || hook._name, hook.group);
const hookResponse = hook.apply(this, [next]);
// support promises
if (typeof hookResponse === 'object') {
hookResponse.then(next, err => next(null, err | {
debug('push hook at index %s', String(i));
this.hooks[event].splice(i, 0, fn);
return;
} | conditional_block |
Clout.js | @property {number} CONFIG 5
* @property {number} MIDDLEWARE 10
* @property {number} MODEL 15
* @property {number} API 20
* @property {number} CONTROLLER 25
* @const
*/
const CORE_PRIORITY = {
CONFIG: 5,
MIDDLEWARE: 10,
MODEL: 15,
API: 20,
CONTROLLER: 25,
};
const CLOUT_MODULE_PATH = path.join(__dirname, '..');
/**
* Clout application
* @class
*/
class Clout extends EventEmitter {
/**
* @constructor
* @param {path} rootDirectory application directory
*/
constructor(rootDirectory) {
super();
this.process = process;
this.handleProcess();
this.rootDirectory = null;
this.package = {};
this.applicationPackage = {};
this.config = {};
this.logger = { debug };
this.app = null;
this.server = {};
this.modules = [];
this.moduleCache = [];
// expose core libraries
this.utils = utils;
this.async = async;
this._ = _;
this.fs = fs;
// allow application hooks (Synchronous)
this.CORE_PRIORITY = CORE_PRIORITY;
this.hooks = {
start: [],
stop: [],
reload: [],
};
// Load clout configuration
this.config = new Config();
this.config.loadFromDir(path.join(__dirname, '../resources/conf'));
this.applicationPackage = {};
// load clout package.json
this.package = require(path.join(__dirname, '../package.json'));
// load clout modules
if (this.package.modules) {
this.addModules(this.package.modules);
}
if (rootDirectory) {
// set application root directory
this.rootDirectory = path.resolve(rootDirectory);
// load application manifest
['package.json', 'clout.json'].forEach((fileName) => {
const filePath = path.resolve(this.rootDirectory, fileName);
if (!fs.existsSync(filePath)) {
return debug(`${fileName} not found`);
}
return _.merge(this.applicationPackage, require(filePath));
});
this.process.title = `[clout-js v${this.package.version}] ${this.applicationPackage.name}`;
// add rootdir to node_modules
module.paths.unshift(path.join(this.rootDirectory, 'node_modules'));
// load modules from application manifest
if (this.applicationPackage.modules) {
this.addModules(this.applicationPackage.modules);
}
}
// append module configuration
this.modules.forEach(module => this.config.loadFromDir(path.join(module.path, 'conf')));
// append application configuration (Overrides module conf)
this.config.loadFromDir(path.join(this.rootDirectory, 'conf'));
// initialize logger
this.logger = new Logger(this);
// 1) load core hooks
// 2) load application hooks
// 3) load module hooks
this.loadHooksFromDir(CLOUT_MODULE_PATH)
.then(this.loadHooksFromDir(this.rootDirectory))
.then(async () => this.modules.map(async module => this.loadHooksFromDir(module.path)))
.then((moduleHooks) => {
this.initialized = true;
return moduleHooks;
})
.catch(err => console.error(err));
}
/**
* hook into clout runtime
* @param {string} event event name
* @param {Function} fn function to execute
* @param {String} fn._name hook name
* @param {String} fn.group hook group
* @param {Number} priority function priority
* @param {Boolean} override override existing
* @example
* // register a function to the hook
* clout.registerHook('start', function (next) {
* next();
* });
* // invoking an error in clout runtime
* clout.registerHook('start', function (next) {
* next(new Error('Error executing function'));
* });
*/
registerHook(event, fn, priority, override) {
const hasPriority = typeof priority !== 'undefined';
const hasEvent = this.hooks[event];
debug('registerHook:event=%s:fn:priority=%s', event, hasEvent, priority);
if (!hasEvent) {
throw new Error('Invalid Hook Event');
}
if (hasPriority) {
fn.priority = priority;
}
// find existing, override
if (override === true) {
debug('override');
for (let i = 0, l = this.hooks[event].length; i < l; i += 1) {
const hook = this.hooks[event][i];
if (hook._name !== null && hook._name === fn._name && hook.group === fn.group) {
debug('match found, overriden');
this.hooks[event][i] = fn;
return;
}
}
}
// push is no priority
if (!fn.priority) {
debug('push hook (no priority)');
this.hooks[event].push(fn);
return;
}
// find the correct place to register hook
for (let i = 0, l = this.hooks[event].length; i < l; i += 1) {
const tmpPriority = this.hooks[event][i].priority || 99999;
if (fn.priority < tmpPriority) {
debug('push hook at index %s', String(i));
this.hooks[event].splice(i, 0, fn);
return;
}
}
debug('push hook (lowest priority yet)');
this.hooks[event].push(fn); | * @param {Path} dir directory
* @return {Promise} promise
*/
loadHooksFromDir(dir) {
const glob = path.join(dir, '/hooks/**/*.js');
const files = utils.getGlobbedFiles(glob);
debug('loadHooksFromDir: %s', dir);
return new Promise((resolve, reject) => {
async.each(files, (file, next) => {
debug('loading hooks from file: %s', String(file));
const hooks = require(file);
const keys = Object.keys(hooks);
keys.forEach((key) => {
const hook = hooks[key];
const args = [];
debug('Loading hook: %s', key);
// create args
if (!hook.event || !hook.fn) {
throw new Error('Hook missing attributes');
}
hook.fn.group = file.split('hooks/')[1].replace('.js', '');
hook.fn._name = key;
args.push(hook.event);
args.push(hook.fn);
if (typeof hook.priority !== 'undefined') {
if (typeof hook.priority === 'string') {
if (!this.CORE_PRIORITY[hook.priority]) {
throw new Error('Invalid priority type');
}
hook.priority = this.CORE_PRIORITY[hook.priority];
}
args.push(hook.priority);
} else {
args.push(null);
}
if (hook.override) {
args.push(true);
}
this.registerHook(...args);
});
next();
}, (err) => {
if (err) {
debug(err);
return reject(err);
}
debug('all hooks loaded from %s', dir);
resolve();
});
});
}
addModules(modules) {
debug('loading modules', JSON.stringify(modules));
modules.forEach(moduleName => this.addModule(moduleName));
}
/**
* Load clout-js node module
* @param {string} moduleName clout node module name
*/
addModule(moduleName) {
if (this.moduleCache.includes(moduleName)) {
debug('module: %s already loaded', moduleName);
return;
}
this.logger.debug('loading module: %s', moduleName);
this.moduleCache.push(moduleName);
const cloutModule = {
name: moduleName,
path: path.dirname(require.resolve(moduleName)),
manifest: {},
};
this.modules.push(cloutModule);
debug(cloutModule);
// load module manifest
['package.json', 'clout.json'].forEach((fileName) => {
const filePath = path.resolve(cloutModule.path, fileName);
if (!fs.existsSync(filePath)) {
return debug(`${fileName} not found`);
}
_.merge(cloutModule.manifest, require(filePath));
});
// load module modules
if (cloutModule.manifest.modules) {
debug('%s loading modules %s', moduleName, cloutModule.manifest.modules);
this.addModules(cloutModule.manifest.modules);
}
}
/**
* Start clout
* @return {Promise} returns a promise
*/
start() {
this.emit('initialized');
if (!this.initialized) {
return new Promise((resolve) => {
setTimeout(() => resolve(this.start()), 100);
});
}
this.emit('start');
return new Promise((resolve, reject) => {
process.nextTick(() => {
async.eachLimit(this.hooks.start, 1, (hook, next) => {
debug('executing', hook.name || hook._name, hook.group);
const hookResponse = hook.apply(this, [next]);
// support promises
if (typeof hookResponse === 'object') {
hookResponse.then(next, err => next(null, err));
| }
/**
* Loads hooks from directory | random_line_split |
Clout.js | property {number} CONFIG 5
* @property {number} MIDDLEWARE 10
* @property {number} MODEL 15
* @property {number} API 20
* @property {number} CONTROLLER 25
* @const
*/
const CORE_PRIORITY = {
CONFIG: 5,
MIDDLEWARE: 10,
MODEL: 15,
API: 20,
CONTROLLER: 25,
};
const CLOUT_MODULE_PATH = path.join(__dirname, '..');
/**
* Clout application
* @class
*/
class Clout extends EventEmitter {
/**
* @constructor
* @param {path} rootDirectory application directory
*/
constructor(rootDirectory) {
super();
this.process = process;
this.handleProcess();
this.rootDirectory = null;
this.package = {};
this.applicationPackage = {};
this.config = {};
this.logger = { debug };
this.app = null;
this.server = {};
this.modules = [];
this.moduleCache = [];
// expose core libraries
this.utils = utils;
this.async = async;
this._ = _;
this.fs = fs;
// allow application hooks (Synchronous)
this.CORE_PRIORITY = CORE_PRIORITY;
this.hooks = {
start: [],
stop: [],
reload: [],
};
// Load clout configuration
this.config = new Config();
this.config.loadFromDir(path.join(__dirname, '../resources/conf'));
this.applicationPackage = {};
// load clout package.json
this.package = require(path.join(__dirname, '../package.json'));
// load clout modules
if (this.package.modules) {
this.addModules(this.package.modules);
}
if (rootDirectory) {
// set application root directory
this.rootDirectory = path.resolve(rootDirectory);
// load application manifest
['package.json', 'clout.json'].forEach((fileName) => {
const filePath = path.resolve(this.rootDirectory, fileName);
if (!fs.existsSync(filePath)) {
return debug(`${fileName} not found`);
}
return _.merge(this.applicationPackage, require(filePath));
});
this.process.title = `[clout-js v${this.package.version}] ${this.applicationPackage.name}`;
// add rootdir to node_modules
module.paths.unshift(path.join(this.rootDirectory, 'node_modules'));
// load modules from application manifest
if (this.applicationPackage.modules) {
this.addModules(this.applicationPackage.modules);
}
}
// append module configuration
this.modules.forEach(module => this.config.loadFromDir(path.join(module.path, 'conf')));
// append application configuration (Overrides module conf)
this.config.loadFromDir(path.join(this.rootDirectory, 'conf'));
// initialize logger
this.logger = new Logger(this);
// 1) load core hooks
// 2) load application hooks
// 3) load module hooks
this.loadHooksFromDir(CLOUT_MODULE_PATH)
.then(this.loadHooksFromDir(this.rootDirectory))
.then(async () => this.modules.map(async module => this.loadHooksFromDir(module.path)))
.then((moduleHooks) => {
this.initialized = true;
return moduleHooks;
})
.catch(err => console.error(err));
}
/**
* hook into clout runtime
* @param {string} event event name
* @param {Function} fn function to execute
* @param {String} fn._name hook name
* @param {String} fn.group hook group
* @param {Number} priority function priority
* @param {Boolean} override override existing
* @example
* // register a function to the hook
* clout.registerHook('start', function (next) {
* next();
* });
* // invoking an error in clout runtime
* clout.registerHook('start', function (next) {
* next(new Error('Error executing function'));
* });
*/
registerHook(event, fn, priority, override) {
const hasPriority = typeof priority !== 'undefined';
const hasEvent = this.hooks[event];
debug('registerHook:event=%s:fn:priority=%s', event, hasEvent, priority);
if (!hasEvent) {
throw new Error('Invalid Hook Event');
}
if (hasPriority) {
fn.priority = priority;
}
// find existing, override
if (override === true) {
debug('override');
for (let i = 0, l = this.hooks[event].length; i < l; i += 1) {
const hook = this.hooks[event][i];
if (hook._name !== null && hook._name === fn._name && hook.group === fn.group) {
debug('match found, overriden');
this.hooks[event][i] = fn;
return;
}
}
}
// push is no priority
if (!fn.priority) {
debug('push hook (no priority)');
this.hooks[event].push(fn);
return;
}
// find the correct place to register hook
for (let i = 0, l = this.hooks[event].length; i < l; i += 1) {
const tmpPriority = this.hooks[event][i].priority || 99999;
if (fn.priority < tmpPriority) {
debug('push hook at index %s', String(i));
this.hooks[event].splice(i, 0, fn);
return;
}
}
debug('push hook (lowest priority yet)');
this.hooks[event].push(fn);
}
/**
* Loads hooks from directory
* @param {Path} dir directory
* @return {Promise} promise
*/
loadHooksFromDir(dir) {
const glob = path.join(dir, '/hooks/**/*.js');
const files = utils.getGlobbedFiles(glob);
debug('loadHooksFromDir: %s', dir);
return new Promise((resolve, reject) => {
async.each(files, (file, next) => {
debug('loading hooks from file: %s', String(file));
const hooks = require(file);
const keys = Object.keys(hooks);
keys.forEach((key) => {
const hook = hooks[key];
const args = [];
debug('Loading hook: %s', key);
// create args
if (!hook.event || !hook.fn) {
throw new Error('Hook missing attributes');
}
hook.fn.group = file.split('hooks/')[1].replace('.js', '');
hook.fn._name = key;
args.push(hook.event);
args.push(hook.fn);
if (typeof hook.priority !== 'undefined') {
if (typeof hook.priority === 'string') {
if (!this.CORE_PRIORITY[hook.priority]) {
throw new Error('Invalid priority type');
}
hook.priority = this.CORE_PRIORITY[hook.priority];
}
args.push(hook.priority);
} else {
args.push(null);
}
if (hook.override) {
args.push(true);
}
this.registerHook(...args);
});
next();
}, (err) => {
if (err) {
debug(err);
return reject(err);
}
debug('all hooks loaded from %s', dir);
resolve();
});
});
}
addModules(modules) |
/**
* Load clout-js node module
* @param {string} moduleName clout node module name
*/
addModule(moduleName) {
if (this.moduleCache.includes(moduleName)) {
debug('module: %s already loaded', moduleName);
return;
}
this.logger.debug('loading module: %s', moduleName);
this.moduleCache.push(moduleName);
const cloutModule = {
name: moduleName,
path: path.dirname(require.resolve(moduleName)),
manifest: {},
};
this.modules.push(cloutModule);
debug(cloutModule);
// load module manifest
['package.json', 'clout.json'].forEach((fileName) => {
const filePath = path.resolve(cloutModule.path, fileName);
if (!fs.existsSync(filePath)) {
return debug(`${fileName} not found`);
}
_.merge(cloutModule.manifest, require(filePath));
});
// load module modules
if (cloutModule.manifest.modules) {
debug('%s loading modules %s', moduleName, cloutModule.manifest.modules);
this.addModules(cloutModule.manifest.modules);
}
}
/**
* Start clout
* @return {Promise} returns a promise
*/
start() {
this.emit('initialized');
if (!this.initialized) {
return new Promise((resolve) => {
setTimeout(() => resolve(this.start()), 100);
});
}
this.emit('start');
return new Promise((resolve, reject) => {
process.nextTick(() => {
async.eachLimit(this.hooks.start, 1, (hook, next) => {
debug('executing', hook.name || hook._name, hook.group);
const hookResponse = hook.apply(this, [next]);
// support promises
if (typeof hookResponse === 'object') {
hookResponse.then(next, err => next(null, err | {
debug('loading modules', JSON.stringify(modules));
modules.forEach(moduleName => this.addModule(moduleName));
} | identifier_body |
Clout.js | @property {number} CONFIG 5
* @property {number} MIDDLEWARE 10
* @property {number} MODEL 15
* @property {number} API 20
* @property {number} CONTROLLER 25
* @const
*/
const CORE_PRIORITY = {
CONFIG: 5,
MIDDLEWARE: 10,
MODEL: 15,
API: 20,
CONTROLLER: 25,
};
const CLOUT_MODULE_PATH = path.join(__dirname, '..');
/**
* Clout application
* @class
*/
class Clout extends EventEmitter {
/**
* @constructor
* @param {path} rootDirectory application directory
*/
constructor(rootDirectory) {
super();
this.process = process;
this.handleProcess();
this.rootDirectory = null;
this.package = {};
this.applicationPackage = {};
this.config = {};
this.logger = { debug };
this.app = null;
this.server = {};
this.modules = [];
this.moduleCache = [];
// expose core libraries
this.utils = utils;
this.async = async;
this._ = _;
this.fs = fs;
// allow application hooks (Synchronous)
this.CORE_PRIORITY = CORE_PRIORITY;
this.hooks = {
start: [],
stop: [],
reload: [],
};
// Load clout configuration
this.config = new Config();
this.config.loadFromDir(path.join(__dirname, '../resources/conf'));
this.applicationPackage = {};
// load clout package.json
this.package = require(path.join(__dirname, '../package.json'));
// load clout modules
if (this.package.modules) {
this.addModules(this.package.modules);
}
if (rootDirectory) {
// set application root directory
this.rootDirectory = path.resolve(rootDirectory);
// load application manifest
['package.json', 'clout.json'].forEach((fileName) => {
const filePath = path.resolve(this.rootDirectory, fileName);
if (!fs.existsSync(filePath)) {
return debug(`${fileName} not found`);
}
return _.merge(this.applicationPackage, require(filePath));
});
this.process.title = `[clout-js v${this.package.version}] ${this.applicationPackage.name}`;
// add rootdir to node_modules
module.paths.unshift(path.join(this.rootDirectory, 'node_modules'));
// load modules from application manifest
if (this.applicationPackage.modules) {
this.addModules(this.applicationPackage.modules);
}
}
// append module configuration
this.modules.forEach(module => this.config.loadFromDir(path.join(module.path, 'conf')));
// append application configuration (Overrides module conf)
this.config.loadFromDir(path.join(this.rootDirectory, 'conf'));
// initialize logger
this.logger = new Logger(this);
// 1) load core hooks
// 2) load application hooks
// 3) load module hooks
this.loadHooksFromDir(CLOUT_MODULE_PATH)
.then(this.loadHooksFromDir(this.rootDirectory))
.then(async () => this.modules.map(async module => this.loadHooksFromDir(module.path)))
.then((moduleHooks) => {
this.initialized = true;
return moduleHooks;
})
.catch(err => console.error(err));
}
/**
* hook into clout runtime
* @param {string} event event name
* @param {Function} fn function to execute
* @param {String} fn._name hook name
* @param {String} fn.group hook group
* @param {Number} priority function priority
* @param {Boolean} override override existing
* @example
* // register a function to the hook
* clout.registerHook('start', function (next) {
* next();
* });
* // invoking an error in clout runtime
* clout.registerHook('start', function (next) {
* next(new Error('Error executing function'));
* });
*/
registerHook(event, fn, priority, override) {
const hasPriority = typeof priority !== 'undefined';
const hasEvent = this.hooks[event];
debug('registerHook:event=%s:fn:priority=%s', event, hasEvent, priority);
if (!hasEvent) {
throw new Error('Invalid Hook Event');
}
if (hasPriority) {
fn.priority = priority;
}
// find existing, override
if (override === true) {
debug('override');
for (let i = 0, l = this.hooks[event].length; i < l; i += 1) {
const hook = this.hooks[event][i];
if (hook._name !== null && hook._name === fn._name && hook.group === fn.group) {
debug('match found, overriden');
this.hooks[event][i] = fn;
return;
}
}
}
// push is no priority
if (!fn.priority) {
debug('push hook (no priority)');
this.hooks[event].push(fn);
return;
}
// find the correct place to register hook
for (let i = 0, l = this.hooks[event].length; i < l; i += 1) {
const tmpPriority = this.hooks[event][i].priority || 99999;
if (fn.priority < tmpPriority) {
debug('push hook at index %s', String(i));
this.hooks[event].splice(i, 0, fn);
return;
}
}
debug('push hook (lowest priority yet)');
this.hooks[event].push(fn);
}
/**
* Loads hooks from directory
* @param {Path} dir directory
* @return {Promise} promise
*/
| (dir) {
const glob = path.join(dir, '/hooks/**/*.js');
const files = utils.getGlobbedFiles(glob);
debug('loadHooksFromDir: %s', dir);
return new Promise((resolve, reject) => {
async.each(files, (file, next) => {
debug('loading hooks from file: %s', String(file));
const hooks = require(file);
const keys = Object.keys(hooks);
keys.forEach((key) => {
const hook = hooks[key];
const args = [];
debug('Loading hook: %s', key);
// create args
if (!hook.event || !hook.fn) {
throw new Error('Hook missing attributes');
}
hook.fn.group = file.split('hooks/')[1].replace('.js', '');
hook.fn._name = key;
args.push(hook.event);
args.push(hook.fn);
if (typeof hook.priority !== 'undefined') {
if (typeof hook.priority === 'string') {
if (!this.CORE_PRIORITY[hook.priority]) {
throw new Error('Invalid priority type');
}
hook.priority = this.CORE_PRIORITY[hook.priority];
}
args.push(hook.priority);
} else {
args.push(null);
}
if (hook.override) {
args.push(true);
}
this.registerHook(...args);
});
next();
}, (err) => {
if (err) {
debug(err);
return reject(err);
}
debug('all hooks loaded from %s', dir);
resolve();
});
});
}
addModules(modules) {
debug('loading modules', JSON.stringify(modules));
modules.forEach(moduleName => this.addModule(moduleName));
}
/**
* Load clout-js node module
* @param {string} moduleName clout node module name
*/
addModule(moduleName) {
if (this.moduleCache.includes(moduleName)) {
debug('module: %s already loaded', moduleName);
return;
}
this.logger.debug('loading module: %s', moduleName);
this.moduleCache.push(moduleName);
const cloutModule = {
name: moduleName,
path: path.dirname(require.resolve(moduleName)),
manifest: {},
};
this.modules.push(cloutModule);
debug(cloutModule);
// load module manifest
['package.json', 'clout.json'].forEach((fileName) => {
const filePath = path.resolve(cloutModule.path, fileName);
if (!fs.existsSync(filePath)) {
return debug(`${fileName} not found`);
}
_.merge(cloutModule.manifest, require(filePath));
});
// load module modules
if (cloutModule.manifest.modules) {
debug('%s loading modules %s', moduleName, cloutModule.manifest.modules);
this.addModules(cloutModule.manifest.modules);
}
}
/**
* Start clout
* @return {Promise} returns a promise
*/
start() {
this.emit('initialized');
if (!this.initialized) {
return new Promise((resolve) => {
setTimeout(() => resolve(this.start()), 100);
});
}
this.emit('start');
return new Promise((resolve, reject) => {
process.nextTick(() => {
async.eachLimit(this.hooks.start, 1, (hook, next) => {
debug('executing', hook.name || hook._name, hook.group);
const hookResponse = hook.apply(this, [next]);
// support promises
if (typeof hookResponse === 'object') {
hookResponse.then(next, err => next(null, err));
| loadHooksFromDir | identifier_name |
htp1.py | name': self.__name,
'slots': [self.slot.as_dict()]
}
class Htp1(PersistentDevice[Htp1State]):
def __init__(self, name: str, config_path: str, cfg: dict, ws_server: WsServer, catalogue: CatalogueProvider):
super().__init__(config_path, name, ws_server)
self.__name = name
self.__catalogue = catalogue
self.__ip = cfg['ip']
self.__channels = cfg['channels']
self.__peq = {}
self.__supports_shelf = True
if not self.__channels:
raise ValueError('No channels supplied for HTP-1')
self.__client = Htp1Client(self.__ip, self)
def _load_initial_state(self) -> Htp1State:
return Htp1State(self.name)
def _merge_state(self, loaded: Htp1State, cached: dict) -> Htp1State:
if 'slots' in cached:
for slot in cached['slots']:
if 'id' in slot:
if slot['id'] == 'HTP1':
if slot['last']:
loaded.slot.last = slot['last']
return loaded
@property
def device_type(self) -> str:
return self.__class__.__name__.lower()
def update(self, params: dict) -> bool:
any_update = False
if 'slots' in params:
for slot in params['slots']:
if slot['id'] == 'HTP1':
if 'entry' in slot:
if slot['entry']:
match = self.__catalogue.find(slot['entry'])
if match:
self.load_filter('HTP1', match)
any_update = True
else:
self.clear_filter('HTP1')
any_update = True
return any_update
def __send(self, to_load: List['PEQ']):
logger.info(f"Sending {len(to_load)} filters")
while len(to_load) < 16:
peq = PEQ(len(to_load), fc=100, q=1, gain=0, filter_type_name='PeakingEQ')
to_load.append(peq)
ops = [peq.as_ops(c, use_shelf=self.__supports_shelf) for peq in to_load for c in self.__peq.keys()]
ops = [op for slot_ops in ops for op in slot_ops if op]
if ops:
self.__client.send('changemso [{"op":"replace","path":"/peq/peqsw","value":true}]')
self.__client.send(f"changemso {json.dumps(ops)}")
else:
logger.warning(f"Nothing to send")
def activate(self, slot: str) -> None:
def __do_it():
self._current_state.slot.active = True
self._hydrate_cache_broadcast(__do_it)
def load_filter(self, slot: str, entry: CatalogueEntry) -> None:
to_load = [PEQ(idx, fc=f['freq'], q=f['q'], gain=f['gain'], filter_type_name=f['type'])
for idx, f in enumerate(entry.filters)]
self._hydrate_cache_broadcast(lambda: self.__do_it(to_load, entry.formatted_title))
def __do_it(self, to_load: List['PEQ'], title: str):
try:
self.__send(to_load)
self._current_state.slot.last = title
except Exception as e:
self._current_state.slot.last = 'ERRUR'
raise e
def clear_filter(self, slot: str) -> None:
self._hydrate_cache_broadcast(lambda: self.__do_it([], 'Empty'))
def mute(self, slot: Optional[str], channel: Optional[int]) -> None:
raise NotImplementedError()
def unmute(self, slot: Optional[str], channel: Optional[int]) -> None:
raise NotImplementedError()
def set_gain(self, slot: Optional[str], channel: Optional[int], gain: float) -> None:
raise NotImplementedError()
def on_mso(self, mso: dict):
logger.info(f"Received {mso}")
version = mso['versions']['swVer']
version = version[1:] if version[0] == 'v' or version[0] == 'V' else version
try:
self.__supports_shelf = semver.parse_version_info(version) > semver.parse_version_info('1.4.0')
except:
logger.error(f"Unable to parse version {mso['versions']['swVer']}, will not send shelf filters")
self.__supports_shelf = False
if not self.__supports_shelf:
logger.error(f"Device version {mso['versions']['swVer']} too old, lacks shelf filter support")
speakers = mso['speakers']['groups']
channels = ['lf', 'rf']
for group in [s for s, v in speakers.items() if 'present' in v and v['present'] is True]:
if group[0:2] == 'lr' and len(group) > 2:
channels.append('l' + group[2:])
channels.append('r' + group[2:])
else:
channels.append(group)
peq_slots = mso['peq']['slots']
filters = {c: [] for c in channels}
unknown_channels = set()
for idx, s in enumerate(peq_slots):
for c in channels:
if c in s['channels']:
filters[c].append(PEQ(idx, s['channels'][c]))
else:
unknown_channels.add(c)
if unknown_channels:
peq_channels = peq_slots[0]['channels'].keys()
logger.error(f"Unknown channels encountered [peq channels: {peq_channels}, unknown: {unknown_channels}]")
for c in filters.keys():
if c in self.__channels:
logger.info(f"Updating PEQ channel {c} with {filters[c]}")
self.__peq[c] = filters[c]
else:
logger.info(f"Discarding filter channel {c} - {filters[c]}")
def on_msoupdate(self, msoupdate: dict):
logger.info(f"Received {msoupdate}")
class Htp1Client:
def __init__(self, ip, listener):
self.__factory = Htp1ClientFactory(listener, f"ws://{ip}/ws/controller")
self.__connector = connectWS(self.__factory)
def send(self, msg: str):
self.__factory.broadcast(msg)
class Htp1Protocol(WebSocketClientProtocol):
def onConnecting(self, transport_details):
logger.info(f"Connecting to {transport_details}")
def onConnect(self, response):
logger.info(f"Connected to {response.peer}") | self.sendMessage('getmso'.encode('utf-8'), isBinary=False)
def onOpen(self):
logger.info("Connected to HTP1")
self.factory.register(self)
def onClose(self, was_clean, code, reason):
if was_clean:
logger.info(f"Disconnected code: {code} reason: {reason}")
else:
logger.warning(f"UNCLEAN! Disconnected code: {code} reason: {reason}")
def onMessage(self, payload, is_binary):
if is_binary:
logger.warning(f"Received {len(payload)} bytes in binary payload, ignoring")
else:
msg = payload.decode('utf8')
if msg.startswith('mso '):
logger.debug(f"Processing mso {msg}")
self.factory.listener.on_mso(json.loads(msg[4:]))
elif msg.startswith('msoupdate '):
logger.debug(f"Processing msoupdate {msg}")
self.factory.listener.on_msoupdate(json.loads(msg[10:]))
else:
logger.info(f"Received unknown payload {msg}")
class Htp1ClientFactory(WebSocketClientFactory, ReconnectingClientFactory):
protocol = Htp1Protocol
maxDelay = 5
initialDelay = 0.5
def __init__(self, listener, *args, **kwargs):
super(Htp1ClientFactory, self).__init__(*args, **kwargs)
self.__clients: List[Htp1Protocol] = []
self.listener = listener
self.setProtocolOptions(version=13)
def clientConnectionFailed(self, connector, reason):
logger.warning(f"Client connection failed {reason} .. retrying ..")
super().clientConnectionFailed(connector, reason)
def clientConnectionLost(self, connector, reason):
logger.warning(f"Client connection failed {reason} .. retrying ..")
super().clientConnectionLost(connector, reason)
def register(self, client: Htp1Protocol):
if client not in self.__clients:
logger.info(f"Registered device {client.peer}")
self.__clients.append(client)
else:
logger.info(f"Ignoring duplicate device {client.peer}")
def unregister(self, client: Htp1Protocol):
if client in self.__clients:
logger.info(f"Unregistering device {client.peer}")
self.__clients.remove(client)
else:
logger.info(f"Ignoring unregistered device {client.peer}")
def broadcast(self, msg):
if self.__clients:
disconnected_clients = []
for c in self.__clients:
logger.info(f"Sending to {c.peer} - {msg}")
try:
c.sendMessage(msg.encode('utf8'))
except Disconnected as e:
logger.exception(f"Failed to send to {c.peer}, discarding")
disconnected | random_line_split |
|
htp1.py | name': self.__name,
'slots': [self.slot.as_dict()]
}
class Htp1(PersistentDevice[Htp1State]):
def __init__(self, name: str, config_path: str, cfg: dict, ws_server: WsServer, catalogue: CatalogueProvider):
super().__init__(config_path, name, ws_server)
self.__name = name
self.__catalogue = catalogue
self.__ip = cfg['ip']
self.__channels = cfg['channels']
self.__peq = {}
self.__supports_shelf = True
if not self.__channels:
raise ValueError('No channels supplied for HTP-1')
self.__client = Htp1Client(self.__ip, self)
def _load_initial_state(self) -> Htp1State:
return Htp1State(self.name)
def _merge_state(self, loaded: Htp1State, cached: dict) -> Htp1State:
if 'slots' in cached:
for slot in cached['slots']:
if 'id' in slot:
if slot['id'] == 'HTP1':
if slot['last']:
loaded.slot.last = slot['last']
return loaded
@property
def device_type(self) -> str:
return self.__class__.__name__.lower()
def update(self, params: dict) -> bool:
any_update = False
if 'slots' in params:
for slot in params['slots']:
if slot['id'] == 'HTP1':
if 'entry' in slot:
if slot['entry']:
match = self.__catalogue.find(slot['entry'])
if match:
self.load_filter('HTP1', match)
any_update = True
else:
self.clear_filter('HTP1')
any_update = True
return any_update
def __send(self, to_load: List['PEQ']):
logger.info(f"Sending {len(to_load)} filters")
while len(to_load) < 16:
|
ops = [peq.as_ops(c, use_shelf=self.__supports_shelf) for peq in to_load for c in self.__peq.keys()]
ops = [op for slot_ops in ops for op in slot_ops if op]
if ops:
self.__client.send('changemso [{"op":"replace","path":"/peq/peqsw","value":true}]')
self.__client.send(f"changemso {json.dumps(ops)}")
else:
logger.warning(f"Nothing to send")
def activate(self, slot: str) -> None:
def __do_it():
self._current_state.slot.active = True
self._hydrate_cache_broadcast(__do_it)
def load_filter(self, slot: str, entry: CatalogueEntry) -> None:
to_load = [PEQ(idx, fc=f['freq'], q=f['q'], gain=f['gain'], filter_type_name=f['type'])
for idx, f in enumerate(entry.filters)]
self._hydrate_cache_broadcast(lambda: self.__do_it(to_load, entry.formatted_title))
def __do_it(self, to_load: List['PEQ'], title: str):
try:
self.__send(to_load)
self._current_state.slot.last = title
except Exception as e:
self._current_state.slot.last = 'ERRUR'
raise e
def clear_filter(self, slot: str) -> None:
self._hydrate_cache_broadcast(lambda: self.__do_it([], 'Empty'))
def mute(self, slot: Optional[str], channel: Optional[int]) -> None:
raise NotImplementedError()
def unmute(self, slot: Optional[str], channel: Optional[int]) -> None:
raise NotImplementedError()
def set_gain(self, slot: Optional[str], channel: Optional[int], gain: float) -> None:
raise NotImplementedError()
def on_mso(self, mso: dict):
logger.info(f"Received {mso}")
version = mso['versions']['swVer']
version = version[1:] if version[0] == 'v' or version[0] == 'V' else version
try:
self.__supports_shelf = semver.parse_version_info(version) > semver.parse_version_info('1.4.0')
except:
logger.error(f"Unable to parse version {mso['versions']['swVer']}, will not send shelf filters")
self.__supports_shelf = False
if not self.__supports_shelf:
logger.error(f"Device version {mso['versions']['swVer']} too old, lacks shelf filter support")
speakers = mso['speakers']['groups']
channels = ['lf', 'rf']
for group in [s for s, v in speakers.items() if 'present' in v and v['present'] is True]:
if group[0:2] == 'lr' and len(group) > 2:
channels.append('l' + group[2:])
channels.append('r' + group[2:])
else:
channels.append(group)
peq_slots = mso['peq']['slots']
filters = {c: [] for c in channels}
unknown_channels = set()
for idx, s in enumerate(peq_slots):
for c in channels:
if c in s['channels']:
filters[c].append(PEQ(idx, s['channels'][c]))
else:
unknown_channels.add(c)
if unknown_channels:
peq_channels = peq_slots[0]['channels'].keys()
logger.error(f"Unknown channels encountered [peq channels: {peq_channels}, unknown: {unknown_channels}]")
for c in filters.keys():
if c in self.__channels:
logger.info(f"Updating PEQ channel {c} with {filters[c]}")
self.__peq[c] = filters[c]
else:
logger.info(f"Discarding filter channel {c} - {filters[c]}")
def on_msoupdate(self, msoupdate: dict):
logger.info(f"Received {msoupdate}")
class Htp1Client:
def __init__(self, ip, listener):
self.__factory = Htp1ClientFactory(listener, f"ws://{ip}/ws/controller")
self.__connector = connectWS(self.__factory)
def send(self, msg: str):
self.__factory.broadcast(msg)
class Htp1Protocol(WebSocketClientProtocol):
def onConnecting(self, transport_details):
logger.info(f"Connecting to {transport_details}")
def onConnect(self, response):
logger.info(f"Connected to {response.peer}")
self.sendMessage('getmso'.encode('utf-8'), isBinary=False)
def onOpen(self):
logger.info("Connected to HTP1")
self.factory.register(self)
def onClose(self, was_clean, code, reason):
if was_clean:
logger.info(f"Disconnected code: {code} reason: {reason}")
else:
logger.warning(f"UNCLEAN! Disconnected code: {code} reason: {reason}")
def onMessage(self, payload, is_binary):
if is_binary:
logger.warning(f"Received {len(payload)} bytes in binary payload, ignoring")
else:
msg = payload.decode('utf8')
if msg.startswith('mso '):
logger.debug(f"Processing mso {msg}")
self.factory.listener.on_mso(json.loads(msg[4:]))
elif msg.startswith('msoupdate '):
logger.debug(f"Processing msoupdate {msg}")
self.factory.listener.on_msoupdate(json.loads(msg[10:]))
else:
logger.info(f"Received unknown payload {msg}")
class Htp1ClientFactory(WebSocketClientFactory, ReconnectingClientFactory):
protocol = Htp1Protocol
maxDelay = 5
initialDelay = 0.5
def __init__(self, listener, *args, **kwargs):
super(Htp1ClientFactory, self).__init__(*args, **kwargs)
self.__clients: List[Htp1Protocol] = []
self.listener = listener
self.setProtocolOptions(version=13)
def clientConnectionFailed(self, connector, reason):
logger.warning(f"Client connection failed {reason} .. retrying ..")
super().clientConnectionFailed(connector, reason)
def clientConnectionLost(self, connector, reason):
logger.warning(f"Client connection failed {reason} .. retrying ..")
super().clientConnectionLost(connector, reason)
def register(self, client: Htp1Protocol):
if client not in self.__clients:
logger.info(f"Registered device {client.peer}")
self.__clients.append(client)
else:
logger.info(f"Ignoring duplicate device {client.peer}")
def unregister(self, client: Htp1Protocol):
if client in self.__clients:
logger.info(f"Unregistering device {client.peer}")
self.__clients.remove(client)
else:
logger.info(f"Ignoring unregistered device {client.peer}")
def broadcast(self, msg):
if self.__clients:
disconnected_clients = []
for c in self.__clients:
logger.info(f"Sending to {c.peer} - {msg}")
try:
c.sendMessage(msg.encode('utf8'))
except Disconnected as e:
logger.exception(f"Failed to send to {c.peer}, discarding")
disconnected | peq = PEQ(len(to_load), fc=100, q=1, gain=0, filter_type_name='PeakingEQ')
to_load.append(peq) | conditional_block |
htp1.py | name': self.__name,
'slots': [self.slot.as_dict()]
}
class Htp1(PersistentDevice[Htp1State]):
def __init__(self, name: str, config_path: str, cfg: dict, ws_server: WsServer, catalogue: CatalogueProvider):
super().__init__(config_path, name, ws_server)
self.__name = name
self.__catalogue = catalogue
self.__ip = cfg['ip']
self.__channels = cfg['channels']
self.__peq = {}
self.__supports_shelf = True
if not self.__channels:
raise ValueError('No channels supplied for HTP-1')
self.__client = Htp1Client(self.__ip, self)
def _load_initial_state(self) -> Htp1State:
return Htp1State(self.name)
def _merge_state(self, loaded: Htp1State, cached: dict) -> Htp1State:
if 'slots' in cached:
for slot in cached['slots']:
if 'id' in slot:
if slot['id'] == 'HTP1':
if slot['last']:
loaded.slot.last = slot['last']
return loaded
@property
def device_type(self) -> str:
return self.__class__.__name__.lower()
def update(self, params: dict) -> bool:
any_update = False
if 'slots' in params:
for slot in params['slots']:
if slot['id'] == 'HTP1':
if 'entry' in slot:
if slot['entry']:
match = self.__catalogue.find(slot['entry'])
if match:
self.load_filter('HTP1', match)
any_update = True
else:
self.clear_filter('HTP1')
any_update = True
return any_update
def __send(self, to_load: List['PEQ']):
logger.info(f"Sending {len(to_load)} filters")
while len(to_load) < 16:
peq = PEQ(len(to_load), fc=100, q=1, gain=0, filter_type_name='PeakingEQ')
to_load.append(peq)
ops = [peq.as_ops(c, use_shelf=self.__supports_shelf) for peq in to_load for c in self.__peq.keys()]
ops = [op for slot_ops in ops for op in slot_ops if op]
if ops:
self.__client.send('changemso [{"op":"replace","path":"/peq/peqsw","value":true}]')
self.__client.send(f"changemso {json.dumps(ops)}")
else:
logger.warning(f"Nothing to send")
def activate(self, slot: str) -> None:
def __do_it():
self._current_state.slot.active = True
self._hydrate_cache_broadcast(__do_it)
def load_filter(self, slot: str, entry: CatalogueEntry) -> None:
to_load = [PEQ(idx, fc=f['freq'], q=f['q'], gain=f['gain'], filter_type_name=f['type'])
for idx, f in enumerate(entry.filters)]
self._hydrate_cache_broadcast(lambda: self.__do_it(to_load, entry.formatted_title))
def __do_it(self, to_load: List['PEQ'], title: str):
try:
self.__send(to_load)
self._current_state.slot.last = title
except Exception as e:
self._current_state.slot.last = 'ERRUR'
raise e
def clear_filter(self, slot: str) -> None:
self._hydrate_cache_broadcast(lambda: self.__do_it([], 'Empty'))
def mute(self, slot: Optional[str], channel: Optional[int]) -> None:
raise NotImplementedError()
def unmute(self, slot: Optional[str], channel: Optional[int]) -> None:
raise NotImplementedError()
def set_gain(self, slot: Optional[str], channel: Optional[int], gain: float) -> None:
raise NotImplementedError()
def on_mso(self, mso: dict):
logger.info(f"Received {mso}")
version = mso['versions']['swVer']
version = version[1:] if version[0] == 'v' or version[0] == 'V' else version
try:
self.__supports_shelf = semver.parse_version_info(version) > semver.parse_version_info('1.4.0')
except:
logger.error(f"Unable to parse version {mso['versions']['swVer']}, will not send shelf filters")
self.__supports_shelf = False
if not self.__supports_shelf:
logger.error(f"Device version {mso['versions']['swVer']} too old, lacks shelf filter support")
speakers = mso['speakers']['groups']
channels = ['lf', 'rf']
for group in [s for s, v in speakers.items() if 'present' in v and v['present'] is True]:
if group[0:2] == 'lr' and len(group) > 2:
channels.append('l' + group[2:])
channels.append('r' + group[2:])
else:
channels.append(group)
peq_slots = mso['peq']['slots']
filters = {c: [] for c in channels}
unknown_channels = set()
for idx, s in enumerate(peq_slots):
for c in channels:
if c in s['channels']:
filters[c].append(PEQ(idx, s['channels'][c]))
else:
unknown_channels.add(c)
if unknown_channels:
peq_channels = peq_slots[0]['channels'].keys()
logger.error(f"Unknown channels encountered [peq channels: {peq_channels}, unknown: {unknown_channels}]")
for c in filters.keys():
if c in self.__channels:
logger.info(f"Updating PEQ channel {c} with {filters[c]}")
self.__peq[c] = filters[c]
else:
logger.info(f"Discarding filter channel {c} - {filters[c]}")
def on_msoupdate(self, msoupdate: dict):
logger.info(f"Received {msoupdate}")
class Htp1Client:
def __init__(self, ip, listener):
self.__factory = Htp1ClientFactory(listener, f"ws://{ip}/ws/controller")
self.__connector = connectWS(self.__factory)
def send(self, msg: str):
self.__factory.broadcast(msg)
class Htp1Protocol(WebSocketClientProtocol):
def onConnecting(self, transport_details):
logger.info(f"Connecting to {transport_details}")
def onConnect(self, response):
logger.info(f"Connected to {response.peer}")
self.sendMessage('getmso'.encode('utf-8'), isBinary=False)
def onOpen(self):
logger.info("Connected to HTP1")
self.factory.register(self)
def onClose(self, was_clean, code, reason):
if was_clean:
logger.info(f"Disconnected code: {code} reason: {reason}")
else:
logger.warning(f"UNCLEAN! Disconnected code: {code} reason: {reason}")
def onMessage(self, payload, is_binary):
if is_binary:
logger.warning(f"Received {len(payload)} bytes in binary payload, ignoring")
else:
msg = payload.decode('utf8')
if msg.startswith('mso '):
logger.debug(f"Processing mso {msg}")
self.factory.listener.on_mso(json.loads(msg[4:]))
elif msg.startswith('msoupdate '):
logger.debug(f"Processing msoupdate {msg}")
self.factory.listener.on_msoupdate(json.loads(msg[10:]))
else:
logger.info(f"Received unknown payload {msg}")
class Htp1ClientFactory(WebSocketClientFactory, ReconnectingClientFactory):
protocol = Htp1Protocol
maxDelay = 5
initialDelay = 0.5
def __init__(self, listener, *args, **kwargs):
|
def clientConnectionFailed(self, connector, reason):
logger.warning(f"Client connection failed {reason} .. retrying ..")
super().clientConnectionFailed(connector, reason)
def clientConnectionLost(self, connector, reason):
logger.warning(f"Client connection failed {reason} .. retrying ..")
super().clientConnectionLost(connector, reason)
def register(self, client: Htp1Protocol):
if client not in self.__clients:
logger.info(f"Registered device {client.peer}")
self.__clients.append(client)
else:
logger.info(f"Ignoring duplicate device {client.peer}")
def unregister(self, client: Htp1Protocol):
if client in self.__clients:
logger.info(f"Unregistering device {client.peer}")
self.__clients.remove(client)
else:
logger.info(f"Ignoring unregistered device {client.peer}")
def broadcast(self, msg):
if self.__clients:
disconnected_clients = []
for c in self.__clients:
logger.info(f"Sending to {c.peer} - {msg}")
try:
c.sendMessage(msg.encode('utf8'))
except Disconnected as e:
logger.exception(f"Failed to send to {c.peer}, discarding")
| super(Htp1ClientFactory, self).__init__(*args, **kwargs)
self.__clients: List[Htp1Protocol] = []
self.listener = listener
self.setProtocolOptions(version=13) | identifier_body |
htp1.py | name': self.__name,
'slots': [self.slot.as_dict()]
}
class Htp1(PersistentDevice[Htp1State]):
def __init__(self, name: str, config_path: str, cfg: dict, ws_server: WsServer, catalogue: CatalogueProvider):
super().__init__(config_path, name, ws_server)
self.__name = name
self.__catalogue = catalogue
self.__ip = cfg['ip']
self.__channels = cfg['channels']
self.__peq = {}
self.__supports_shelf = True
if not self.__channels:
raise ValueError('No channels supplied for HTP-1')
self.__client = Htp1Client(self.__ip, self)
def _load_initial_state(self) -> Htp1State:
return Htp1State(self.name)
def _merge_state(self, loaded: Htp1State, cached: dict) -> Htp1State:
if 'slots' in cached:
for slot in cached['slots']:
if 'id' in slot:
if slot['id'] == 'HTP1':
if slot['last']:
loaded.slot.last = slot['last']
return loaded
@property
def device_type(self) -> str:
return self.__class__.__name__.lower()
def update(self, params: dict) -> bool:
any_update = False
if 'slots' in params:
for slot in params['slots']:
if slot['id'] == 'HTP1':
if 'entry' in slot:
if slot['entry']:
match = self.__catalogue.find(slot['entry'])
if match:
self.load_filter('HTP1', match)
any_update = True
else:
self.clear_filter('HTP1')
any_update = True
return any_update
def __send(self, to_load: List['PEQ']):
logger.info(f"Sending {len(to_load)} filters")
while len(to_load) < 16:
peq = PEQ(len(to_load), fc=100, q=1, gain=0, filter_type_name='PeakingEQ')
to_load.append(peq)
ops = [peq.as_ops(c, use_shelf=self.__supports_shelf) for peq in to_load for c in self.__peq.keys()]
ops = [op for slot_ops in ops for op in slot_ops if op]
if ops:
self.__client.send('changemso [{"op":"replace","path":"/peq/peqsw","value":true}]')
self.__client.send(f"changemso {json.dumps(ops)}")
else:
logger.warning(f"Nothing to send")
def activate(self, slot: str) -> None:
def __do_it():
self._current_state.slot.active = True
self._hydrate_cache_broadcast(__do_it)
def load_filter(self, slot: str, entry: CatalogueEntry) -> None:
to_load = [PEQ(idx, fc=f['freq'], q=f['q'], gain=f['gain'], filter_type_name=f['type'])
for idx, f in enumerate(entry.filters)]
self._hydrate_cache_broadcast(lambda: self.__do_it(to_load, entry.formatted_title))
def __do_it(self, to_load: List['PEQ'], title: str):
try:
self.__send(to_load)
self._current_state.slot.last = title
except Exception as e:
self._current_state.slot.last = 'ERRUR'
raise e
def clear_filter(self, slot: str) -> None:
self._hydrate_cache_broadcast(lambda: self.__do_it([], 'Empty'))
def mute(self, slot: Optional[str], channel: Optional[int]) -> None:
raise NotImplementedError()
def unmute(self, slot: Optional[str], channel: Optional[int]) -> None:
raise NotImplementedError()
def set_gain(self, slot: Optional[str], channel: Optional[int], gain: float) -> None:
raise NotImplementedError()
def on_mso(self, mso: dict):
logger.info(f"Received {mso}")
version = mso['versions']['swVer']
version = version[1:] if version[0] == 'v' or version[0] == 'V' else version
try:
self.__supports_shelf = semver.parse_version_info(version) > semver.parse_version_info('1.4.0')
except:
logger.error(f"Unable to parse version {mso['versions']['swVer']}, will not send shelf filters")
self.__supports_shelf = False
if not self.__supports_shelf:
logger.error(f"Device version {mso['versions']['swVer']} too old, lacks shelf filter support")
speakers = mso['speakers']['groups']
channels = ['lf', 'rf']
for group in [s for s, v in speakers.items() if 'present' in v and v['present'] is True]:
if group[0:2] == 'lr' and len(group) > 2:
channels.append('l' + group[2:])
channels.append('r' + group[2:])
else:
channels.append(group)
peq_slots = mso['peq']['slots']
filters = {c: [] for c in channels}
unknown_channels = set()
for idx, s in enumerate(peq_slots):
for c in channels:
if c in s['channels']:
filters[c].append(PEQ(idx, s['channels'][c]))
else:
unknown_channels.add(c)
if unknown_channels:
peq_channels = peq_slots[0]['channels'].keys()
logger.error(f"Unknown channels encountered [peq channels: {peq_channels}, unknown: {unknown_channels}]")
for c in filters.keys():
if c in self.__channels:
logger.info(f"Updating PEQ channel {c} with {filters[c]}")
self.__peq[c] = filters[c]
else:
logger.info(f"Discarding filter channel {c} - {filters[c]}")
def on_msoupdate(self, msoupdate: dict):
logger.info(f"Received {msoupdate}")
class Htp1Client:
def __init__(self, ip, listener):
self.__factory = Htp1ClientFactory(listener, f"ws://{ip}/ws/controller")
self.__connector = connectWS(self.__factory)
def send(self, msg: str):
self.__factory.broadcast(msg)
class Htp1Protocol(WebSocketClientProtocol):
def onConnecting(self, transport_details):
logger.info(f"Connecting to {transport_details}")
def onConnect(self, response):
logger.info(f"Connected to {response.peer}")
self.sendMessage('getmso'.encode('utf-8'), isBinary=False)
def onOpen(self):
logger.info("Connected to HTP1")
self.factory.register(self)
def onClose(self, was_clean, code, reason):
if was_clean:
logger.info(f"Disconnected code: {code} reason: {reason}")
else:
logger.warning(f"UNCLEAN! Disconnected code: {code} reason: {reason}")
def onMessage(self, payload, is_binary):
if is_binary:
logger.warning(f"Received {len(payload)} bytes in binary payload, ignoring")
else:
msg = payload.decode('utf8')
if msg.startswith('mso '):
logger.debug(f"Processing mso {msg}")
self.factory.listener.on_mso(json.loads(msg[4:]))
elif msg.startswith('msoupdate '):
logger.debug(f"Processing msoupdate {msg}")
self.factory.listener.on_msoupdate(json.loads(msg[10:]))
else:
logger.info(f"Received unknown payload {msg}")
class Htp1ClientFactory(WebSocketClientFactory, ReconnectingClientFactory):
protocol = Htp1Protocol
maxDelay = 5
initialDelay = 0.5
def __init__(self, listener, *args, **kwargs):
super(Htp1ClientFactory, self).__init__(*args, **kwargs)
self.__clients: List[Htp1Protocol] = []
self.listener = listener
self.setProtocolOptions(version=13)
def clientConnectionFailed(self, connector, reason):
logger.warning(f"Client connection failed {reason} .. retrying ..")
super().clientConnectionFailed(connector, reason)
def clientConnectionLost(self, connector, reason):
logger.warning(f"Client connection failed {reason} .. retrying ..")
super().clientConnectionLost(connector, reason)
def register(self, client: Htp1Protocol):
if client not in self.__clients:
logger.info(f"Registered device {client.peer}")
self.__clients.append(client)
else:
logger.info(f"Ignoring duplicate device {client.peer}")
def | (self, client: Htp1Protocol):
if client in self.__clients:
logger.info(f"Unregistering device {client.peer}")
self.__clients.remove(client)
else:
logger.info(f"Ignoring unregistered device {client.peer}")
def broadcast(self, msg):
if self.__clients:
disconnected_clients = []
for c in self.__clients:
logger.info(f"Sending to {c.peer} - {msg}")
try:
c.sendMessage(msg.encode('utf8'))
except Disconnected as e:
logger.exception(f"Failed to send to {c.peer}, discarding")
| unregister | identifier_name |
Final of the Electromagnet Project.py | tt = float(raw_input("Enter the thickness of the load in mm :"))
llcm = ll * float(2.54)
wwcm = ww * float(2.54)
ttcm = tt / float(10)
VV = float(llcm * wwcm * ttcm)
density = float(7.81)
mgg = VV * density
mg = mgg / float(1000)
elif material == 2:
ll = float(raw_input("Enter the length of the load in inches :"))
ww = float(raw_input("Enter the width of the load in inches :"))
tt = float(raw_input("Enter the thickness of the load in mm :"))
llcm = ll * float(2.54)
wwcm = ww * float(2.54)
ttcm = tt / float(10)
VV = float(llcm * wwcm * ttcm)
density = float(7.83)
mgg = VV * density
mg = mgg / float(1000)
elif material == 3:
ll = float(raw_input("Enter the length of the load in inches :"))
ww = float(raw_input("Enter the width of the load in inches :"))
tt = float(raw_input("Enter the thickness of the load in mm :"))
llcm = ll * float(2.54)
wwcm = ww * float(2.54)
ttcm = tt / float(10)
VV = float(llcm * wwcm * ttcm)
density = float(2.7)
mgg = VV * density
mg = mgg / float(1000)
elif material == 4:
ll = float(raw_input("Enter the length of the load in inches :"))
ww = float(raw_input("Enter the width of the load in inches :"))
tt = float(raw_input("Enter the thickness of the load in mm :"))
llcm = ll * float(2.54)
wwcm = ww * float(2.54)
ttcm = tt / float(10)
VV = float(llcm * wwcm * ttcm)
density = float(4.51)
mgg = VV * density
mg = mgg / float(1000)
elif material == 5:
ll = float(raw_input("Enter the length of the load in inches :"))
ww = float(raw_input("Enter the width of the load in inches :"))
tt = float(raw_input("Enter the thickness of the load in mm :"))
llcm = ll * float(2.54)
wwcm = ww * float(2.54)
ttcm = tt / float(10)
VV = float(llcm * wwcm * ttcm)
density = float(8.70)
mgg = VV * density
mg = mgg / float(1000)
elif material == 10:
ll = float(raw_input("Enter the length of the load in inches :"))
ww = float(raw_input("Enter the width of the load in inches :"))
tt = float(raw_input("Enter the thickness of the load in mm :"))
llcm = ll * float(2.54)
wwcm = ww * float(2.54)
ttcm = tt / float(10)
VV = float(llcm * wwcm * ttcm)
density = float(raw_input("Enter the density of the material in g/cm^3 :"))
mgg = VV * density
mg = mgg / float(1000)
else:
print " Wrong Entered Number, Please Enter again"
print "The weight of the load is %f in kg" %mg
print "Fmg = The gravitational force produced by load "
Fmg = float(9.8 * mg)
print "Fmg is %f in N" %(Fmg)
# In this step, calculation of magnetic force is covered. Magnetic flux density is limited around 1.5 T due to saturation. Cross sectional area of the core is calculated based on the weight of the load. And, to get better area parameters, there is an optional re-adjusted area. Our design is based on circular area because of the stacking factor improvement and lower losses of the core.
#
# The force exerted by an electromagnet on a section of core material is:
# $$F = B^2A/2\mu_0$$
#
# In[4]:
B = float(raw_input ("Enter an appropriate B value in T : ")) # Nominal 0.6 - 1.5 T
A = Fmg * 2* uo / (B * B)
print "Cross Sectional Area (min) in m^2 : %f" %A
A = float(raw_input ("Enter Cross Sectional Area (A) in m^2 : ")) # TO GET BETTER APPROXIMATION OF AREA
dia_core = sqrt( A / pi )
dia_core = dia_core * float(1000)
print "For circular design ---> diameter of core : %f mm" %dia_core # CIRCULAR DESIGN
F = B * B * A / (2 * uo)
print "Force is %f in N and Fmg is %f in N" %(F,Fmg)
while True:
if F < Fmg:
print ("Not sufficient MMF, Enter new design values")
A = float(raw_input("Enter Cross Sectional Area (A) in m^2 : "))
F = B * B * A / (2 * uo)
else:
break
# Throughout this part,
# Firstly, lifting distance is requested and pre_design core length (roughly) is requested. For unsufficient core length, code gives an error at the end. According to flux path in both core and air gap, MMF is calculated. After entering number of turns for given MMF value, current on the wire is obtained.
# ### Selection of the material
# Solid steel is generally best, in terms of economics, for the yoke, or frame, of static field devices. The mass of material required to efficiently carry flux between the far poles of an assembly make anything else impossible to justify. The reluctance of this part of the magnetic circuit is relatively low, even without annealing, compared to that of the working gap, so the losses associated with driving flux through the material are a small portion of overall losses
#
# ### Losses of Electromagnet by choosing current
# The only power consumed in a DC electromagnet is due to the resistance of the windings, and is dissipated as heat.Since the MMF is proportional to the product NI,the number of turns in the windings N and the current I can be chosen to minimize heat losses, as long as their product is constant. Since the power dissipation,
#
# $$P = I^2R$$
#
# ,increases with the square of the current but only increases approximately linearly with the number of windings, the power lost in the windings can be minimized by reducing I and increasing the number of turns N proportionally, or using thicker wire to reduce the resistance; However, the limit to increasing N or lowering the resistance is that the windings take up more room between the magnet’s core pieces. If the area available for the windings is filled up, more turns require going to a smaller diameter of wire, which has higher resistance, which cancels the advantage of using more turns.
#
# ## Electromagnet Design Procedure (for U shaped)
# U shaped electromagnet is shown below. It has two airgap and two wire turn area. The core is in a circular form.
#
# 
#
# Applying Ampere's Law to the circuit,
#
# $$\sum_{}^\ i = 2Ni = \int_{S}^{} H dl = H_{s}l_{s} + 2H_{g}l_{g}$$
#
# where Hs is the value of H in the core path, which is assumed to be constant and ls is the total length of the core path. At the boundary the magnetic flux densities are equal.
# $$ B_{s} = B_{g}$$
#
# Actually,
#
# $$H_{s}l_{s} << 2H_{g}l_{g}$$
#
# However, nothing is neglected to get accurate result. Therefore,
#
# $$F = 2Ni = \phi(R_{s} + 2R{g})$$
#
# where:
# $$R_{s} = l_{s}/\mu\mu_0A_{s}$$
# $$R_{g} = l_{g}/\mu_0A_{g}$$
#
# In[5]:
# CALCULATION OF CURRENT
L_gap = float(raw_input("Enter the lifting distance in mm :")) # FOR CONTACTED LIFTING => L_gap = 0. | material = float(raw_input("The load material: "))
if material == 1:
ll = float(raw_input("Enter the length of the load in inches :"))
ww = float(raw_input("Enter the width of the load in inches :")) | random_line_split |
|
Final of the Electromagnet Project.py | _Sum = Sum_{lengthofwire} * T $$
#
# The total resistance of the wire is calculated by using the AWG table.
# $$ R_{total} = Total_Sum * R_{ohm / m}$$
#
# The inductance of the coil is obtained as:
# $$ L = N \phi / I$$
# where:
# - L is the total inductance of the coil
# - I is the current on the wire
#
# The total weight of the wire is calculated by using the diameter and density of the selected material. Material is mostly selected copper, which has the density 8.96 g/cm^3.
# $$ Volume_{wire} = \pi * (d/2)^2 * Total_sum$$
# $$ Weight_{wire} = Volume_{wire} * density_{material}$$
#
# The required voltage is obtained as:
# $$ V_{required} = I * R_{total}$$
# In[7]:
stacking = float(raw_input ("Enter the stacking factor : %f")) #roughly 0.9
# dia_meter is obtained
# dia_wire is obtained
# turn number is obtained
print "The entered length of the core is %f m" %L_core
L_former = float(raw_input("Enter the Length of Former in cm :"))
L_former = float (L_former * 10)
L_coremm = L_core * 1000 # Turn Lcore from m to mm
if L_former > L_coremm:
print " L_former is longer than L_core, Please re-enter the L_core values"
print "L_core value is %f" %L_core
L_former = float(raw_input("Enter the Length of Former in cm :"))
L_former = float (L_former * 10)
first = int (L_former * stacking / diameter_wire)
print " The maximum number of winding in the first layer is %f" %first
n = float(N) / first
n = int(n) + 1
print "the total number of layers required to give the total number of turns : %f" %n
Sum_length_wire = (float(n) / 2) * float(2 * pi * diameter_wire + (n-1) * diameter_wire)
Sum_length_wire = Sum_length_wire / float(10) # in cm
if N < first :
Sum_length_wire = Sum_length_wire * float (N)
else:
Sum_length_wire = Sum_length_wire * float (first)
print "Total Length of wire is %f (cm)" %Sum_length_wire
Sum_length_wire = Sum_length_wire / float(100) #in m
R_total = resistance_wire * Sum_length_wire
print " The resistance of the wire is %f ohm " %R_total # TOO LOW VALUE OF RESISTANCES !!!!!!!!!
flux = float( B * A)
Inductance = float(N) * flux / I
print " The inductance of the coil is %f H" %Inductance
dia_wiree = diameter_wire / float(10)
dia_wire = dia_wiree / float(2)
density_copper = 8.96 # g / cm^-3
weight = pi* dia_wire * dia_wire * Sum_length_wire * density_copper
print " The weight of the wire is %f g" %weight
V_required = I * R_total
print " Required voltage for the electromagnet is %f V" %V_required
# # U-shaped Electromagnet Design Prodecure !!More Accurate!!
# At this part is an optional area if the user is trying to design U-shaped electromagnet, far more accurately. It needs more entries on core.
# U-shaped electromagnet is shown below.
# 
#
# The user is used the obtained results and parameters, which are obtained above, to check more accurately.
# All reluctances are included into the design below.
# The magnetic circuit of the U-shaped electromagnet is shown below.
#
# 
#
# Due to U-shape, there are two airgap.The force of attraction across each air gap between the electromagnet and total force is expressed, respectively.
# $$M_{airgap} = NI/2$$
# $$R_{airgap} = g / \mu_0 A_{pole}$$
#
# $$F_{magnet} = \mu_0 (NI)^2 A_{pole} / 4g^2$$
#
# where,
# - M is the magneto-motive force across the each air gap
# - g is the length of the airgap
# - R_airgap is the reluctance of the each airgap
# Fmagnet equation shows that the magnetic force is a nonlinear function of both current and air gap length. And also, for a constant current the force decreases with increasing air gap, hence it has a negative stiffness coefficient. There is therefore no point of equilibrium between two magnetised bodies and so the open-loop force-air gap characteristic of an electromagnet is unstable.
# The assumption of uniform air gap flux distribution is valid only for air gaps that are much smaller than the pole width. At larger air gaps, flux fringing increases the effective air gap flux area and hence decreases the air gap reluctance.
#
# Directly lifting force is expressed below. For the loads in Parker Steel Company, the lateral offset (y) is zero.
# $$F_{lift} = F_{magnet}(1+(2g/\pi p)(1-(y/g)tan^-1(y/g)))$$
# where,
# - p is the pole width
# - y is the lateral offset between the electromagnet and track poles. It is set to zero.
#
# By considering zero lateral offset,
# $$R_{airgap} = g / (\mu_0 l (p + 2g/\pi))$$
#
# Magnet and track path reluctance is calculated by summing the exact dimension entries.
# $$R_{magnet} = (2h + w + 2p) / \mu_M \mu_0 lp$$
# $$R_{track} = (t + 4p)/ \mu_T \mu_0 lp$$
# where,
# - w is the width between the electromagnet poles
# - p is the pole width
# - h is the pole height above the yoke
# - l is the length of the electromagnet
# - t is the width between the track poles
# - μM is the relavite permeability of electromagnet
# - μT is the relavite permeability of track
#
# Leakage flux is calculated by neglecting fringe flux. Because air gap is so small for fringe flux calculation.
# $$R_{leakage} = w / \mu_0 l (h/2) $$
#
# All the parameters are obtained in the U-shaped magnetic circuit, shown above. After the analyzing the circuit,
# $$ M_{airgap} = (R_{airgap}R_{leakage}/ (R_{track}+2R_{airgap})(R_{leakage}+R_{magnet}+R_{magnet}R_{leakage})) M_{coils}$$
# where,
# $$M_{coils} = NI$$
#
# By substituting M_airgap into the function;
# $$F_{magnet} = M_{airgap}^2/g R_{airgap} $$
#
# $$F_{magnet} = (R_{airgap}R_{leakage}^2/g((R_{track}+2R_{airgap})(R_{leakage}+R_{magnet}+R_{magnet}R_{leakage}))^2)M_{coils}^2$$
#
# F_lift is calculated by substituting F_magnet into F_lift function.
#
# In[ ]:
print "Welcome to the Electromagnet Design Code"
print " U-SHAPE Electromagnet is covered in detailed "
# In[ ]:
while True:
l = f | loat(raw_input("Enter the length of Electromagnet in mm "))
w = float(raw_input("Enter the Width between the electromagnet pole pieces in mm "))
h = float(raw_input("Enter the Height of the pole pieces above the yoke in mm "))
p = float(raw_input("Enter the Width of the pole pieces in mm "))
if l >= (2 * h + w + 2 * p):
print "The length of Electromagnet ERROR"
print "l should be equal to (2 * h + w + 2 * p)"
l = float(raw_input("Enter the length of Electromagnet in mm "))
w = float(raw_input("Enter the Width between the electromagnet pole pieces in mm "))
h = float(raw_input("Enter the Height of the pole pieces above the yoke in mm "))
p = float(raw_input("Enter the Width of the pole pieces in mm "))
else:
break
break
| conditional_block |
|
aa_changes.rs | ,
pos: AaRefPosition,
qry_seq: &[Nuc],
ref_seq: &[Nuc],
ref_tr: &CdsTranslation,
qry_tr: &CdsTranslation,
) -> Self {
let ref_aa = ref_tr.seq[pos.as_usize()];
let qry_aa = qry_tr.seq[pos.as_usize()];
let nuc_ranges = cds_codon_pos_to_ref_range(cds, pos);
let ref_triplet = nuc_ranges
.iter()
.flat_map(|(range, strand)| {
let mut nucs = ref_seq[range.to_std()].to_vec();
if strand == &GeneStrand::Reverse {
reverse_complement_in_place(&mut nucs);
}
nucs
})
.collect_vec();
let qry_triplet = nuc_ranges
.iter()
.flat_map(|(range, strand)| {
let mut nucs = qry_seq[range.clamp_range(0, qry_seq.len()).to_std()].to_vec();
if strand == &GeneStrand::Reverse {
reverse_complement_in_place(&mut nucs);
}
nucs
})
.collect_vec();
let nuc_ranges = nuc_ranges.into_iter().map(|(range, _)| range).collect_vec();
Self {
cds_name: cds.name.clone(),
pos,
ref_aa,
qry_aa,
nuc_pos: nuc_ranges[0].begin,
nuc_ranges,
ref_triplet,
qry_triplet,
}
}
#[inline]
pub fn is_mutated_or_deleted(&self) -> bool {
is_aa_mutated_or_deleted(self.ref_aa, self.qry_aa)
}
}
#[derive(Clone, Debug, Default, Serialize, Deserialize, schemars::JsonSchema)]
#[serde(rename_all = "camelCase")]
pub struct AaChangesGroup {
name: String,
range: AaRefRange,
changes: Vec<AaChangeWithContext>,
nuc_subs: Vec<NucSub>,
nuc_dels: Vec<NucDelRange>,
}
impl AaChangesGroup {
pub fn new(name: impl AsRef<str>) -> Self {
Self::with_changes(name, vec![])
}
pub fn with_changes(name: impl AsRef<str>, changes: Vec<AaChangeWithContext>) -> Self {
Self {
name: name.as_ref().to_owned(),
range: Self::find_codon_range(&changes),
changes,
nuc_subs: vec![],
nuc_dels: vec![],
}
}
pub fn push(&mut self, change: AaChangeWithContext) {
self.changes.push(change);
self.range = Self::find_codon_range(&self.changes);
}
pub fn last(&self) -> Option<&AaChangeWithContext> {
self.changes.last()
}
fn find_codon_range(changes: &[AaChangeWithContext]) -> AaRefRange {
match changes.iter().minmax_by_key(|change| change.pos) {
MinMaxResult::NoElements => AaRefRange::from_isize(0, 0),
MinMaxResult::OneElement(one) => AaRefRange::new(one.pos, one.pos + 1),
MinMaxResult::MinMax(first, last) => AaRefRange::new(first.pos, last.pos + 1),
}
}
}
#[derive(Clone, Debug, Default, Serialize, Deserialize, schemars::JsonSchema)]
#[serde(rename_all = "camelCase")]
pub struct FindAaChangesOutput {
pub aa_changes_groups: Vec<AaChangesGroup>,
pub aa_substitutions: Vec<AaSub>,
pub aa_deletions: Vec<AaDel>,
pub nuc_to_aa_muts: BTreeMap<String, Vec<AaSub>>,
}
/// Finds aminoacid substitutions and deletions in query peptides relative to reference peptides, in all genes
///
/// ## Precondition
/// Nucleotide sequences and peptides are required to be stripped from insertions
pub fn find_aa_changes(
ref_seq: &[Nuc],
qry_seq: &[Nuc],
ref_translation: &Translation,
qry_translation: &Translation,
gene_map: &GeneMap,
nuc_subs: &[NucSub],
nuc_dels: &[NucDelRange],
) -> Result<FindAaChangesOutput, Report> |
changes.aa_substitutions.sort();
changes.aa_deletions.sort();
changes.nuc_to_aa_muts.iter_mut().for_each(|(_, vals)| {
vals.sort();
vals.dedup();
});
Ok(changes)
}
/// Finds aminoacid substitutions and deletions in query peptides relative to reference peptides, in one gene
///
/// ## Precondition
/// Nucleotide sequences and peptides are required to be stripped from insertions
///
///
/// ## Implementation details
/// We compare reference and query peptides (extracted by the preceding call to Nextalign),
/// one aminoacid at at time, and deduce changes. We then report the change and relevant nucleotide context surrounding
/// this change.
/// Previously we reported one-to-one mapping of aminoacid changes to corresponding nucleotide changes. However, it
/// was not always accurate, because if there are multiple nucleotide changes in a codon, the direct correspondence
/// might not always be established without knowing the order in which nucleotide changes have occurred. And in the
/// context of Nextclade we don't have this information.
fn find_aa_changes_for_cds(
cds: &Cds,
qry_seq: &[Nuc],
ref_seq: &[Nuc],
ref_tr: &CdsTranslation,
qry_tr: &CdsTranslation,
nuc_subs: &[NucSub],
nuc_dels: &[NucDelRange],
) -> FindAaChangesOutput {
assert_eq!(ref_tr.seq.len(), qry_tr.seq.len());
assert_eq!(qry_seq.len(), ref_seq.len());
let aa_alignment_ranges = &qry_tr.alignment_ranges;
let mut aa_changes_groups = vec![AaChangesGroup::new(&cds.name)];
let mut curr_group = aa_changes_groups.last_mut().unwrap();
for codon in AaRefRange::from_usize(0, qry_tr.seq.len()).iter() {
if !is_codon_sequenced(aa_alignment_ranges, codon) {
continue;
}
let ref_aa = ref_tr.seq[codon.as_usize()];
let qry_aa = qry_tr.seq[codon.as_usize()];
if is_aa_mutated_or_deleted(ref_aa, qry_aa) {
match curr_group.last() {
// If current group is empty, then we are about to insert the first codon into the first group.
None => {
if codon > 0 && is_codon_sequenced(aa_alignment_ranges, codon - 1) {
// Also prepend one codon to the left, for additional context, to start the group
curr_group.push(AaChangeWithContext::new(
cds,
codon - 1,
qry_seq,
ref_seq,
ref_tr,
qry_tr,
));
}
// The current codon itself
curr_group.push(AaChangeWithContext::new(cds, codon, qry_seq, ref_seq, ref_tr, qry_tr));
}
// Current group is not empty
Some(prev) => {
// If previous codon in the group is adjacent or almost adjacent (there is 1 item in between),
// then append to the group.
if codon <= prev.pos + 2 {
// If previous codon in the group is not exactly adjacent, there is 1 item in between,
// then cover the hole by inserting previous codon.
if codon == prev.pos + 2 && is_codon_sequenced(aa_alignment_ranges, codon - 1) {
curr_group.push(AaChangeWithContext::new(
cds,
codon - 1,
qry_seq,
ref_seq,
ref_tr,
qry_tr,
));
}
// And insert the current codon
curr_group.push(AaChangeWithContext::new(cds, codon, qry_seq, ref_seq, ref_tr, qry_tr));
}
// If previous codon in the group is not adjacent, then terminate the current group and start a new group.
else {
// Add one codon to the right, for additional context, to finalize the current group
if is_codon_sequ | {
let mut changes = qry_translation
.iter_cdses()
.map(|(qry_name, qry_cds_tr)| {
let ref_cds_tr = ref_translation.get_cds(qry_name)?;
let cds = gene_map.get_cds(&qry_cds_tr.name)?;
Ok(find_aa_changes_for_cds(
cds, qry_seq, ref_seq, ref_cds_tr, qry_cds_tr, nuc_subs, nuc_dels,
))
})
.collect::<Result<Vec<FindAaChangesOutput>, Report>>()?
.into_iter()
// Merge changes from all CDSes into one struct
.fold(FindAaChangesOutput::default(), |mut output, changes| {
output.aa_changes_groups.extend(changes.aa_changes_groups);
output.aa_substitutions.extend(changes.aa_substitutions);
output.aa_deletions.extend(changes.aa_deletions);
extend_map_of_vecs(&mut output.nuc_to_aa_muts, changes.nuc_to_aa_muts);
output
}); | identifier_body |
aa_changes.rs | ,
pos: AaRefPosition,
qry_seq: &[Nuc],
ref_seq: &[Nuc],
ref_tr: &CdsTranslation,
qry_tr: &CdsTranslation,
) -> Self {
let ref_aa = ref_tr.seq[pos.as_usize()];
let qry_aa = qry_tr.seq[pos.as_usize()];
let nuc_ranges = cds_codon_pos_to_ref_range(cds, pos);
let ref_triplet = nuc_ranges
.iter()
.flat_map(|(range, strand)| {
let mut nucs = ref_seq[range.to_std()].to_vec();
if strand == &GeneStrand::Reverse {
reverse_complement_in_place(&mut nucs);
}
nucs
})
.collect_vec();
let qry_triplet = nuc_ranges
.iter()
.flat_map(|(range, strand)| {
let mut nucs = qry_seq[range.clamp_range(0, qry_seq.len()).to_std()].to_vec();
if strand == &GeneStrand::Reverse {
reverse_complement_in_place(&mut nucs);
}
nucs
})
.collect_vec();
let nuc_ranges = nuc_ranges.into_iter().map(|(range, _)| range).collect_vec();
Self {
cds_name: cds.name.clone(),
pos,
ref_aa,
qry_aa,
nuc_pos: nuc_ranges[0].begin,
nuc_ranges,
ref_triplet,
qry_triplet,
}
}
#[inline]
pub fn is_mutated_or_deleted(&self) -> bool {
is_aa_mutated_or_deleted(self.ref_aa, self.qry_aa)
}
}
#[derive(Clone, Debug, Default, Serialize, Deserialize, schemars::JsonSchema)]
#[serde(rename_all = "camelCase")]
pub struct AaChangesGroup {
name: String,
range: AaRefRange,
changes: Vec<AaChangeWithContext>,
nuc_subs: Vec<NucSub>,
nuc_dels: Vec<NucDelRange>,
}
impl AaChangesGroup {
pub fn new(name: impl AsRef<str>) -> Self {
Self::with_changes(name, vec![])
}
pub fn with_changes(name: impl AsRef<str>, changes: Vec<AaChangeWithContext>) -> Self {
Self {
name: name.as_ref().to_owned(),
range: Self::find_codon_range(&changes),
changes,
nuc_subs: vec![],
nuc_dels: vec![],
}
}
pub fn push(&mut self, change: AaChangeWithContext) {
self.changes.push(change);
self.range = Self::find_codon_range(&self.changes);
}
pub fn last(&self) -> Option<&AaChangeWithContext> {
self.changes.last()
}
fn find_codon_range(changes: &[AaChangeWithContext]) -> AaRefRange {
match changes.iter().minmax_by_key(|change| change.pos) {
MinMaxResult::NoElements => AaRefRange::from_isize(0, 0),
MinMaxResult::OneElement(one) => AaRefRange::new(one.pos, one.pos + 1),
MinMaxResult::MinMax(first, last) => AaRefRange::new(first.pos, last.pos + 1),
}
}
}
#[derive(Clone, Debug, Default, Serialize, Deserialize, schemars::JsonSchema)]
#[serde(rename_all = "camelCase")]
pub struct FindAaChangesOutput {
pub aa_changes_groups: Vec<AaChangesGroup>,
pub aa_substitutions: Vec<AaSub>,
pub aa_deletions: Vec<AaDel>,
pub nuc_to_aa_muts: BTreeMap<String, Vec<AaSub>>,
}
/// Finds aminoacid substitutions and deletions in query peptides relative to reference peptides, in all genes
///
/// ## Precondition
/// Nucleotide sequences and peptides are required to be stripped from insertions
pub fn find_aa_changes(
ref_seq: &[Nuc],
qry_seq: &[Nuc],
ref_translation: &Translation,
qry_translation: &Translation,
gene_map: &GeneMap,
nuc_subs: &[NucSub],
nuc_dels: &[NucDelRange],
) -> Result<FindAaChangesOutput, Report> {
let mut changes = qry_translation
.iter_cdses()
.map(|(qry_name, qry_cds_tr)| {
let ref_cds_tr = ref_translation.get_cds(qry_name)?;
let cds = gene_map.get_cds(&qry_cds_tr.name)?;
Ok(find_aa_changes_for_cds(
cds, qry_seq, ref_seq, ref_cds_tr, qry_cds_tr, nuc_subs, nuc_dels,
))
})
.collect::<Result<Vec<FindAaChangesOutput>, Report>>()?
.into_iter()
// Merge changes from all CDSes into one struct
.fold(FindAaChangesOutput::default(), |mut output, changes| {
output.aa_changes_groups.extend(changes.aa_changes_groups);
output.aa_substitutions.extend(changes.aa_substitutions);
output.aa_deletions.extend(changes.aa_deletions);
extend_map_of_vecs(&mut output.nuc_to_aa_muts, changes.nuc_to_aa_muts);
output
});
changes.aa_substitutions.sort();
changes.aa_deletions.sort();
changes.nuc_to_aa_muts.iter_mut().for_each(|(_, vals)| {
vals.sort();
vals.dedup();
});
Ok(changes)
}
/// Finds aminoacid substitutions and deletions in query peptides relative to reference peptides, in one gene
///
/// ## Precondition
/// Nucleotide sequences and peptides are required to be stripped from insertions
///
///
/// ## Implementation details
/// We compare reference and query peptides (extracted by the preceding call to Nextalign),
/// one aminoacid at at time, and deduce changes. We then report the change and relevant nucleotide context surrounding
/// this change.
/// Previously we reported one-to-one mapping of aminoacid changes to corresponding nucleotide changes. However, it
/// was not always accurate, because if there are multiple nucleotide changes in a codon, the direct correspondence
/// might not always be established without knowing the order in which nucleotide changes have occurred. And in the
/// context of Nextclade we don't have this information.
fn find_aa_changes_for_cds(
cds: &Cds,
qry_seq: &[Nuc],
ref_seq: &[Nuc],
ref_tr: &CdsTranslation,
qry_tr: &CdsTranslation,
nuc_subs: &[NucSub],
nuc_dels: &[NucDelRange],
) -> FindAaChangesOutput {
assert_eq!(ref_tr.seq.len(), qry_tr.seq.len());
assert_eq!(qry_seq.len(), ref_seq.len());
let aa_alignment_ranges = &qry_tr.alignment_ranges;
let mut aa_changes_groups = vec![AaChangesGroup::new(&cds.name)];
let mut curr_group = aa_changes_groups.last_mut().unwrap();
for codon in AaRefRange::from_usize(0, qry_tr.seq.len()).iter() {
if !is_codon_sequenced(aa_alignment_ranges, codon) {
continue;
}
let ref_aa = ref_tr.seq[codon.as_usize()];
let qry_aa = qry_tr.seq[codon.as_usize()];
if is_aa_mutated_or_deleted(ref_aa, qry_aa) {
match curr_group.last() {
// If current group is empty, then we are about to insert the first codon into the first group.
None => {
if codon > 0 && is_codon_sequenced(aa_alignment_ranges, codon - 1) {
// Also prepend one codon to the left, for additional context, to start the group
curr_group.push(AaChangeWithContext::new(
cds,
codon - 1,
qry_seq,
ref_seq,
ref_tr,
qry_tr,
));
}
// The current codon itself
curr_group.push(AaChangeWithContext::new(cds, codon, qry_seq, ref_seq, ref_tr, qry_tr));
}
// Current group is not empty
Some(prev) => {
// If previous codon in the group is adjacent or almost adjacent (there is 1 item in between), | // then append to the group.
if codon <= prev.pos + 2 {
// If previous codon in the group is not exactly adjacent, there is 1 item in between,
// then cover the hole by inserting previous codon.
if codon == prev.pos + 2 && is_codon_sequenced(aa_alignment_ranges, codon - 1) {
curr_group.push(AaChangeWithContext::new(
cds,
codon - 1,
qry_seq,
ref_seq,
ref_tr,
qry_tr,
));
}
// And insert the current codon
curr_group.push(AaChangeWithContext::new(cds, codon, qry_seq, ref_seq, ref_tr, qry_tr));
}
// If previous codon in the group is not adjacent, then terminate the current group and start a new group.
else {
// Add one codon to the right, for additional context, to finalize the current group
if is_codon_sequenced(aa | random_line_split |
|
aa_changes.rs | ,
pos: AaRefPosition,
qry_seq: &[Nuc],
ref_seq: &[Nuc],
ref_tr: &CdsTranslation,
qry_tr: &CdsTranslation,
) -> Self {
let ref_aa = ref_tr.seq[pos.as_usize()];
let qry_aa = qry_tr.seq[pos.as_usize()];
let nuc_ranges = cds_codon_pos_to_ref_range(cds, pos);
let ref_triplet = nuc_ranges
.iter()
.flat_map(|(range, strand)| {
let mut nucs = ref_seq[range.to_std()].to_vec();
if strand == &GeneStrand::Reverse {
reverse_complement_in_place(&mut nucs);
}
nucs
})
.collect_vec();
let qry_triplet = nuc_ranges
.iter()
.flat_map(|(range, strand)| {
let mut nucs = qry_seq[range.clamp_range(0, qry_seq.len()).to_std()].to_vec();
if strand == &GeneStrand::Reverse {
reverse_complement_in_place(&mut nucs);
}
nucs
})
.collect_vec();
let nuc_ranges = nuc_ranges.into_iter().map(|(range, _)| range).collect_vec();
Self {
cds_name: cds.name.clone(),
pos,
ref_aa,
qry_aa,
nuc_pos: nuc_ranges[0].begin,
nuc_ranges,
ref_triplet,
qry_triplet,
}
}
#[inline]
pub fn is_mutated_or_deleted(&self) -> bool {
is_aa_mutated_or_deleted(self.ref_aa, self.qry_aa)
}
}
#[derive(Clone, Debug, Default, Serialize, Deserialize, schemars::JsonSchema)]
#[serde(rename_all = "camelCase")]
pub struct AaChangesGroup {
name: String,
range: AaRefRange,
changes: Vec<AaChangeWithContext>,
nuc_subs: Vec<NucSub>,
nuc_dels: Vec<NucDelRange>,
}
impl AaChangesGroup {
pub fn new(name: impl AsRef<str>) -> Self {
Self::with_changes(name, vec![])
}
pub fn with_changes(name: impl AsRef<str>, changes: Vec<AaChangeWithContext>) -> Self {
Self {
name: name.as_ref().to_owned(),
range: Self::find_codon_range(&changes),
changes,
nuc_subs: vec![],
nuc_dels: vec![],
}
}
pub fn push(&mut self, change: AaChangeWithContext) {
self.changes.push(change);
self.range = Self::find_codon_range(&self.changes);
}
pub fn last(&self) -> Option<&AaChangeWithContext> {
self.changes.last()
}
fn find_codon_range(changes: &[AaChangeWithContext]) -> AaRefRange {
match changes.iter().minmax_by_key(|change| change.pos) {
MinMaxResult::NoElements => AaRefRange::from_isize(0, 0),
MinMaxResult::OneElement(one) => AaRefRange::new(one.pos, one.pos + 1),
MinMaxResult::MinMax(first, last) => AaRefRange::new(first.pos, last.pos + 1),
}
}
}
#[derive(Clone, Debug, Default, Serialize, Deserialize, schemars::JsonSchema)]
#[serde(rename_all = "camelCase")]
pub struct FindAaChangesOutput {
pub aa_changes_groups: Vec<AaChangesGroup>,
pub aa_substitutions: Vec<AaSub>,
pub aa_deletions: Vec<AaDel>,
pub nuc_to_aa_muts: BTreeMap<String, Vec<AaSub>>,
}
/// Finds aminoacid substitutions and deletions in query peptides relative to reference peptides, in all genes
///
/// ## Precondition
/// Nucleotide sequences and peptides are required to be stripped from insertions
pub fn find_aa_changes(
ref_seq: &[Nuc],
qry_seq: &[Nuc],
ref_translation: &Translation,
qry_translation: &Translation,
gene_map: &GeneMap,
nuc_subs: &[NucSub],
nuc_dels: &[NucDelRange],
) -> Result<FindAaChangesOutput, Report> {
let mut changes = qry_translation
.iter_cdses()
.map(|(qry_name, qry_cds_tr)| {
let ref_cds_tr = ref_translation.get_cds(qry_name)?;
let cds = gene_map.get_cds(&qry_cds_tr.name)?;
Ok(find_aa_changes_for_cds(
cds, qry_seq, ref_seq, ref_cds_tr, qry_cds_tr, nuc_subs, nuc_dels,
))
})
.collect::<Result<Vec<FindAaChangesOutput>, Report>>()?
.into_iter()
// Merge changes from all CDSes into one struct
.fold(FindAaChangesOutput::default(), |mut output, changes| {
output.aa_changes_groups.extend(changes.aa_changes_groups);
output.aa_substitutions.extend(changes.aa_substitutions);
output.aa_deletions.extend(changes.aa_deletions);
extend_map_of_vecs(&mut output.nuc_to_aa_muts, changes.nuc_to_aa_muts);
output
});
changes.aa_substitutions.sort();
changes.aa_deletions.sort();
changes.nuc_to_aa_muts.iter_mut().for_each(|(_, vals)| {
vals.sort();
vals.dedup();
});
Ok(changes)
}
/// Finds aminoacid substitutions and deletions in query peptides relative to reference peptides, in one gene
///
/// ## Precondition
/// Nucleotide sequences and peptides are required to be stripped from insertions
///
///
/// ## Implementation details
/// We compare reference and query peptides (extracted by the preceding call to Nextalign),
/// one aminoacid at at time, and deduce changes. We then report the change and relevant nucleotide context surrounding
/// this change.
/// Previously we reported one-to-one mapping of aminoacid changes to corresponding nucleotide changes. However, it
/// was not always accurate, because if there are multiple nucleotide changes in a codon, the direct correspondence
/// might not always be established without knowing the order in which nucleotide changes have occurred. And in the
/// context of Nextclade we don't have this information.
fn find_aa_changes_for_cds(
cds: &Cds,
qry_seq: &[Nuc],
ref_seq: &[Nuc],
ref_tr: &CdsTranslation,
qry_tr: &CdsTranslation,
nuc_subs: &[NucSub],
nuc_dels: &[NucDelRange],
) -> FindAaChangesOutput {
assert_eq!(ref_tr.seq.len(), qry_tr.seq.len());
assert_eq!(qry_seq.len(), ref_seq.len());
let aa_alignment_ranges = &qry_tr.alignment_ranges;
let mut aa_changes_groups = vec![AaChangesGroup::new(&cds.name)];
let mut curr_group = aa_changes_groups.last_mut().unwrap();
for codon in AaRefRange::from_usize(0, qry_tr.seq.len()).iter() {
if !is_codon_sequenced(aa_alignment_ranges, codon) {
continue;
}
let ref_aa = ref_tr.seq[codon.as_usize()];
let qry_aa = qry_tr.seq[codon.as_usize()];
if is_aa_mutated_or_deleted(ref_aa, qry_aa) {
match curr_group.last() {
// If current group is empty, then we are about to insert the first codon into the first group.
None => |
// Current group is not empty
Some(prev) => {
// If previous codon in the group is adjacent or almost adjacent (there is 1 item in between),
// then append to the group.
if codon <= prev.pos + 2 {
// If previous codon in the group is not exactly adjacent, there is 1 item in between,
// then cover the hole by inserting previous codon.
if codon == prev.pos + 2 && is_codon_sequenced(aa_alignment_ranges, codon - 1) {
curr_group.push(AaChangeWithContext::new(
cds,
codon - 1,
qry_seq,
ref_seq,
ref_tr,
qry_tr,
));
}
// And insert the current codon
curr_group.push(AaChangeWithContext::new(cds, codon, qry_seq, ref_seq, ref_tr, qry_tr));
}
// If previous codon in the group is not adjacent, then terminate the current group and start a new group.
else {
// Add one codon to the right, for additional context, to finalize the current group
if is_codon_sequ | {
if codon > 0 && is_codon_sequenced(aa_alignment_ranges, codon - 1) {
// Also prepend one codon to the left, for additional context, to start the group
curr_group.push(AaChangeWithContext::new(
cds,
codon - 1,
qry_seq,
ref_seq,
ref_tr,
qry_tr,
));
}
// The current codon itself
curr_group.push(AaChangeWithContext::new(cds, codon, qry_seq, ref_seq, ref_tr, qry_tr));
} | conditional_block |
aa_changes.rs | {
pub cds_name: String,
pub pos: AaRefPosition,
pub ref_aa: Aa,
pub qry_aa: Aa,
pub nuc_pos: NucRefGlobalPosition,
#[schemars(with = "String")]
#[serde(serialize_with = "serde_serialize_seq")]
#[serde(deserialize_with = "serde_deserialize_seq")]
pub ref_triplet: Vec<Nuc>,
#[schemars(with = "String")]
#[serde(serialize_with = "serde_serialize_seq")]
#[serde(deserialize_with = "serde_deserialize_seq")]
pub qry_triplet: Vec<Nuc>,
pub nuc_ranges: Vec<NucRefGlobalRange>,
}
impl AaChangeWithContext {
pub fn new(
cds: &Cds,
pos: AaRefPosition,
qry_seq: &[Nuc],
ref_seq: &[Nuc],
ref_tr: &CdsTranslation,
qry_tr: &CdsTranslation,
) -> Self {
let ref_aa = ref_tr.seq[pos.as_usize()];
let qry_aa = qry_tr.seq[pos.as_usize()];
let nuc_ranges = cds_codon_pos_to_ref_range(cds, pos);
let ref_triplet = nuc_ranges
.iter()
.flat_map(|(range, strand)| {
let mut nucs = ref_seq[range.to_std()].to_vec();
if strand == &GeneStrand::Reverse {
reverse_complement_in_place(&mut nucs);
}
nucs
})
.collect_vec();
let qry_triplet = nuc_ranges
.iter()
.flat_map(|(range, strand)| {
let mut nucs = qry_seq[range.clamp_range(0, qry_seq.len()).to_std()].to_vec();
if strand == &GeneStrand::Reverse {
reverse_complement_in_place(&mut nucs);
}
nucs
})
.collect_vec();
let nuc_ranges = nuc_ranges.into_iter().map(|(range, _)| range).collect_vec();
Self {
cds_name: cds.name.clone(),
pos,
ref_aa,
qry_aa,
nuc_pos: nuc_ranges[0].begin,
nuc_ranges,
ref_triplet,
qry_triplet,
}
}
#[inline]
pub fn is_mutated_or_deleted(&self) -> bool {
is_aa_mutated_or_deleted(self.ref_aa, self.qry_aa)
}
}
#[derive(Clone, Debug, Default, Serialize, Deserialize, schemars::JsonSchema)]
#[serde(rename_all = "camelCase")]
pub struct AaChangesGroup {
name: String,
range: AaRefRange,
changes: Vec<AaChangeWithContext>,
nuc_subs: Vec<NucSub>,
nuc_dels: Vec<NucDelRange>,
}
impl AaChangesGroup {
pub fn new(name: impl AsRef<str>) -> Self {
Self::with_changes(name, vec![])
}
pub fn with_changes(name: impl AsRef<str>, changes: Vec<AaChangeWithContext>) -> Self {
Self {
name: name.as_ref().to_owned(),
range: Self::find_codon_range(&changes),
changes,
nuc_subs: vec![],
nuc_dels: vec![],
}
}
pub fn push(&mut self, change: AaChangeWithContext) {
self.changes.push(change);
self.range = Self::find_codon_range(&self.changes);
}
pub fn last(&self) -> Option<&AaChangeWithContext> {
self.changes.last()
}
fn find_codon_range(changes: &[AaChangeWithContext]) -> AaRefRange {
match changes.iter().minmax_by_key(|change| change.pos) {
MinMaxResult::NoElements => AaRefRange::from_isize(0, 0),
MinMaxResult::OneElement(one) => AaRefRange::new(one.pos, one.pos + 1),
MinMaxResult::MinMax(first, last) => AaRefRange::new(first.pos, last.pos + 1),
}
}
}
#[derive(Clone, Debug, Default, Serialize, Deserialize, schemars::JsonSchema)]
#[serde(rename_all = "camelCase")]
pub struct FindAaChangesOutput {
pub aa_changes_groups: Vec<AaChangesGroup>,
pub aa_substitutions: Vec<AaSub>,
pub aa_deletions: Vec<AaDel>,
pub nuc_to_aa_muts: BTreeMap<String, Vec<AaSub>>,
}
/// Finds aminoacid substitutions and deletions in query peptides relative to reference peptides, in all genes
///
/// ## Precondition
/// Nucleotide sequences and peptides are required to be stripped from insertions
pub fn find_aa_changes(
ref_seq: &[Nuc],
qry_seq: &[Nuc],
ref_translation: &Translation,
qry_translation: &Translation,
gene_map: &GeneMap,
nuc_subs: &[NucSub],
nuc_dels: &[NucDelRange],
) -> Result<FindAaChangesOutput, Report> {
let mut changes = qry_translation
.iter_cdses()
.map(|(qry_name, qry_cds_tr)| {
let ref_cds_tr = ref_translation.get_cds(qry_name)?;
let cds = gene_map.get_cds(&qry_cds_tr.name)?;
Ok(find_aa_changes_for_cds(
cds, qry_seq, ref_seq, ref_cds_tr, qry_cds_tr, nuc_subs, nuc_dels,
))
})
.collect::<Result<Vec<FindAaChangesOutput>, Report>>()?
.into_iter()
// Merge changes from all CDSes into one struct
.fold(FindAaChangesOutput::default(), |mut output, changes| {
output.aa_changes_groups.extend(changes.aa_changes_groups);
output.aa_substitutions.extend(changes.aa_substitutions);
output.aa_deletions.extend(changes.aa_deletions);
extend_map_of_vecs(&mut output.nuc_to_aa_muts, changes.nuc_to_aa_muts);
output
});
changes.aa_substitutions.sort();
changes.aa_deletions.sort();
changes.nuc_to_aa_muts.iter_mut().for_each(|(_, vals)| {
vals.sort();
vals.dedup();
});
Ok(changes)
}
/// Finds aminoacid substitutions and deletions in query peptides relative to reference peptides, in one gene
///
/// ## Precondition
/// Nucleotide sequences and peptides are required to be stripped from insertions
///
///
/// ## Implementation details
/// We compare reference and query peptides (extracted by the preceding call to Nextalign),
/// one aminoacid at at time, and deduce changes. We then report the change and relevant nucleotide context surrounding
/// this change.
/// Previously we reported one-to-one mapping of aminoacid changes to corresponding nucleotide changes. However, it
/// was not always accurate, because if there are multiple nucleotide changes in a codon, the direct correspondence
/// might not always be established without knowing the order in which nucleotide changes have occurred. And in the
/// context of Nextclade we don't have this information.
fn find_aa_changes_for_cds(
cds: &Cds,
qry_seq: &[Nuc],
ref_seq: &[Nuc],
ref_tr: &CdsTranslation,
qry_tr: &CdsTranslation,
nuc_subs: &[NucSub],
nuc_dels: &[NucDelRange],
) -> FindAaChangesOutput {
assert_eq!(ref_tr.seq.len(), qry_tr.seq.len());
assert_eq!(qry_seq.len(), ref_seq.len());
let aa_alignment_ranges = &qry_tr.alignment_ranges;
let mut aa_changes_groups = vec![AaChangesGroup::new(&cds.name)];
let mut curr_group = aa_changes_groups.last_mut().unwrap();
for codon in AaRefRange::from_usize(0, qry_tr.seq.len()).iter() {
if !is_codon_sequenced(aa_alignment_ranges, codon) {
continue;
}
let ref_aa = ref_tr.seq[codon.as_usize()];
let qry_aa = qry_tr.seq[codon.as_usize()];
if is_aa_mutated_or_deleted(ref_aa, qry_aa) {
match curr_group.last() {
// If current group is empty, then we are about to insert the first codon into the first group.
None => {
if codon > 0 && is_codon_sequenced(aa_alignment_ranges, codon - 1) {
// Also prepend one codon to the left, for additional context, to start the group
curr_group.push(AaChangeWithContext::new(
cds,
codon - 1,
qry_seq,
ref_seq,
ref_tr,
qry_tr,
));
}
// The current codon itself
curr_group.push(AaChangeWithContext::new(cds, codon, qry_seq, ref_seq, ref_tr, qry_tr));
}
// Current group is not empty
Some(prev) => {
// If previous codon in the group is adjacent or almost adjacent (there is 1 item in between),
// then append to the group.
if codon <= prev.pos + 2 {
// If previous codon in the group is not exactly adjacent, there is 1 item in between,
// then cover the | AaChangeWithContext | identifier_name |
|
lib.rs | inbound requests from other
/// instances.
///
/// The public listener forwards requests to a local socket (TCP or UNIX).
///
/// The private listener routes requests to service-discovery-aware load-balancer.
///
pub struct Main<G> {
config: config::Config,
control_listener: BoundPort,
inbound_listener: BoundPort,
outbound_listener: BoundPort,
metrics_listener: BoundPort,
get_original_dst: G,
runtime: MainRuntime,
}
impl<G> Main<G>
where
G: GetOriginalDst + Clone + Send + 'static,
{
pub fn new<R>(
config: config::Config,
get_original_dst: G,
runtime: R
) -> Self
where
R: Into<MainRuntime>,
{
let control_listener = BoundPort::new(config.control_listener.addr)
.expect("controller listener bind");
let inbound_listener = BoundPort::new(config.public_listener.addr)
.expect("public listener bind");
let outbound_listener = BoundPort::new(config.private_listener.addr)
.expect("private listener bind");
let runtime = runtime.into();
let metrics_listener = BoundPort::new(config.metrics_listener.addr)
.expect("metrics listener bind");
Main {
config,
control_listener,
inbound_listener,
outbound_listener,
metrics_listener,
get_original_dst,
runtime,
}
}
pub fn control_addr(&self) -> SocketAddr {
self.control_listener.local_addr()
}
pub fn inbound_addr(&self) -> SocketAddr {
self.inbound_listener.local_addr()
}
pub fn outbound_addr(&self) -> SocketAddr {
self.outbound_listener.local_addr()
}
pub fn metrics_addr(&self) -> SocketAddr {
self.metrics_listener.local_addr()
}
pub fn run_until<F>(self, shutdown_signal: F)
where
F: Future<Item = (), Error = ()> + Send + 'static,
{
let process_ctx = ctx::Process::new(&self.config);
let Main {
config,
control_listener,
inbound_listener,
outbound_listener,
metrics_listener,
get_original_dst,
mut runtime,
} = self;
let control_host_and_port = config.control_host_and_port.clone();
info!("using controller at {:?}", control_host_and_port);
info!("routing on {:?}", outbound_listener.local_addr());
info!(
"proxying on {:?} to {:?}",
inbound_listener.local_addr(),
config.private_forward
);
info!(
"serving Prometheus metrics on {:?}",
metrics_listener.local_addr(),
);
info!(
"protocol detection disabled for inbound ports {:?}",
config.inbound_ports_disable_protocol_detection,
);
info!(
"protocol detection disabled for outbound ports {:?}",
config.outbound_ports_disable_protocol_detection,
);
let (taps, observe) = control::Observe::new(100);
let (sensors, telemetry) = telemetry::new(
&process_ctx,
config.event_buffer_capacity,
config.metrics_retain_idle,
&taps,
);
let (tls_client_config, tls_server_config, tls_cfg_bg) =
tls::watch_for_config_changes(
config.tls_settings.as_ref(),
sensors.tls_config(),
);
let controller_tls = config.tls_settings.as_ref().and_then(|settings| {
settings.controller_identity.as_ref().map(|controller_identity| {
tls::ConnectionConfig {
identity: controller_identity.clone(),
config: tls_client_config.clone(),
}
})
});
let (dns_resolver, dns_bg) = dns::Resolver::from_system_config_and_env(&config)
.unwrap_or_else(|e| {
// TODO: Make DNS configuration infallible.
panic!("invalid DNS configuration: {:?}", e);
});
let (resolver, resolver_bg) = control::destination::new(
dns_resolver.clone(),
config.namespaces.clone(),
control_host_and_port,
controller_tls,
);
let (drain_tx, drain_rx) = drain::channel();
let bind = Bind::new(tls_client_config).with_sensors(sensors.clone());
// Setup the public listener. This will listen on a publicly accessible
// address and listen for inbound connections that should be forwarded
// to the managed application (private destination).
let inbound = {
let ctx = ctx::Proxy::inbound(&process_ctx);
let bind = bind.clone().with_ctx(ctx.clone());
let default_addr = config.private_forward.map(|a| a.into());
let router = Router::new(
Inbound::new(default_addr, bind),
config.inbound_router_capacity,
config.inbound_router_max_idle_age,
);
let tls_settings = config.tls_settings.as_ref().map(|settings| {
tls::ConnectionConfig {
identity: settings.pod_identity.clone(),
config: tls_server_config
}
});
serve(
inbound_listener,
tls_settings,
router,
config.private_connect_timeout,
config.inbound_ports_disable_protocol_detection,
ctx,
sensors.clone(),
get_original_dst.clone(),
drain_rx.clone(),
)
};
// Setup the private listener. This will listen on a locally accessible
// address and listen for outbound requests that should be routed
// to a remote service (public destination).
let outbound = {
let ctx = ctx::Proxy::outbound(&process_ctx);
let bind = bind.clone().with_ctx(ctx.clone());
let router = Router::new(
Outbound::new(bind, resolver, config.bind_timeout),
config.outbound_router_capacity,
config.outbound_router_max_idle_age,
);
serve(
outbound_listener,
Conditional::None(tls::ReasonForNoTls::InternalTraffic),
router,
config.public_connect_timeout,
config.outbound_ports_disable_protocol_detection,
ctx,
sensors,
get_original_dst,
drain_rx,
)
};
trace!("running");
let (_tx, admin_shutdown_signal) = futures::sync::oneshot::channel::<()>();
{
thread::Builder::new()
.name("admin".into())
.spawn(move || {
use conduit_proxy_controller_grpc::tap::server::TapServer;
let mut rt = current_thread::Runtime::new()
.expect("initialize admin thread runtime");
let tap = serve_tap(control_listener, TapServer::new(observe));
let metrics_server = telemetry.serve_metrics(metrics_listener);
let fut = ::logging::admin().bg("resolver").future(resolver_bg)
.join5(
::logging::admin().bg("telemetry").future(telemetry),
tap.map_err(|_| {}),
metrics_server.map_err(|_| {}),
::logging::admin().bg("dns-resolver").future(dns_bg),
)
// There's no `Future::join6` combinator...
.join(::logging::admin().bg("tls-config").future(tls_cfg_bg))
.map(|_| {});
rt.spawn(Box::new(fut));
let shutdown = admin_shutdown_signal.then(|_| Ok::<(), ()>(()));
rt.block_on(shutdown).expect("admin");
trace!("admin shutdown finished");
})
.expect("initialize controller api thread");
trace!("controller client thread spawned");
}
let fut = inbound
.join(outbound)
.map(|_| ())
.map_err(|err| error!("main error: {:?}", err));
runtime.spawn(Box::new(fut));
trace!("main task spawned");
let shutdown_signal = shutdown_signal.and_then(move |()| {
debug!("shutdown signaled");
drain_tx.drain()
});
runtime.run_until(shutdown_signal).expect("executor");
debug!("shutdown complete");
}
}
fn serve<R, B, E, F, G>(
bound_port: BoundPort,
tls_config: tls::ConditionalConnectionConfig<tls::ServerConfigWatch>,
router: Router<R>,
tcp_connect_timeout: Duration,
disable_protocol_detection_ports: IndexSet<u16>,
proxy_ctx: Arc<ctx::Proxy>,
sensors: telemetry::Sensors,
get_orig_dst: G,
drain_rx: drain::Watch,
) -> impl Future<Item = (), Error = io::Error> + Send + 'static
where
B: tower_h2::Body + Default + Send + 'static,
B::Data: Send,
<B::Data as ::bytes::IntoBuf>::Buf: Send,
E: Error + Send + 'static,
F: Error + Send + 'static,
R: Recognize<
Request = http::Request<HttpBody>,
Response = http::Response<B>,
Error = E,
RouteError = F,
>
+ Send + Sync + 'static,
R::Key: Send,
R::Service: Send,
<R::Service as tower_service::Service>::Future: Send,
Router<R>: Send,
G: GetOriginalDst + Send + 'static,
{
let stack = Arc::new(NewServiceFn::new(move || {
// Clone the router handle
let router = router.clone();
// Map errors to appropriate response error codes.
let map_err = MapErr::new(router, |e| {
match e {
RouteError::Route(r) => {
error!(" turning route error: {} into 500", r);
http::StatusCode::INTERNAL_SERVER_ERROR
}
RouteError::Inner(i) => | {
error!("turning {} into 500", i);
http::StatusCode::INTERNAL_SERVER_ERROR
} | conditional_block |
|
lib.rs | -discovery-aware load-balancer.
///
pub struct Main<G> {
config: config::Config,
control_listener: BoundPort,
inbound_listener: BoundPort,
outbound_listener: BoundPort,
metrics_listener: BoundPort,
get_original_dst: G,
runtime: MainRuntime,
}
impl<G> Main<G>
where
G: GetOriginalDst + Clone + Send + 'static,
{
pub fn new<R>(
config: config::Config,
get_original_dst: G,
runtime: R
) -> Self
where
R: Into<MainRuntime>,
{
let control_listener = BoundPort::new(config.control_listener.addr)
.expect("controller listener bind");
let inbound_listener = BoundPort::new(config.public_listener.addr)
.expect("public listener bind");
let outbound_listener = BoundPort::new(config.private_listener.addr)
.expect("private listener bind");
let runtime = runtime.into();
let metrics_listener = BoundPort::new(config.metrics_listener.addr)
.expect("metrics listener bind");
Main {
config,
control_listener,
inbound_listener,
outbound_listener,
metrics_listener,
get_original_dst,
runtime,
}
}
pub fn control_addr(&self) -> SocketAddr {
self.control_listener.local_addr()
}
pub fn inbound_addr(&self) -> SocketAddr {
self.inbound_listener.local_addr()
}
pub fn outbound_addr(&self) -> SocketAddr {
self.outbound_listener.local_addr()
}
pub fn metrics_addr(&self) -> SocketAddr {
self.metrics_listener.local_addr()
}
pub fn run_until<F>(self, shutdown_signal: F)
where
F: Future<Item = (), Error = ()> + Send + 'static,
{
let process_ctx = ctx::Process::new(&self.config);
let Main {
config,
control_listener,
inbound_listener,
outbound_listener,
metrics_listener,
get_original_dst,
mut runtime,
} = self;
let control_host_and_port = config.control_host_and_port.clone();
info!("using controller at {:?}", control_host_and_port);
info!("routing on {:?}", outbound_listener.local_addr());
info!(
"proxying on {:?} to {:?}",
inbound_listener.local_addr(),
config.private_forward
);
info!(
"serving Prometheus metrics on {:?}",
metrics_listener.local_addr(),
);
info!(
"protocol detection disabled for inbound ports {:?}",
config.inbound_ports_disable_protocol_detection,
);
info!(
"protocol detection disabled for outbound ports {:?}",
config.outbound_ports_disable_protocol_detection,
);
let (taps, observe) = control::Observe::new(100);
let (sensors, telemetry) = telemetry::new(
&process_ctx,
config.event_buffer_capacity,
config.metrics_retain_idle,
&taps,
);
let (tls_client_config, tls_server_config, tls_cfg_bg) =
tls::watch_for_config_changes(
config.tls_settings.as_ref(),
sensors.tls_config(),
);
let controller_tls = config.tls_settings.as_ref().and_then(|settings| {
settings.controller_identity.as_ref().map(|controller_identity| {
tls::ConnectionConfig {
identity: controller_identity.clone(),
config: tls_client_config.clone(),
}
})
});
let (dns_resolver, dns_bg) = dns::Resolver::from_system_config_and_env(&config)
.unwrap_or_else(|e| {
// TODO: Make DNS configuration infallible.
panic!("invalid DNS configuration: {:?}", e);
});
let (resolver, resolver_bg) = control::destination::new(
dns_resolver.clone(),
config.namespaces.clone(),
control_host_and_port,
controller_tls,
);
let (drain_tx, drain_rx) = drain::channel();
let bind = Bind::new(tls_client_config).with_sensors(sensors.clone());
// Setup the public listener. This will listen on a publicly accessible
// address and listen for inbound connections that should be forwarded
// to the managed application (private destination).
let inbound = {
let ctx = ctx::Proxy::inbound(&process_ctx);
let bind = bind.clone().with_ctx(ctx.clone());
let default_addr = config.private_forward.map(|a| a.into());
let router = Router::new(
Inbound::new(default_addr, bind),
config.inbound_router_capacity,
config.inbound_router_max_idle_age,
);
let tls_settings = config.tls_settings.as_ref().map(|settings| {
tls::ConnectionConfig {
identity: settings.pod_identity.clone(),
config: tls_server_config
}
});
serve(
inbound_listener,
tls_settings,
router,
config.private_connect_timeout,
config.inbound_ports_disable_protocol_detection,
ctx,
sensors.clone(),
get_original_dst.clone(),
drain_rx.clone(),
)
};
// Setup the private listener. This will listen on a locally accessible
// address and listen for outbound requests that should be routed
// to a remote service (public destination).
let outbound = {
let ctx = ctx::Proxy::outbound(&process_ctx);
let bind = bind.clone().with_ctx(ctx.clone());
let router = Router::new(
Outbound::new(bind, resolver, config.bind_timeout),
config.outbound_router_capacity,
config.outbound_router_max_idle_age,
);
serve(
outbound_listener,
Conditional::None(tls::ReasonForNoTls::InternalTraffic),
router,
config.public_connect_timeout,
config.outbound_ports_disable_protocol_detection,
ctx,
sensors,
get_original_dst,
drain_rx,
)
};
trace!("running");
let (_tx, admin_shutdown_signal) = futures::sync::oneshot::channel::<()>();
{
thread::Builder::new()
.name("admin".into())
.spawn(move || {
use conduit_proxy_controller_grpc::tap::server::TapServer;
let mut rt = current_thread::Runtime::new()
.expect("initialize admin thread runtime");
let tap = serve_tap(control_listener, TapServer::new(observe));
let metrics_server = telemetry.serve_metrics(metrics_listener);
let fut = ::logging::admin().bg("resolver").future(resolver_bg)
.join5(
::logging::admin().bg("telemetry").future(telemetry),
tap.map_err(|_| {}),
metrics_server.map_err(|_| {}),
::logging::admin().bg("dns-resolver").future(dns_bg),
)
// There's no `Future::join6` combinator...
.join(::logging::admin().bg("tls-config").future(tls_cfg_bg))
.map(|_| {});
rt.spawn(Box::new(fut));
let shutdown = admin_shutdown_signal.then(|_| Ok::<(), ()>(()));
rt.block_on(shutdown).expect("admin");
trace!("admin shutdown finished");
})
.expect("initialize controller api thread");
trace!("controller client thread spawned");
}
let fut = inbound
.join(outbound)
.map(|_| ())
.map_err(|err| error!("main error: {:?}", err));
runtime.spawn(Box::new(fut));
trace!("main task spawned");
let shutdown_signal = shutdown_signal.and_then(move |()| {
debug!("shutdown signaled");
drain_tx.drain()
});
runtime.run_until(shutdown_signal).expect("executor");
debug!("shutdown complete");
}
}
fn serve<R, B, E, F, G>(
bound_port: BoundPort,
tls_config: tls::ConditionalConnectionConfig<tls::ServerConfigWatch>,
router: Router<R>,
tcp_connect_timeout: Duration,
disable_protocol_detection_ports: IndexSet<u16>,
proxy_ctx: Arc<ctx::Proxy>,
sensors: telemetry::Sensors,
get_orig_dst: G,
drain_rx: drain::Watch,
) -> impl Future<Item = (), Error = io::Error> + Send + 'static
where
B: tower_h2::Body + Default + Send + 'static,
B::Data: Send,
<B::Data as ::bytes::IntoBuf>::Buf: Send,
E: Error + Send + 'static,
F: Error + Send + 'static,
R: Recognize<
Request = http::Request<HttpBody>,
Response = http::Response<B>,
Error = E,
RouteError = F,
>
+ Send + Sync + 'static,
R::Key: Send,
R::Service: Send,
<R::Service as tower_service::Service>::Future: Send,
Router<R>: Send,
G: GetOriginalDst + Send + 'static,
| {
let stack = Arc::new(NewServiceFn::new(move || {
// Clone the router handle
let router = router.clone();
// Map errors to appropriate response error codes.
let map_err = MapErr::new(router, |e| {
match e {
RouteError::Route(r) => {
error!(" turning route error: {} into 500", r);
http::StatusCode::INTERNAL_SERVER_ERROR
}
RouteError::Inner(i) => {
error!("turning {} into 500", i);
http::StatusCode::INTERNAL_SERVER_ERROR
}
RouteError::NotRecognized => {
error!("turning route not recognized error into 500");
http::StatusCode::INTERNAL_SERVER_ERROR
} | identifier_body |
|
lib.rs | pub mod ctx;
mod dns;
mod drain;
pub mod fs_watch;
mod inbound;
mod logging;
mod map_err;
mod outbound;
pub mod stream;
pub mod task;
pub mod telemetry;
mod transparency;
mod transport;
pub mod timeout;
mod tower_fn; // TODO: move to tower-fn
mod watch_service; // TODO: move to tower
use bind::Bind;
use conditional::Conditional;
use connection::BoundPort;
use inbound::Inbound;
use map_err::MapErr;
use task::MainRuntime;
use transparency::{HttpBody, Server};
pub use transport::{AddrInfo, GetOriginalDst, SoOriginalDst, tls};
use outbound::Outbound;
pub use watch_service::WatchService;
/// Runs a sidecar proxy.
///
/// The proxy binds two listeners:
///
/// - a private socket (TCP or UNIX) for outbound requests to other instances;
/// - and a public socket (TCP and optionally TLS) for inbound requests from other
/// instances.
///
/// The public listener forwards requests to a local socket (TCP or UNIX).
///
/// The private listener routes requests to service-discovery-aware load-balancer.
///
pub struct Main<G> {
config: config::Config,
control_listener: BoundPort,
inbound_listener: BoundPort,
outbound_listener: BoundPort,
metrics_listener: BoundPort,
get_original_dst: G,
runtime: MainRuntime,
}
impl<G> Main<G>
where
G: GetOriginalDst + Clone + Send + 'static,
{
pub fn new<R>(
config: config::Config,
get_original_dst: G,
runtime: R
) -> Self
where
R: Into<MainRuntime>,
{
let control_listener = BoundPort::new(config.control_listener.addr)
.expect("controller listener bind");
let inbound_listener = BoundPort::new(config.public_listener.addr)
.expect("public listener bind");
let outbound_listener = BoundPort::new(config.private_listener.addr)
.expect("private listener bind");
let runtime = runtime.into();
let metrics_listener = BoundPort::new(config.metrics_listener.addr)
.expect("metrics listener bind");
Main {
config,
control_listener,
inbound_listener,
outbound_listener,
metrics_listener,
get_original_dst,
runtime,
}
}
pub fn control_addr(&self) -> SocketAddr {
self.control_listener.local_addr()
}
pub fn inbound_addr(&self) -> SocketAddr {
self.inbound_listener.local_addr()
}
pub fn outbound_addr(&self) -> SocketAddr {
self.outbound_listener.local_addr()
}
pub fn metrics_addr(&self) -> SocketAddr {
self.metrics_listener.local_addr()
}
pub fn run_until<F>(self, shutdown_signal: F)
where
F: Future<Item = (), Error = ()> + Send + 'static,
{
let process_ctx = ctx::Process::new(&self.config);
let Main {
config,
control_listener,
inbound_listener,
outbound_listener,
metrics_listener,
get_original_dst,
mut runtime,
} = self;
let control_host_and_port = config.control_host_and_port.clone();
info!("using controller at {:?}", control_host_and_port);
info!("routing on {:?}", outbound_listener.local_addr());
info!(
"proxying on {:?} to {:?}",
inbound_listener.local_addr(),
config.private_forward
);
info!(
"serving Prometheus metrics on {:?}",
metrics_listener.local_addr(),
);
info!(
"protocol detection disabled for inbound ports {:?}",
config.inbound_ports_disable_protocol_detection,
);
info!(
"protocol detection disabled for outbound ports {:?}",
config.outbound_ports_disable_protocol_detection,
);
let (taps, observe) = control::Observe::new(100);
let (sensors, telemetry) = telemetry::new(
&process_ctx,
config.event_buffer_capacity,
config.metrics_retain_idle,
&taps,
);
let (tls_client_config, tls_server_config, tls_cfg_bg) =
tls::watch_for_config_changes(
config.tls_settings.as_ref(),
sensors.tls_config(),
);
let controller_tls = config.tls_settings.as_ref().and_then(|settings| {
settings.controller_identity.as_ref().map(|controller_identity| {
tls::ConnectionConfig {
identity: controller_identity.clone(),
config: tls_client_config.clone(),
}
})
});
let (dns_resolver, dns_bg) = dns::Resolver::from_system_config_and_env(&config)
.unwrap_or_else(|e| {
// TODO: Make DNS configuration infallible.
panic!("invalid DNS configuration: {:?}", e);
});
let (resolver, resolver_bg) = control::destination::new(
dns_resolver.clone(),
config.namespaces.clone(),
control_host_and_port,
controller_tls,
);
let (drain_tx, drain_rx) = drain::channel();
let bind = Bind::new(tls_client_config).with_sensors(sensors.clone());
// Setup the public listener. This will listen on a publicly accessible
// address and listen for inbound connections that should be forwarded
// to the managed application (private destination).
let inbound = {
let ctx = ctx::Proxy::inbound(&process_ctx);
let bind = bind.clone().with_ctx(ctx.clone());
let default_addr = config.private_forward.map(|a| a.into());
let router = Router::new(
Inbound::new(default_addr, bind),
config.inbound_router_capacity,
config.inbound_router_max_idle_age,
);
let tls_settings = config.tls_settings.as_ref().map(|settings| {
tls::ConnectionConfig {
identity: settings.pod_identity.clone(),
config: tls_server_config
}
});
serve(
inbound_listener,
tls_settings,
router,
config.private_connect_timeout,
config.inbound_ports_disable_protocol_detection,
ctx,
sensors.clone(),
get_original_dst.clone(),
drain_rx.clone(),
)
};
// Setup the private listener. This will listen on a locally accessible
// address and listen for outbound requests that should be routed
// to a remote service (public destination).
let outbound = {
let ctx = ctx::Proxy::outbound(&process_ctx);
let bind = bind.clone().with_ctx(ctx.clone());
let router = Router::new(
Outbound::new(bind, resolver, config.bind_timeout),
config.outbound_router_capacity,
config.outbound_router_max_idle_age,
);
serve(
outbound_listener,
Conditional::None(tls::ReasonForNoTls::InternalTraffic),
router,
config.public_connect_timeout,
config.outbound_ports_disable_protocol_detection,
ctx,
sensors,
get_original_dst,
drain_rx,
)
};
trace!("running");
let (_tx, admin_shutdown_signal) = futures::sync::oneshot::channel::<()>();
{
thread::Builder::new()
.name("admin".into())
.spawn(move || {
use conduit_proxy_controller_grpc::tap::server::TapServer;
let mut rt = current_thread::Runtime::new()
.expect("initialize admin thread runtime");
let tap = serve_tap(control_listener, TapServer::new(observe));
let metrics_server = telemetry.serve_metrics(metrics_listener);
let fut = ::logging::admin().bg("resolver").future(resolver_bg)
.join5(
::logging::admin().bg("telemetry").future(telemetry),
tap.map_err(|_| {}),
metrics_server.map_err(|_| {}),
::logging::admin().bg("dns-resolver").future(dns_bg),
)
// There's no `Future::join6` combinator...
.join(::logging::admin().bg("tls-config").future(tls_cfg_bg))
.map(|_| {});
rt.spawn(Box::new(fut));
let shutdown = admin_shutdown_signal.then(|_| Ok::<(), ()>(()));
rt.block_on(shutdown).expect("admin");
trace!("admin shutdown finished");
})
.expect("initialize controller api thread");
trace!("controller client thread spawned");
}
let fut = inbound
.join(outbound)
.map(|_| ())
.map_err(|err| error!("main error: {:?}", err));
runtime.spawn(Box::new(fut));
trace!("main task spawned");
let shutdown_signal = shutdown_signal.and_then(move |()| {
debug!("shutdown signaled");
drain_tx.drain()
});
runtime.run_until(shutdown_signal).expect("executor");
debug!("shutdown complete");
}
}
fn serve<R, B, E, F, G>(
bound_port: BoundPort,
tls_config: tls::ConditionalConnectionConfig<tls::ServerConfigWatch>,
router: Router<R>,
tcp_connect_timeout: Duration,
disable_protocol_detection_ports: IndexSet<u16>,
proxy_ctx: Arc<ctx::Proxy>,
sensors: telemetry::Sensors,
get_orig_dst: G,
drain_rx: drain::Watch,
) -> impl Future<Item = (), Error = io::Error> + Send + 'static
where
B: tower_h2::Body + Default + Send + 'static,
B::Data: Send,
<B::Data as ::bytes::IntoBuf>::Buf: Send,
E: Error + Send + 'static,
F: Error + Send + 'static,
R: Recognize<
Request = http::Request<HttpBody>,
Response = http:: | pub mod control;
pub mod convert; | random_line_split |
|
lib.rs | mod outbound;
pub mod stream;
pub mod task;
pub mod telemetry;
mod transparency;
mod transport;
pub mod timeout;
mod tower_fn; // TODO: move to tower-fn
mod watch_service; // TODO: move to tower
use bind::Bind;
use conditional::Conditional;
use connection::BoundPort;
use inbound::Inbound;
use map_err::MapErr;
use task::MainRuntime;
use transparency::{HttpBody, Server};
pub use transport::{AddrInfo, GetOriginalDst, SoOriginalDst, tls};
use outbound::Outbound;
pub use watch_service::WatchService;
/// Runs a sidecar proxy.
///
/// The proxy binds two listeners:
///
/// - a private socket (TCP or UNIX) for outbound requests to other instances;
/// - and a public socket (TCP and optionally TLS) for inbound requests from other
/// instances.
///
/// The public listener forwards requests to a local socket (TCP or UNIX).
///
/// The private listener routes requests to service-discovery-aware load-balancer.
///
pub struct Main<G> {
config: config::Config,
control_listener: BoundPort,
inbound_listener: BoundPort,
outbound_listener: BoundPort,
metrics_listener: BoundPort,
get_original_dst: G,
runtime: MainRuntime,
}
impl<G> Main<G>
where
G: GetOriginalDst + Clone + Send + 'static,
{
pub fn new<R>(
config: config::Config,
get_original_dst: G,
runtime: R
) -> Self
where
R: Into<MainRuntime>,
{
let control_listener = BoundPort::new(config.control_listener.addr)
.expect("controller listener bind");
let inbound_listener = BoundPort::new(config.public_listener.addr)
.expect("public listener bind");
let outbound_listener = BoundPort::new(config.private_listener.addr)
.expect("private listener bind");
let runtime = runtime.into();
let metrics_listener = BoundPort::new(config.metrics_listener.addr)
.expect("metrics listener bind");
Main {
config,
control_listener,
inbound_listener,
outbound_listener,
metrics_listener,
get_original_dst,
runtime,
}
}
pub fn control_addr(&self) -> SocketAddr {
self.control_listener.local_addr()
}
pub fn | (&self) -> SocketAddr {
self.inbound_listener.local_addr()
}
pub fn outbound_addr(&self) -> SocketAddr {
self.outbound_listener.local_addr()
}
pub fn metrics_addr(&self) -> SocketAddr {
self.metrics_listener.local_addr()
}
pub fn run_until<F>(self, shutdown_signal: F)
where
F: Future<Item = (), Error = ()> + Send + 'static,
{
let process_ctx = ctx::Process::new(&self.config);
let Main {
config,
control_listener,
inbound_listener,
outbound_listener,
metrics_listener,
get_original_dst,
mut runtime,
} = self;
let control_host_and_port = config.control_host_and_port.clone();
info!("using controller at {:?}", control_host_and_port);
info!("routing on {:?}", outbound_listener.local_addr());
info!(
"proxying on {:?} to {:?}",
inbound_listener.local_addr(),
config.private_forward
);
info!(
"serving Prometheus metrics on {:?}",
metrics_listener.local_addr(),
);
info!(
"protocol detection disabled for inbound ports {:?}",
config.inbound_ports_disable_protocol_detection,
);
info!(
"protocol detection disabled for outbound ports {:?}",
config.outbound_ports_disable_protocol_detection,
);
let (taps, observe) = control::Observe::new(100);
let (sensors, telemetry) = telemetry::new(
&process_ctx,
config.event_buffer_capacity,
config.metrics_retain_idle,
&taps,
);
let (tls_client_config, tls_server_config, tls_cfg_bg) =
tls::watch_for_config_changes(
config.tls_settings.as_ref(),
sensors.tls_config(),
);
let controller_tls = config.tls_settings.as_ref().and_then(|settings| {
settings.controller_identity.as_ref().map(|controller_identity| {
tls::ConnectionConfig {
identity: controller_identity.clone(),
config: tls_client_config.clone(),
}
})
});
let (dns_resolver, dns_bg) = dns::Resolver::from_system_config_and_env(&config)
.unwrap_or_else(|e| {
// TODO: Make DNS configuration infallible.
panic!("invalid DNS configuration: {:?}", e);
});
let (resolver, resolver_bg) = control::destination::new(
dns_resolver.clone(),
config.namespaces.clone(),
control_host_and_port,
controller_tls,
);
let (drain_tx, drain_rx) = drain::channel();
let bind = Bind::new(tls_client_config).with_sensors(sensors.clone());
// Setup the public listener. This will listen on a publicly accessible
// address and listen for inbound connections that should be forwarded
// to the managed application (private destination).
let inbound = {
let ctx = ctx::Proxy::inbound(&process_ctx);
let bind = bind.clone().with_ctx(ctx.clone());
let default_addr = config.private_forward.map(|a| a.into());
let router = Router::new(
Inbound::new(default_addr, bind),
config.inbound_router_capacity,
config.inbound_router_max_idle_age,
);
let tls_settings = config.tls_settings.as_ref().map(|settings| {
tls::ConnectionConfig {
identity: settings.pod_identity.clone(),
config: tls_server_config
}
});
serve(
inbound_listener,
tls_settings,
router,
config.private_connect_timeout,
config.inbound_ports_disable_protocol_detection,
ctx,
sensors.clone(),
get_original_dst.clone(),
drain_rx.clone(),
)
};
// Setup the private listener. This will listen on a locally accessible
// address and listen for outbound requests that should be routed
// to a remote service (public destination).
let outbound = {
let ctx = ctx::Proxy::outbound(&process_ctx);
let bind = bind.clone().with_ctx(ctx.clone());
let router = Router::new(
Outbound::new(bind, resolver, config.bind_timeout),
config.outbound_router_capacity,
config.outbound_router_max_idle_age,
);
serve(
outbound_listener,
Conditional::None(tls::ReasonForNoTls::InternalTraffic),
router,
config.public_connect_timeout,
config.outbound_ports_disable_protocol_detection,
ctx,
sensors,
get_original_dst,
drain_rx,
)
};
trace!("running");
let (_tx, admin_shutdown_signal) = futures::sync::oneshot::channel::<()>();
{
thread::Builder::new()
.name("admin".into())
.spawn(move || {
use conduit_proxy_controller_grpc::tap::server::TapServer;
let mut rt = current_thread::Runtime::new()
.expect("initialize admin thread runtime");
let tap = serve_tap(control_listener, TapServer::new(observe));
let metrics_server = telemetry.serve_metrics(metrics_listener);
let fut = ::logging::admin().bg("resolver").future(resolver_bg)
.join5(
::logging::admin().bg("telemetry").future(telemetry),
tap.map_err(|_| {}),
metrics_server.map_err(|_| {}),
::logging::admin().bg("dns-resolver").future(dns_bg),
)
// There's no `Future::join6` combinator...
.join(::logging::admin().bg("tls-config").future(tls_cfg_bg))
.map(|_| {});
rt.spawn(Box::new(fut));
let shutdown = admin_shutdown_signal.then(|_| Ok::<(), ()>(()));
rt.block_on(shutdown).expect("admin");
trace!("admin shutdown finished");
})
.expect("initialize controller api thread");
trace!("controller client thread spawned");
}
let fut = inbound
.join(outbound)
.map(|_| ())
.map_err(|err| error!("main error: {:?}", err));
runtime.spawn(Box::new(fut));
trace!("main task spawned");
let shutdown_signal = shutdown_signal.and_then(move |()| {
debug!("shutdown signaled");
drain_tx.drain()
});
runtime.run_until(shutdown_signal).expect("executor");
debug!("shutdown complete");
}
}
fn serve<R, B, E, F, G>(
bound_port: BoundPort,
tls_config: tls::ConditionalConnectionConfig<tls::ServerConfigWatch>,
router: Router<R>,
tcp_connect_timeout: Duration,
disable_protocol_detection_ports: IndexSet<u16>,
proxy_ctx: Arc<ctx::Proxy>,
sensors: telemetry::Sensors,
get_orig_dst: G,
drain_rx: drain::Watch,
) -> impl Future<Item = (), Error = io::Error> + Send + 'static
where
B: tower_h2::Body + Default + Send + 'static,
B::Data: Send,
<B::Data as ::bytes::IntoBuf>::Buf: Send,
E: Error + Send + 'static,
F: Error + Send + 'static,
R: Recognize<
Request = http::Request<HttpBody>,
Response = http::Response<B>,
Error = E,
RouteError = F,
>
+ Send + Sync + 'static,
R::Key: Send | inbound_addr | identifier_name |
auto_aliasing.go | type changes from `T to List[T]` or vice versa. To
// avoid these breaking changes, this method undoes any upstream changes to MaxItems using
// [SchemaInfo.MaxItemsOne] overrides. This happens until a major version change is
// detected, and then overrides are cleared. Effectively this makes sure that upstream
// MaxItems changes are deferred until the next major version.
//
// Implementation note: to operate correctly this method needs to keep a persistent track
// of a database of past decision history. This is currently done by doing reads and
// writes to `providerInfo.GetMetadata()`, which is assumed to be persisted across
// provider releases. The bridge framework keeps this information written out to an opaque
// `bridge-metadata.json` blob which is expected to be stored in source control to persist
// across releases.
//
// Panics if [ProviderInfo.ApplyAutoAliases] would return an error.
func (info *ProviderInfo) MustApplyAutoAliases() {
err := info.ApplyAutoAliases()
contract.AssertNoErrorf(err, "Failed to apply aliases")
}
// Automatically applies backwards compatibility best practices.
//
// Specifically, [ApplyAutoAliases] may perform the following actions:
//
// - Call [ProviderInfo.RenameResourceWithAlias] or [ProviderInfo.RenameDataSource]
// - Edit [ResourceInfo.Aliases]
// - Edit [SchemaInfo.MaxItemsOne]
//
// The goal is to always maximize backwards compatibility and reduce breaking changes for
// the users of the Pulumi providers.
//
// Resource aliases help mask TF provider resource renames or changes in mapped tokens so
// older programs continue to work. See [ResourceInfo.RenameResourceWithAlias] and
// [ResourceInfo.Aliases] for more details.
//
// [SchemaInfo.MaxItemsOne] changes are also important because they involve flattening and
// pluralizing names. Collections (lists or sets) marked with MaxItems=1 are projected as
// scalar types in Pulumi SDKs. Therefore changes to the MaxItems property may be breaking
// the compilation of programs as the type changes from `T to List[T]` or vice versa. To
// avoid these breaking changes, this method undoes any upstream changes to MaxItems using
// [SchemaInfo.MaxItemsOne] overrides. This happens until a major version change is
// detected, and then overrides are cleared. Effectively this makes sure that upstream
// MaxItems changes are deferred until the next major version.
//
// Implementation note: to operate correctly this method needs to keep a persistent track
// of a database of past decision history. This is currently done by doing reads and
// writes to `providerInfo.GetMetadata()`, which is assumed to be persisted across
// provider releases. The bridge framework keeps this information written out to an opaque
// `bridge-metadata.json` blob which is expected to be stored in source control to persist
// across releas
func (info *ProviderInfo) ApplyAutoAliases() error {
artifact := info.GetMetadata()
hist, err := getHistory(artifact)
if err != nil {
return err
}
var currentVersion int
// If version is missing, we assume the current version is the most recent major
// version in mentioned in history.
if info.Version != "" {
v, err := semver.NewVersion(info.Version)
if err != nil {
return err
}
currentVersion = int(v.Major())
} else {
for _, r := range hist.Resources {
for _, p := range r.Past {
if p.MajorVersion > currentVersion {
currentVersion = p.MajorVersion
}
}
}
for _, d := range hist.DataSources {
for _, p := range d.Past {
if p.MajorVersion > currentVersion {
currentVersion = p.MajorVersion
}
}
}
}
rMap := info.P.ResourcesMap()
dMap := info.P.DataSourcesMap()
// Applying resource aliases adds new resources to providerInfo.Resources. To keep
// this process deterministic, we don't apply resource aliases until all resources
// have been examined.
//
// The same logic applies to datasources.
applyAliases := []func(){}
for tfToken, computed := range info.Resources {
r, _ := rMap.GetOk(tfToken)
aliasResource(info, r, &applyAliases, hist.Resources,
computed, tfToken, currentVersion)
}
for tfToken, computed := range info.DataSources {
ds, _ := dMap.GetOk(tfToken)
aliasDataSource(info, ds, &applyAliases, hist.DataSources,
computed, tfToken, currentVersion)
}
for _, f := range applyAliases {
f()
}
if err := md.Set(artifact, aliasMetadataKey, hist); err != nil {
// Set fails only when `hist` is not serializable. Because `hist` is
// composed of marshallable, non-cyclic types, this is impossible.
contract.AssertNoErrorf(err, "History failed to serialize")
}
return nil
}
const aliasMetadataKey = "auto-aliasing"
func getHistory(artifact ProviderMetadata) (aliasHistory, error) |
func aliasResource(
p *ProviderInfo, res shim.Resource,
applyResourceAliases *[]func(),
hist map[string]*tokenHistory[tokens.Type], computed *ResourceInfo,
tfToken string, version int,
) {
prev, hasPrev := hist[tfToken]
if !hasPrev {
// It's not in the history, so it must be new. Stick it in the history for
// next time.
hist[tfToken] = &tokenHistory[tokens.Type]{
Current: computed.Tok,
}
} else {
// We don't do this eagerly because aliasResource is called while
// iterating over p.Resources which aliasOrRenameResource mutates.
*applyResourceAliases = append(*applyResourceAliases,
func() { aliasOrRenameResource(p, computed, tfToken, prev, version) })
}
// Apply Aliasing to MaxItemOne by traversing the field tree and applying the
// stored value.
//
// Note: If the user explicitly sets a MaxItemOne value, that value is respected
// and overwrites the current history.'
if res == nil {
return
}
// If we are behind the major version, reset the fields and the major version.
if hist[tfToken].MajorVersion < version {
hist[tfToken].MajorVersion = version
hist[tfToken].Fields = nil
}
applyResourceMaxItemsOneAliasing(res, &hist[tfToken].Fields, &computed.Fields)
}
// applyResourceMaxItemsOneAliasing traverses a shim.Resource, applying walk to each field in the resource.
func applyResourceMaxItemsOneAliasing(
r shim.Resource, hist *map[string]*fieldHistory, info *map[string]*SchemaInfo,
) (bool, bool) {
if r == nil {
return hist != nil, info != nil
}
m := r.Schema()
if m == nil {
return hist != nil, info != nil
}
var rHasH, rHasI bool
m.Range(func(k string, v shim.Schema) bool {
h, hasH := getNonNil(hist, k)
i, hasI := getNonNil(info, k)
fieldHasHist, fieldHasInfo := applyMaxItemsOneAliasing(v, h, i)
hasH = hasH || fieldHasHist
hasI = hasI || fieldHasInfo
if !hasH {
delete(*hist, k)
}
if !hasI {
delete(*info, k)
}
rHasH = rHasH || hasH
rHasI = rHasI || hasI
return true
})
return rHasH, rHasI
}
// When walking the schema tree for a resource, we create mirroring trees in
// *fieldHistory and *SchemaInfo. To avoid polluting either tree (and
// interfering with other actions such as SetAutonaming), we clean up the paths
// that we created but did not store any information into.
//
// For example, consider the schema for a field of type `Object{ Key1:
// List[String] }`. The schema tree for this field looks like this:
//
// Object:
// Fields:
// Key1:
// List:
// Elem:
// String
//
// When we walk the tree, we create an almost identical history tree:
//
// Object:
// Fields:
// Key1:
// List:
// MaxItemsOne: false
// Elem:
// String
//
// We stored the additional piece of information `MaxItemsOne: false`. We need to
// keep enough of the tree to maintain that information, but no more. We can
// discard the unnecessary `Elem: String`.
//
// This keeps the tree as clean as possible for other processes which expect a
// `nil` element when making changes. Since other processes (like SetAutonaming)
| {
hist, ok, err := md.Get[aliasHistory](artifact, aliasMetadataKey)
if err != nil {
return aliasHistory{}, err
}
if !ok {
hist = aliasHistory{
Resources: map[string]*tokenHistory[tokens.Type]{},
DataSources: map[string]*tokenHistory[tokens.ModuleMember]{},
}
}
return hist, nil
} | identifier_body |
auto_aliasing.go | MaxItems changes are deferred until the next major version.
//
// Implementation note: to operate correctly this method needs to keep a persistent track
// of a database of past decision history. This is currently done by doing reads and
// writes to `providerInfo.GetMetadata()`, which is assumed to be persisted across
// provider releases. The bridge framework keeps this information written out to an opaque
// `bridge-metadata.json` blob which is expected to be stored in source control to persist
// across releases.
//
// Panics if [ProviderInfo.ApplyAutoAliases] would return an error.
func (info *ProviderInfo) MustApplyAutoAliases() {
err := info.ApplyAutoAliases()
contract.AssertNoErrorf(err, "Failed to apply aliases")
}
// Automatically applies backwards compatibility best practices.
//
// Specifically, [ApplyAutoAliases] may perform the following actions:
//
// - Call [ProviderInfo.RenameResourceWithAlias] or [ProviderInfo.RenameDataSource]
// - Edit [ResourceInfo.Aliases]
// - Edit [SchemaInfo.MaxItemsOne]
//
// The goal is to always maximize backwards compatibility and reduce breaking changes for
// the users of the Pulumi providers.
//
// Resource aliases help mask TF provider resource renames or changes in mapped tokens so
// older programs continue to work. See [ResourceInfo.RenameResourceWithAlias] and
// [ResourceInfo.Aliases] for more details.
//
// [SchemaInfo.MaxItemsOne] changes are also important because they involve flattening and
// pluralizing names. Collections (lists or sets) marked with MaxItems=1 are projected as
// scalar types in Pulumi SDKs. Therefore changes to the MaxItems property may be breaking
// the compilation of programs as the type changes from `T to List[T]` or vice versa. To
// avoid these breaking changes, this method undoes any upstream changes to MaxItems using
// [SchemaInfo.MaxItemsOne] overrides. This happens until a major version change is
// detected, and then overrides are cleared. Effectively this makes sure that upstream
// MaxItems changes are deferred until the next major version.
//
// Implementation note: to operate correctly this method needs to keep a persistent track
// of a database of past decision history. This is currently done by doing reads and
// writes to `providerInfo.GetMetadata()`, which is assumed to be persisted across
// provider releases. The bridge framework keeps this information written out to an opaque
// `bridge-metadata.json` blob which is expected to be stored in source control to persist
// across releas
func (info *ProviderInfo) ApplyAutoAliases() error {
artifact := info.GetMetadata()
hist, err := getHistory(artifact)
if err != nil {
return err
}
var currentVersion int
// If version is missing, we assume the current version is the most recent major
// version in mentioned in history.
if info.Version != "" {
v, err := semver.NewVersion(info.Version)
if err != nil {
return err
}
currentVersion = int(v.Major())
} else {
for _, r := range hist.Resources {
for _, p := range r.Past {
if p.MajorVersion > currentVersion {
currentVersion = p.MajorVersion
}
}
}
for _, d := range hist.DataSources {
for _, p := range d.Past {
if p.MajorVersion > currentVersion {
currentVersion = p.MajorVersion
}
}
}
}
rMap := info.P.ResourcesMap()
dMap := info.P.DataSourcesMap()
// Applying resource aliases adds new resources to providerInfo.Resources. To keep
// this process deterministic, we don't apply resource aliases until all resources
// have been examined.
//
// The same logic applies to datasources.
applyAliases := []func(){}
for tfToken, computed := range info.Resources {
r, _ := rMap.GetOk(tfToken)
aliasResource(info, r, &applyAliases, hist.Resources,
computed, tfToken, currentVersion)
}
for tfToken, computed := range info.DataSources {
ds, _ := dMap.GetOk(tfToken)
aliasDataSource(info, ds, &applyAliases, hist.DataSources,
computed, tfToken, currentVersion)
}
for _, f := range applyAliases {
f()
}
if err := md.Set(artifact, aliasMetadataKey, hist); err != nil {
// Set fails only when `hist` is not serializable. Because `hist` is
// composed of marshallable, non-cyclic types, this is impossible.
contract.AssertNoErrorf(err, "History failed to serialize")
}
return nil
}
const aliasMetadataKey = "auto-aliasing"
func getHistory(artifact ProviderMetadata) (aliasHistory, error) {
hist, ok, err := md.Get[aliasHistory](artifact, aliasMetadataKey)
if err != nil {
return aliasHistory{}, err
}
if !ok {
hist = aliasHistory{
Resources: map[string]*tokenHistory[tokens.Type]{},
DataSources: map[string]*tokenHistory[tokens.ModuleMember]{},
}
}
return hist, nil
}
func aliasResource(
p *ProviderInfo, res shim.Resource,
applyResourceAliases *[]func(),
hist map[string]*tokenHistory[tokens.Type], computed *ResourceInfo,
tfToken string, version int,
) {
prev, hasPrev := hist[tfToken]
if !hasPrev {
// It's not in the history, so it must be new. Stick it in the history for
// next time.
hist[tfToken] = &tokenHistory[tokens.Type]{
Current: computed.Tok,
}
} else {
// We don't do this eagerly because aliasResource is called while
// iterating over p.Resources which aliasOrRenameResource mutates.
*applyResourceAliases = append(*applyResourceAliases,
func() { aliasOrRenameResource(p, computed, tfToken, prev, version) })
}
// Apply Aliasing to MaxItemOne by traversing the field tree and applying the
// stored value.
//
// Note: If the user explicitly sets a MaxItemOne value, that value is respected
// and overwrites the current history.'
if res == nil {
return
}
// If we are behind the major version, reset the fields and the major version.
if hist[tfToken].MajorVersion < version {
hist[tfToken].MajorVersion = version
hist[tfToken].Fields = nil
}
applyResourceMaxItemsOneAliasing(res, &hist[tfToken].Fields, &computed.Fields)
}
// applyResourceMaxItemsOneAliasing traverses a shim.Resource, applying walk to each field in the resource.
func applyResourceMaxItemsOneAliasing(
r shim.Resource, hist *map[string]*fieldHistory, info *map[string]*SchemaInfo,
) (bool, bool) {
if r == nil {
return hist != nil, info != nil
}
m := r.Schema()
if m == nil {
return hist != nil, info != nil
}
var rHasH, rHasI bool
m.Range(func(k string, v shim.Schema) bool {
h, hasH := getNonNil(hist, k)
i, hasI := getNonNil(info, k)
fieldHasHist, fieldHasInfo := applyMaxItemsOneAliasing(v, h, i)
hasH = hasH || fieldHasHist
hasI = hasI || fieldHasInfo
if !hasH {
delete(*hist, k)
}
if !hasI {
delete(*info, k)
}
rHasH = rHasH || hasH
rHasI = rHasI || hasI
return true
})
return rHasH, rHasI
}
// When walking the schema tree for a resource, we create mirroring trees in
// *fieldHistory and *SchemaInfo. To avoid polluting either tree (and
// interfering with other actions such as SetAutonaming), we clean up the paths
// that we created but did not store any information into.
//
// For example, consider the schema for a field of type `Object{ Key1:
// List[String] }`. The schema tree for this field looks like this:
//
// Object:
// Fields:
// Key1:
// List:
// Elem:
// String
//
// When we walk the tree, we create an almost identical history tree:
//
// Object:
// Fields:
// Key1:
// List:
// MaxItemsOne: false
// Elem:
// String
//
// We stored the additional piece of information `MaxItemsOne: false`. We need to
// keep enough of the tree to maintain that information, but no more. We can
// discard the unnecessary `Elem: String`.
//
// This keeps the tree as clean as possible for other processes which expect a
// `nil` element when making changes. Since other processes (like SetAutonaming)
// act on edge nodes (like our String), this allows us to inter-operate with them
// without interference.
//
// applyMaxItemsOneAliasing traverses a generic shim.Schema recursively, applying fieldHistory to
// SchemaInfo and vise versa as necessary to avoid breaking changes in the
// resulting sdk.
func | applyMaxItemsOneAliasing | identifier_name |
|
auto_aliasing.go | type changes from `T to List[T]` or vice versa. To
// avoid these breaking changes, this method undoes any upstream changes to MaxItems using
// [SchemaInfo.MaxItemsOne] overrides. This happens until a major version change is
// detected, and then overrides are cleared. Effectively this makes sure that upstream
// MaxItems changes are deferred until the next major version.
//
// Implementation note: to operate correctly this method needs to keep a persistent track
// of a database of past decision history. This is currently done by doing reads and
// writes to `providerInfo.GetMetadata()`, which is assumed to be persisted across
// provider releases. The bridge framework keeps this information written out to an opaque
// `bridge-metadata.json` blob which is expected to be stored in source control to persist
// across releases.
//
// Panics if [ProviderInfo.ApplyAutoAliases] would return an error.
func (info *ProviderInfo) MustApplyAutoAliases() {
err := info.ApplyAutoAliases()
contract.AssertNoErrorf(err, "Failed to apply aliases")
}
// Automatically applies backwards compatibility best practices.
//
// Specifically, [ApplyAutoAliases] may perform the following actions:
//
// - Call [ProviderInfo.RenameResourceWithAlias] or [ProviderInfo.RenameDataSource]
// - Edit [ResourceInfo.Aliases]
// - Edit [SchemaInfo.MaxItemsOne]
//
// The goal is to always maximize backwards compatibility and reduce breaking changes for
// the users of the Pulumi providers.
//
// Resource aliases help mask TF provider resource renames or changes in mapped tokens so
// older programs continue to work. See [ResourceInfo.RenameResourceWithAlias] and
// [ResourceInfo.Aliases] for more details.
//
// [SchemaInfo.MaxItemsOne] changes are also important because they involve flattening and
// pluralizing names. Collections (lists or sets) marked with MaxItems=1 are projected as
// scalar types in Pulumi SDKs. Therefore changes to the MaxItems property may be breaking
// the compilation of programs as the type changes from `T to List[T]` or vice versa. To
// avoid these breaking changes, this method undoes any upstream changes to MaxItems using
// [SchemaInfo.MaxItemsOne] overrides. This happens until a major version change is
// detected, and then overrides are cleared. Effectively this makes sure that upstream
// MaxItems changes are deferred until the next major version.
//
// Implementation note: to operate correctly this method needs to keep a persistent track
// of a database of past decision history. This is currently done by doing reads and
// writes to `providerInfo.GetMetadata()`, which is assumed to be persisted across
// provider releases. The bridge framework keeps this information written out to an opaque
// `bridge-metadata.json` blob which is expected to be stored in source control to persist
// across releas
func (info *ProviderInfo) ApplyAutoAliases() error {
artifact := info.GetMetadata()
hist, err := getHistory(artifact)
if err != nil {
return err
}
var currentVersion int
// If version is missing, we assume the current version is the most recent major
// version in mentioned in history.
if info.Version != "" {
v, err := semver.NewVersion(info.Version)
if err != nil {
return err
}
currentVersion = int(v.Major())
} else {
for _, r := range hist.Resources {
for _, p := range r.Past {
if p.MajorVersion > currentVersion {
currentVersion = p.MajorVersion
}
}
}
for _, d := range hist.DataSources {
for _, p := range d.Past {
if p.MajorVersion > currentVersion {
currentVersion = p.MajorVersion
}
}
}
}
rMap := info.P.ResourcesMap()
dMap := info.P.DataSourcesMap()
// Applying resource aliases adds new resources to providerInfo.Resources. To keep
// this process deterministic, we don't apply resource aliases until all resources
// have been examined.
//
// The same logic applies to datasources.
applyAliases := []func(){}
for tfToken, computed := range info.Resources {
r, _ := rMap.GetOk(tfToken)
aliasResource(info, r, &applyAliases, hist.Resources,
computed, tfToken, currentVersion)
}
for tfToken, computed := range info.DataSources {
ds, _ := dMap.GetOk(tfToken)
aliasDataSource(info, ds, &applyAliases, hist.DataSources,
computed, tfToken, currentVersion)
}
for _, f := range applyAliases {
f()
}
if err := md.Set(artifact, aliasMetadataKey, hist); err != nil {
// Set fails only when `hist` is not serializable. Because `hist` is
// composed of marshallable, non-cyclic types, this is impossible.
contract.AssertNoErrorf(err, "History failed to serialize")
}
return nil
}
const aliasMetadataKey = "auto-aliasing"
func getHistory(artifact ProviderMetadata) (aliasHistory, error) {
hist, ok, err := md.Get[aliasHistory](artifact, aliasMetadataKey)
if err != nil {
return aliasHistory{}, err
}
if !ok |
return hist, nil
}
func aliasResource(
p *ProviderInfo, res shim.Resource,
applyResourceAliases *[]func(),
hist map[string]*tokenHistory[tokens.Type], computed *ResourceInfo,
tfToken string, version int,
) {
prev, hasPrev := hist[tfToken]
if !hasPrev {
// It's not in the history, so it must be new. Stick it in the history for
// next time.
hist[tfToken] = &tokenHistory[tokens.Type]{
Current: computed.Tok,
}
} else {
// We don't do this eagerly because aliasResource is called while
// iterating over p.Resources which aliasOrRenameResource mutates.
*applyResourceAliases = append(*applyResourceAliases,
func() { aliasOrRenameResource(p, computed, tfToken, prev, version) })
}
// Apply Aliasing to MaxItemOne by traversing the field tree and applying the
// stored value.
//
// Note: If the user explicitly sets a MaxItemOne value, that value is respected
// and overwrites the current history.'
if res == nil {
return
}
// If we are behind the major version, reset the fields and the major version.
if hist[tfToken].MajorVersion < version {
hist[tfToken].MajorVersion = version
hist[tfToken].Fields = nil
}
applyResourceMaxItemsOneAliasing(res, &hist[tfToken].Fields, &computed.Fields)
}
// applyResourceMaxItemsOneAliasing traverses a shim.Resource, applying walk to each field in the resource.
func applyResourceMaxItemsOneAliasing(
r shim.Resource, hist *map[string]*fieldHistory, info *map[string]*SchemaInfo,
) (bool, bool) {
if r == nil {
return hist != nil, info != nil
}
m := r.Schema()
if m == nil {
return hist != nil, info != nil
}
var rHasH, rHasI bool
m.Range(func(k string, v shim.Schema) bool {
h, hasH := getNonNil(hist, k)
i, hasI := getNonNil(info, k)
fieldHasHist, fieldHasInfo := applyMaxItemsOneAliasing(v, h, i)
hasH = hasH || fieldHasHist
hasI = hasI || fieldHasInfo
if !hasH {
delete(*hist, k)
}
if !hasI {
delete(*info, k)
}
rHasH = rHasH || hasH
rHasI = rHasI || hasI
return true
})
return rHasH, rHasI
}
// When walking the schema tree for a resource, we create mirroring trees in
// *fieldHistory and *SchemaInfo. To avoid polluting either tree (and
// interfering with other actions such as SetAutonaming), we clean up the paths
// that we created but did not store any information into.
//
// For example, consider the schema for a field of type `Object{ Key1:
// List[String] }`. The schema tree for this field looks like this:
//
// Object:
// Fields:
// Key1:
// List:
// Elem:
// String
//
// When we walk the tree, we create an almost identical history tree:
//
// Object:
// Fields:
// Key1:
// List:
// MaxItemsOne: false
// Elem:
// String
//
// We stored the additional piece of information `MaxItemsOne: false`. We need to
// keep enough of the tree to maintain that information, but no more. We can
// discard the unnecessary `Elem: String`.
//
// This keeps the tree as clean as possible for other processes which expect a
// `nil` element when making changes. Since other processes (like SetAutonaming)
| {
hist = aliasHistory{
Resources: map[string]*tokenHistory[tokens.Type]{},
DataSources: map[string]*tokenHistory[tokens.ModuleMember]{},
}
} | conditional_block |
auto_aliasing.go | the type changes from `T to List[T]` or vice versa. To
// avoid these breaking changes, this method undoes any upstream changes to MaxItems using
// [SchemaInfo.MaxItemsOne] overrides. This happens until a major version change is
// detected, and then overrides are cleared. Effectively this makes sure that upstream
// MaxItems changes are deferred until the next major version.
//
// Implementation note: to operate correctly this method needs to keep a persistent track
// of a database of past decision history. This is currently done by doing reads and
// writes to `providerInfo.GetMetadata()`, which is assumed to be persisted across
// provider releases. The bridge framework keeps this information written out to an opaque
// `bridge-metadata.json` blob which is expected to be stored in source control to persist
// across releases.
//
// Panics if [ProviderInfo.ApplyAutoAliases] would return an error.
func (info *ProviderInfo) MustApplyAutoAliases() {
err := info.ApplyAutoAliases()
contract.AssertNoErrorf(err, "Failed to apply aliases")
}
// Automatically applies backwards compatibility best practices.
//
// Specifically, [ApplyAutoAliases] may perform the following actions:
//
// - Call [ProviderInfo.RenameResourceWithAlias] or [ProviderInfo.RenameDataSource]
// - Edit [ResourceInfo.Aliases]
// - Edit [SchemaInfo.MaxItemsOne]
//
// The goal is to always maximize backwards compatibility and reduce breaking changes for
// the users of the Pulumi providers.
//
// Resource aliases help mask TF provider resource renames or changes in mapped tokens so
// older programs continue to work. See [ResourceInfo.RenameResourceWithAlias] and
// [ResourceInfo.Aliases] for more details.
// | // the compilation of programs as the type changes from `T to List[T]` or vice versa. To
// avoid these breaking changes, this method undoes any upstream changes to MaxItems using
// [SchemaInfo.MaxItemsOne] overrides. This happens until a major version change is
// detected, and then overrides are cleared. Effectively this makes sure that upstream
// MaxItems changes are deferred until the next major version.
//
// Implementation note: to operate correctly this method needs to keep a persistent track
// of a database of past decision history. This is currently done by doing reads and
// writes to `providerInfo.GetMetadata()`, which is assumed to be persisted across
// provider releases. The bridge framework keeps this information written out to an opaque
// `bridge-metadata.json` blob which is expected to be stored in source control to persist
// across releas
func (info *ProviderInfo) ApplyAutoAliases() error {
artifact := info.GetMetadata()
hist, err := getHistory(artifact)
if err != nil {
return err
}
var currentVersion int
// If version is missing, we assume the current version is the most recent major
// version in mentioned in history.
if info.Version != "" {
v, err := semver.NewVersion(info.Version)
if err != nil {
return err
}
currentVersion = int(v.Major())
} else {
for _, r := range hist.Resources {
for _, p := range r.Past {
if p.MajorVersion > currentVersion {
currentVersion = p.MajorVersion
}
}
}
for _, d := range hist.DataSources {
for _, p := range d.Past {
if p.MajorVersion > currentVersion {
currentVersion = p.MajorVersion
}
}
}
}
rMap := info.P.ResourcesMap()
dMap := info.P.DataSourcesMap()
// Applying resource aliases adds new resources to providerInfo.Resources. To keep
// this process deterministic, we don't apply resource aliases until all resources
// have been examined.
//
// The same logic applies to datasources.
applyAliases := []func(){}
for tfToken, computed := range info.Resources {
r, _ := rMap.GetOk(tfToken)
aliasResource(info, r, &applyAliases, hist.Resources,
computed, tfToken, currentVersion)
}
for tfToken, computed := range info.DataSources {
ds, _ := dMap.GetOk(tfToken)
aliasDataSource(info, ds, &applyAliases, hist.DataSources,
computed, tfToken, currentVersion)
}
for _, f := range applyAliases {
f()
}
if err := md.Set(artifact, aliasMetadataKey, hist); err != nil {
// Set fails only when `hist` is not serializable. Because `hist` is
// composed of marshallable, non-cyclic types, this is impossible.
contract.AssertNoErrorf(err, "History failed to serialize")
}
return nil
}
const aliasMetadataKey = "auto-aliasing"
func getHistory(artifact ProviderMetadata) (aliasHistory, error) {
hist, ok, err := md.Get[aliasHistory](artifact, aliasMetadataKey)
if err != nil {
return aliasHistory{}, err
}
if !ok {
hist = aliasHistory{
Resources: map[string]*tokenHistory[tokens.Type]{},
DataSources: map[string]*tokenHistory[tokens.ModuleMember]{},
}
}
return hist, nil
}
func aliasResource(
p *ProviderInfo, res shim.Resource,
applyResourceAliases *[]func(),
hist map[string]*tokenHistory[tokens.Type], computed *ResourceInfo,
tfToken string, version int,
) {
prev, hasPrev := hist[tfToken]
if !hasPrev {
// It's not in the history, so it must be new. Stick it in the history for
// next time.
hist[tfToken] = &tokenHistory[tokens.Type]{
Current: computed.Tok,
}
} else {
// We don't do this eagerly because aliasResource is called while
// iterating over p.Resources which aliasOrRenameResource mutates.
*applyResourceAliases = append(*applyResourceAliases,
func() { aliasOrRenameResource(p, computed, tfToken, prev, version) })
}
// Apply Aliasing to MaxItemOne by traversing the field tree and applying the
// stored value.
//
// Note: If the user explicitly sets a MaxItemOne value, that value is respected
// and overwrites the current history.'
if res == nil {
return
}
// If we are behind the major version, reset the fields and the major version.
if hist[tfToken].MajorVersion < version {
hist[tfToken].MajorVersion = version
hist[tfToken].Fields = nil
}
applyResourceMaxItemsOneAliasing(res, &hist[tfToken].Fields, &computed.Fields)
}
// applyResourceMaxItemsOneAliasing traverses a shim.Resource, applying walk to each field in the resource.
func applyResourceMaxItemsOneAliasing(
r shim.Resource, hist *map[string]*fieldHistory, info *map[string]*SchemaInfo,
) (bool, bool) {
if r == nil {
return hist != nil, info != nil
}
m := r.Schema()
if m == nil {
return hist != nil, info != nil
}
var rHasH, rHasI bool
m.Range(func(k string, v shim.Schema) bool {
h, hasH := getNonNil(hist, k)
i, hasI := getNonNil(info, k)
fieldHasHist, fieldHasInfo := applyMaxItemsOneAliasing(v, h, i)
hasH = hasH || fieldHasHist
hasI = hasI || fieldHasInfo
if !hasH {
delete(*hist, k)
}
if !hasI {
delete(*info, k)
}
rHasH = rHasH || hasH
rHasI = rHasI || hasI
return true
})
return rHasH, rHasI
}
// When walking the schema tree for a resource, we create mirroring trees in
// *fieldHistory and *SchemaInfo. To avoid polluting either tree (and
// interfering with other actions such as SetAutonaming), we clean up the paths
// that we created but did not store any information into.
//
// For example, consider the schema for a field of type `Object{ Key1:
// List[String] }`. The schema tree for this field looks like this:
//
// Object:
// Fields:
// Key1:
// List:
// Elem:
// String
//
// When we walk the tree, we create an almost identical history tree:
//
// Object:
// Fields:
// Key1:
// List:
// MaxItemsOne: false
// Elem:
// String
//
// We stored the additional piece of information `MaxItemsOne: false`. We need to
// keep enough of the tree to maintain that information, but no more. We can
// discard the unnecessary `Elem: String`.
//
// This keeps the tree as clean as possible for other processes which expect a
// `nil` element when making changes. Since other processes (like SetAutonaming)
// act | // [SchemaInfo.MaxItemsOne] changes are also important because they involve flattening and
// pluralizing names. Collections (lists or sets) marked with MaxItems=1 are projected as
// scalar types in Pulumi SDKs. Therefore changes to the MaxItems property may be breaking | random_line_split |
cht.rs | <N>(cht_size: N, block_num: N) -> Option<N>
where
N: Clone + AtLeast32Bit,
{
let block_cht_num = block_to_cht_number(cht_size.clone(), block_num.clone())?;
let two = N::one() + N::one();
if block_cht_num < two {
return None
}
let cht_start = start_number(cht_size, block_cht_num.clone());
if cht_start != block_num {
return None
}
Some(block_cht_num - two)
}
/// Returns Some(max_cht_number) if CHT has ever been built given maximal canonical block number.
pub fn max_cht_number<N>(cht_size: N, max_canonical_block: N) -> Option<N>
where
N: Clone + AtLeast32Bit,
{
let max_cht_number = block_to_cht_number(cht_size, max_canonical_block)?;
let two = N::one() + N::one();
if max_cht_number < two {
return None
}
Some(max_cht_number - two)
}
/// Compute a CHT root from an iterator of block hashes. Fails if shorter than
/// SIZE items. The items are assumed to proceed sequentially from `start_number(cht_num)`.
/// Discards the trie's nodes.
pub fn compute_root<Header, Hasher, I>(
cht_size: Header::Number,
cht_num: Header::Number,
hashes: I,
) -> ClientResult<Hasher::Out>
where
Header: HeaderT,
Hasher: hash_db::Hasher,
Hasher::Out: Ord,
I: IntoIterator<Item = ClientResult<Option<Header::Hash>>>,
{
use sp_trie::TrieConfiguration;
Ok(sp_trie::trie_types::Layout::<Hasher>::trie_root(build_pairs::<Header, I>(
cht_size, cht_num, hashes,
)?))
}
/// Build CHT-based header proof.
pub fn build_proof<Header, Hasher, BlocksI, HashesI>(
cht_size: Header::Number,
cht_num: Header::Number,
blocks: BlocksI,
hashes: HashesI,
) -> ClientResult<StorageProof>
where
Header: HeaderT,
Hasher: hash_db::Hasher,
Hasher::Out: Ord + codec::Codec,
BlocksI: IntoIterator<Item = Header::Number>,
HashesI: IntoIterator<Item = ClientResult<Option<Header::Hash>>>,
{
let transaction = build_pairs::<Header, _>(cht_size, cht_num, hashes)?
.into_iter()
.map(|(k, v)| (k, Some(v)))
.collect::<Vec<_>>();
let mut storage = InMemoryBackend::<Hasher>::default().update(vec![(None, transaction)]);
let trie_storage =
storage.as_trie_backend().expect("InMemoryState::as_trie_backend always returns Some; qed");
prove_read_on_trie_backend(
trie_storage,
blocks.into_iter().map(|number| encode_cht_key(number)),
)
.map_err(ClientError::Execution)
}
/// Check CHT-based header proof.
pub fn check_proof<Header, Hasher>(
local_root: Header::Hash,
local_number: Header::Number,
remote_hash: Header::Hash,
remote_proof: StorageProof,
) -> ClientResult<()>
where
Header: HeaderT,
Hasher: hash_db::Hasher,
Hasher::Out: Ord + codec::Codec,
{
do_check_proof::<Header, Hasher, _>(
local_root,
local_number,
remote_hash,
move |local_root, local_cht_key| {
read_proof_check::<Hasher, _>(
local_root,
remote_proof,
::std::iter::once(local_cht_key),
)
.map(|mut map| map.remove(local_cht_key).expect("checked proof of local_cht_key; qed"))
.map_err(|e| ClientError::from(e))
},
)
}
/// Check CHT-based header proof on pre-created proving backend.
pub fn check_proof_on_proving_backend<Header, Hasher>(
local_root: Header::Hash,
local_number: Header::Number,
remote_hash: Header::Hash,
proving_backend: &TrieBackend<MemoryDB<Hasher>, Hasher>,
) -> ClientResult<()>
where
Header: HeaderT,
Hasher: hash_db::Hasher,
Hasher::Out: Ord + codec::Codec,
{
do_check_proof::<Header, Hasher, _>(
local_root,
local_number,
remote_hash,
|_, local_cht_key| {
read_proof_check_on_proving_backend::<Hasher>(proving_backend, local_cht_key)
.map_err(|e| ClientError::from(e))
},
)
}
/// Check CHT-based header proof using passed checker function.
fn do_check_proof<Header, Hasher, F>(
local_root: Header::Hash,
local_number: Header::Number,
remote_hash: Header::Hash,
checker: F,
) -> ClientResult<()>
where
Header: HeaderT,
Hasher: hash_db::Hasher,
Hasher::Out: Ord,
F: FnOnce(Hasher::Out, &[u8]) -> ClientResult<Option<Vec<u8>>>,
{
let root: Hasher::Out = convert_hash(&local_root);
let local_cht_key = encode_cht_key(local_number);
let local_cht_value = checker(root, &local_cht_key)?;
let local_cht_value = local_cht_value.ok_or_else(|| ClientError::InvalidCHTProof)?;
let local_hash =
decode_cht_value(&local_cht_value).ok_or_else(|| ClientError::InvalidCHTProof)?;
match &local_hash[..] == remote_hash.as_ref() {
true => Ok(()),
false => Err(ClientError::InvalidCHTProof.into()),
}
}
/// Group ordered blocks by CHT number and call functor with blocks of each group.
pub fn for_each_cht_group<Header, I, F, P>(
cht_size: Header::Number,
blocks: I,
mut functor: F,
mut functor_param: P,
) -> ClientResult<()>
where
Header: HeaderT,
I: IntoIterator<Item = Header::Number>,
F: FnMut(P, Header::Number, Vec<Header::Number>) -> ClientResult<P>,
{
let mut current_cht_num = None;
let mut current_cht_blocks = Vec::new();
for block in blocks {
let new_cht_num = match block_to_cht_number(cht_size, block) {
Some(new_cht_num) => new_cht_num,
None =>
return Err(ClientError::Backend(format!(
"Cannot compute CHT root for the block #{}",
block
))
.into()),
};
let advance_to_next_cht = current_cht_num.is_some() && current_cht_num != Some(new_cht_num);
if advance_to_next_cht {
let current_cht_num = current_cht_num.expect(
"advance_to_next_cht is true;
it is true only when current_cht_num is Some; qed",
);
assert!(
new_cht_num > current_cht_num,
"for_each_cht_group only supports ordered iterators"
);
functor_param =
functor(functor_param, current_cht_num, std::mem::take(&mut current_cht_blocks))?;
}
current_cht_blocks.push(block);
current_cht_num = Some(new_cht_num);
}
if let Some(current_cht_num) = current_cht_num {
functor(functor_param, current_cht_num, std::mem::take(&mut current_cht_blocks))?;
}
Ok(())
}
/// Build pairs for computing CHT.
fn build_pairs<Header, I>(
cht_size: Header::Number,
cht_num: Header::Number,
hashes: I,
) -> ClientResult<Vec<(Vec<u8>, Vec<u8>)>>
where
Header: HeaderT,
I: IntoIterator<Item = ClientResult<Option<Header::Hash>>>,
{
let start_num = start_number(cht_size, cht_num);
let mut pairs = Vec::new();
let mut hash_index = Header::Number::zero();
for hash in hashes.into_iter() {
let hash =
hash?.ok_or_else(|| ClientError::from(ClientError::MissingHashRequiredForCHT))?;
pairs.push((encode_cht_key(start_num + hash_index).to_vec(), encode_cht_value(hash)));
hash_index += Header::Number::one();
if hash_index == cht_size {
break
}
}
if hash_index == cht_size {
Ok(pairs)
} else {
Err(ClientError::MissingHashRequiredForCHT)
}
}
/// Get the starting block of a given CHT.
/// CHT 0 includes block 1...SIZE,
/// CHT 1 includes block SIZE + 1 ... 2*SIZE
/// More generally: CHT N includes block (1 + N*SIZE)...((N+1)*SIZE).
/// This is because the genesis hash is assumed to be known
/// and including it would be redundant.
pub fn start_number<N: AtLeast32Bit>(cht_size: N, cht_num: N) -> N {
(cht_num * cht_size) + N::one()
}
/// | is_build_required | identifier_name |
|
cht.rs |
/// Returns Some(cht_number) if CHT is need to be built when the block with given number is
/// canonized.
pub fn is_build_required<N>(cht_size: N, block_num: N) -> Option<N>
where
N: Clone + AtLeast32Bit,
{
let block_cht_num = block_to_cht_number(cht_size.clone(), block_num.clone())?;
let two = N::one() + N::one();
if block_cht_num < two {
return None
}
let cht_start = start_number(cht_size, block_cht_num.clone());
if cht_start != block_num {
return None
}
Some(block_cht_num - two)
}
/// Returns Some(max_cht_number) if CHT has ever been built given maximal canonical block number.
pub fn max_cht_number<N>(cht_size: N, max_canonical_block: N) -> Option<N>
where
N: Clone + AtLeast32Bit,
{
let max_cht_number = block_to_cht_number(cht_size, max_canonical_block)?;
let two = N::one() + N::one();
if max_cht_number < two {
return None
}
Some(max_cht_number - two)
}
/// Compute a CHT root from an iterator of block hashes. Fails if shorter than
/// SIZE items. The items are assumed to proceed sequentially from `start_number(cht_num)`.
/// Discards the trie's nodes.
pub fn compute_root<Header, Hasher, I>(
cht_size: Header::Number,
cht_num: Header::Number,
hashes: I,
) -> ClientResult<Hasher::Out>
where
Header: HeaderT,
Hasher: hash_db::Hasher,
Hasher::Out: Ord,
I: IntoIterator<Item = ClientResult<Option<Header::Hash>>>,
{
use sp_trie::TrieConfiguration;
Ok(sp_trie::trie_types::Layout::<Hasher>::trie_root(build_pairs::<Header, I>(
cht_size, cht_num, hashes,
)?))
}
/// Build CHT-based header proof.
pub fn build_proof<Header, Hasher, BlocksI, HashesI>(
cht_size: Header::Number,
cht_num: Header::Number,
blocks: BlocksI,
hashes: HashesI,
) -> ClientResult<StorageProof>
where
Header: HeaderT,
Hasher: hash_db::Hasher,
Hasher::Out: Ord + codec::Codec,
BlocksI: IntoIterator<Item = Header::Number>,
HashesI: IntoIterator<Item = ClientResult<Option<Header::Hash>>>,
{
let transaction = build_pairs::<Header, _>(cht_size, cht_num, hashes)?
.into_iter()
.map(|(k, v)| (k, Some(v)))
.collect::<Vec<_>>();
let mut storage = InMemoryBackend::<Hasher>::default().update(vec![(None, transaction)]);
let trie_storage =
storage.as_trie_backend().expect("InMemoryState::as_trie_backend always returns Some; qed");
prove_read_on_trie_backend(
trie_storage,
blocks.into_iter().map(|number| encode_cht_key(number)),
)
.map_err(ClientError::Execution)
}
/// Check CHT-based header proof.
pub fn check_proof<Header, Hasher>(
local_root: Header::Hash,
local_number: Header::Number,
remote_hash: Header::Hash,
remote_proof: StorageProof,
) -> ClientResult<()>
where
Header: HeaderT,
Hasher: hash_db::Hasher,
Hasher::Out: Ord + codec::Codec,
{
do_check_proof::<Header, Hasher, _>(
local_root,
local_number,
remote_hash,
move |local_root, local_cht_key| {
read_proof_check::<Hasher, _>(
local_root,
remote_proof,
::std::iter::once(local_cht_key),
)
.map(|mut map| map.remove(local_cht_key).expect("checked proof of local_cht_key; qed"))
.map_err(|e| ClientError::from(e))
},
)
}
/// Check CHT-based header proof on pre-created proving backend.
pub fn check_proof_on_proving_backend<Header, Hasher>(
local_root: Header::Hash,
local_number: Header::Number,
remote_hash: Header::Hash,
proving_backend: &TrieBackend<MemoryDB<Hasher>, Hasher>,
) -> ClientResult<()>
where
Header: HeaderT,
Hasher: hash_db::Hasher,
Hasher::Out: Ord + codec::Codec,
{
do_check_proof::<Header, Hasher, _>(
local_root,
local_number,
remote_hash,
|_, local_cht_key| {
read_proof_check_on_proving_backend::<Hasher>(proving_backend, local_cht_key)
.map_err(|e| ClientError::from(e))
},
)
}
/// Check CHT-based header proof using passed checker function.
fn do_check_proof<Header, Hasher, F>(
local_root: Header::Hash,
local_number: Header::Number,
remote_hash: Header::Hash,
checker: F,
) -> ClientResult<()>
where
Header: HeaderT,
Hasher: hash_db::Hasher,
Hasher::Out: Ord,
F: FnOnce(Hasher::Out, &[u8]) -> ClientResult<Option<Vec<u8>>>,
{
let root: Hasher::Out = convert_hash(&local_root);
let local_cht_key = encode_cht_key(local_number);
let local_cht_value = checker(root, &local_cht_key)?;
let local_cht_value = local_cht_value.ok_or_else(|| ClientError::InvalidCHTProof)?;
let local_hash =
decode_cht_value(&local_cht_value).ok_or_else(|| ClientError::InvalidCHTProof)?;
match &local_hash[..] == remote_hash.as_ref() {
true => Ok(()),
false => Err(ClientError::InvalidCHTProof.into()),
}
}
/// Group ordered blocks by CHT number and call functor with blocks of each group.
pub fn for_each_cht_group<Header, I, F, P>(
cht_size: Header::Number,
blocks: I,
mut functor: F,
mut functor_param: P,
) -> ClientResult<()>
where
Header: HeaderT,
I: IntoIterator<Item = Header::Number>,
F: FnMut(P, Header::Number, Vec<Header::Number>) -> ClientResult<P>,
{
let mut current_cht_num = None;
let mut current_cht_blocks = Vec::new();
for block in blocks {
let new_cht_num = match block_to_cht_number(cht_size, block) {
Some(new_cht_num) => new_cht_num,
None =>
return Err(ClientError::Backend(format!(
"Cannot compute CHT root for the block #{}",
block
))
.into()),
};
let advance_to_next_cht = current_cht_num.is_some() && current_cht_num != Some(new_cht_num);
if advance_to_next_cht {
let current_cht_num = current_cht_num.expect(
"advance_to_next_cht is true;
it is true only when current_cht_num is Some; qed",
);
assert!(
new_cht_num > current_cht_num,
"for_each_cht_group only supports ordered iterators"
);
functor_param =
functor(functor_param, current_cht_num, std::mem::take(&mut current_cht_blocks))?;
}
current_cht_blocks.push(block);
current_cht_num = Some(new_cht_num);
}
if let Some(current_cht_num) = current_cht_num {
functor(functor_param, current_cht_num, std::mem::take(&mut current_cht_blocks))?;
}
Ok(())
}
/// Build pairs for computing CHT.
fn build_pairs<Header, I>(
cht_size: Header::Number,
cht_num: Header::Number,
hashes: I,
) -> ClientResult<Vec<(Vec<u8>, Vec<u8>)>>
where
Header: HeaderT,
I: IntoIterator<Item = ClientResult<Option<Header::Hash>>>,
{
let start_num = start_number(cht_size, cht_num);
let mut pairs = Vec::new();
let mut hash_index = Header::Number::zero();
for hash in hashes.into_iter() {
let hash =
hash?.ok_or_else(|| ClientError::from(ClientError::MissingHashRequiredForCHT))?;
pairs.push((encode_cht_key(start_num + hash_index).to_vec(), encode_cht_value(hash)));
hash_index += Header::Number::one();
if hash_index == cht_size {
break
}
}
if hash_index == cht_size {
Ok(pairs)
} else {
Err(ClientError::MissingHashRequiredForCHT)
}
}
/// Get the starting block of a given CHT.
/// CHT 0 includes block 1...SIZE,
/// CHT 1 includes block SIZE + 1 ... 2*SIZE
/// More generally: CHT N includes block (1 + N*SIZE)...((N+1)*SIZE).
/// This is because the genesis hash is assumed to be known
/// and including it would be redundant.
pub fn start_number<N: | {
SIZE.into()
} | identifier_body |
|
cht.rs |
let cht_start = start_number(cht_size, block_cht_num.clone());
if cht_start != block_num {
return None
}
Some(block_cht_num - two)
}
/// Returns Some(max_cht_number) if CHT has ever been built given maximal canonical block number.
pub fn max_cht_number<N>(cht_size: N, max_canonical_block: N) -> Option<N>
where
N: Clone + AtLeast32Bit,
{
let max_cht_number = block_to_cht_number(cht_size, max_canonical_block)?;
let two = N::one() + N::one();
if max_cht_number < two {
return None
}
Some(max_cht_number - two)
}
/// Compute a CHT root from an iterator of block hashes. Fails if shorter than
/// SIZE items. The items are assumed to proceed sequentially from `start_number(cht_num)`.
/// Discards the trie's nodes.
pub fn compute_root<Header, Hasher, I>(
cht_size: Header::Number,
cht_num: Header::Number,
hashes: I,
) -> ClientResult<Hasher::Out>
where
Header: HeaderT,
Hasher: hash_db::Hasher,
Hasher::Out: Ord,
I: IntoIterator<Item = ClientResult<Option<Header::Hash>>>,
{
use sp_trie::TrieConfiguration;
Ok(sp_trie::trie_types::Layout::<Hasher>::trie_root(build_pairs::<Header, I>(
cht_size, cht_num, hashes,
)?))
}
/// Build CHT-based header proof.
pub fn build_proof<Header, Hasher, BlocksI, HashesI>(
cht_size: Header::Number,
cht_num: Header::Number,
blocks: BlocksI,
hashes: HashesI,
) -> ClientResult<StorageProof>
where
Header: HeaderT,
Hasher: hash_db::Hasher,
Hasher::Out: Ord + codec::Codec,
BlocksI: IntoIterator<Item = Header::Number>,
HashesI: IntoIterator<Item = ClientResult<Option<Header::Hash>>>,
{
let transaction = build_pairs::<Header, _>(cht_size, cht_num, hashes)?
.into_iter()
.map(|(k, v)| (k, Some(v)))
.collect::<Vec<_>>();
let mut storage = InMemoryBackend::<Hasher>::default().update(vec![(None, transaction)]);
let trie_storage =
storage.as_trie_backend().expect("InMemoryState::as_trie_backend always returns Some; qed");
prove_read_on_trie_backend(
trie_storage,
blocks.into_iter().map(|number| encode_cht_key(number)),
)
.map_err(ClientError::Execution)
}
/// Check CHT-based header proof.
pub fn check_proof<Header, Hasher>(
local_root: Header::Hash,
local_number: Header::Number,
remote_hash: Header::Hash,
remote_proof: StorageProof,
) -> ClientResult<()>
where
Header: HeaderT,
Hasher: hash_db::Hasher,
Hasher::Out: Ord + codec::Codec,
{
do_check_proof::<Header, Hasher, _>(
local_root,
local_number,
remote_hash,
move |local_root, local_cht_key| {
read_proof_check::<Hasher, _>(
local_root,
remote_proof,
::std::iter::once(local_cht_key),
)
.map(|mut map| map.remove(local_cht_key).expect("checked proof of local_cht_key; qed"))
.map_err(|e| ClientError::from(e))
},
)
}
/// Check CHT-based header proof on pre-created proving backend.
pub fn check_proof_on_proving_backend<Header, Hasher>(
local_root: Header::Hash,
local_number: Header::Number,
remote_hash: Header::Hash,
proving_backend: &TrieBackend<MemoryDB<Hasher>, Hasher>,
) -> ClientResult<()>
where
Header: HeaderT,
Hasher: hash_db::Hasher,
Hasher::Out: Ord + codec::Codec,
{
do_check_proof::<Header, Hasher, _>(
local_root,
local_number,
remote_hash,
|_, local_cht_key| {
read_proof_check_on_proving_backend::<Hasher>(proving_backend, local_cht_key)
.map_err(|e| ClientError::from(e))
},
)
}
/// Check CHT-based header proof using passed checker function.
fn do_check_proof<Header, Hasher, F>(
local_root: Header::Hash,
local_number: Header::Number,
remote_hash: Header::Hash,
checker: F,
) -> ClientResult<()>
where
Header: HeaderT,
Hasher: hash_db::Hasher,
Hasher::Out: Ord,
F: FnOnce(Hasher::Out, &[u8]) -> ClientResult<Option<Vec<u8>>>,
{
let root: Hasher::Out = convert_hash(&local_root);
let local_cht_key = encode_cht_key(local_number);
let local_cht_value = checker(root, &local_cht_key)?;
let local_cht_value = local_cht_value.ok_or_else(|| ClientError::InvalidCHTProof)?;
let local_hash =
decode_cht_value(&local_cht_value).ok_or_else(|| ClientError::InvalidCHTProof)?;
match &local_hash[..] == remote_hash.as_ref() {
true => Ok(()),
false => Err(ClientError::InvalidCHTProof.into()),
}
}
/// Group ordered blocks by CHT number and call functor with blocks of each group.
pub fn for_each_cht_group<Header, I, F, P>(
cht_size: Header::Number,
blocks: I,
mut functor: F,
mut functor_param: P,
) -> ClientResult<()>
where
Header: HeaderT,
I: IntoIterator<Item = Header::Number>,
F: FnMut(P, Header::Number, Vec<Header::Number>) -> ClientResult<P>,
{
let mut current_cht_num = None;
let mut current_cht_blocks = Vec::new();
for block in blocks {
let new_cht_num = match block_to_cht_number(cht_size, block) {
Some(new_cht_num) => new_cht_num,
None =>
return Err(ClientError::Backend(format!(
"Cannot compute CHT root for the block #{}",
block
))
.into()),
};
let advance_to_next_cht = current_cht_num.is_some() && current_cht_num != Some(new_cht_num);
if advance_to_next_cht {
let current_cht_num = current_cht_num.expect(
"advance_to_next_cht is true;
it is true only when current_cht_num is Some; qed",
);
assert!(
new_cht_num > current_cht_num,
"for_each_cht_group only supports ordered iterators"
);
functor_param =
functor(functor_param, current_cht_num, std::mem::take(&mut current_cht_blocks))?;
}
current_cht_blocks.push(block);
current_cht_num = Some(new_cht_num);
}
if let Some(current_cht_num) = current_cht_num {
functor(functor_param, current_cht_num, std::mem::take(&mut current_cht_blocks))?;
}
Ok(())
}
/// Build pairs for computing CHT.
fn build_pairs<Header, I>(
cht_size: Header::Number,
cht_num: Header::Number,
hashes: I,
) -> ClientResult<Vec<(Vec<u8>, Vec<u8>)>>
where
Header: HeaderT,
I: IntoIterator<Item = ClientResult<Option<Header::Hash>>>,
{
let start_num = start_number(cht_size, cht_num);
let mut pairs = Vec::new();
let mut hash_index = Header::Number::zero();
for hash in hashes.into_iter() {
let hash =
hash?.ok_or_else(|| ClientError::from(ClientError::MissingHashRequiredForCHT))?;
pairs.push((encode_cht_key(start_num + hash_index).to_vec(), encode_cht_value(hash)));
hash_index += Header::Number::one();
if hash_index == cht_size {
break
}
}
if hash_index == cht_size {
Ok(pairs)
} else {
Err(ClientError::MissingHashRequiredForCHT)
}
}
/// Get the starting block of a given CHT.
/// CHT 0 includes block 1...SIZE,
/// CHT 1 includes block SIZE + 1 ... 2*SIZE
/// More generally: CHT N includes block (1 + N*SIZE)...((N+1)*SIZE).
/// This is because the genesis hash is assumed to be known
/// and including it would be redundant.
pub fn start_number<N: AtLeast32Bit>(cht_size: N, cht_num: N) -> N {
(cht_num * cht_size) + N::one()
}
/// Get the ending block of a given CHT.
pub fn end_number<N: AtLeast32Bit>(cht_size: N, cht_num: N) -> N {
(cht_num + N::one()) * cht_size
}
/// Convert a block number to a CHT number.
/// Returns `None` for `block_num | {
return None
} | conditional_block |
|
cht.rs | 32Bit,
{
let max_cht_number = block_to_cht_number(cht_size, max_canonical_block)?;
let two = N::one() + N::one();
if max_cht_number < two {
return None
}
Some(max_cht_number - two)
}
| /// SIZE items. The items are assumed to proceed sequentially from `start_number(cht_num)`.
/// Discards the trie's nodes.
pub fn compute_root<Header, Hasher, I>(
cht_size: Header::Number,
cht_num: Header::Number,
hashes: I,
) -> ClientResult<Hasher::Out>
where
Header: HeaderT,
Hasher: hash_db::Hasher,
Hasher::Out: Ord,
I: IntoIterator<Item = ClientResult<Option<Header::Hash>>>,
{
use sp_trie::TrieConfiguration;
Ok(sp_trie::trie_types::Layout::<Hasher>::trie_root(build_pairs::<Header, I>(
cht_size, cht_num, hashes,
)?))
}
/// Build CHT-based header proof.
pub fn build_proof<Header, Hasher, BlocksI, HashesI>(
cht_size: Header::Number,
cht_num: Header::Number,
blocks: BlocksI,
hashes: HashesI,
) -> ClientResult<StorageProof>
where
Header: HeaderT,
Hasher: hash_db::Hasher,
Hasher::Out: Ord + codec::Codec,
BlocksI: IntoIterator<Item = Header::Number>,
HashesI: IntoIterator<Item = ClientResult<Option<Header::Hash>>>,
{
let transaction = build_pairs::<Header, _>(cht_size, cht_num, hashes)?
.into_iter()
.map(|(k, v)| (k, Some(v)))
.collect::<Vec<_>>();
let mut storage = InMemoryBackend::<Hasher>::default().update(vec![(None, transaction)]);
let trie_storage =
storage.as_trie_backend().expect("InMemoryState::as_trie_backend always returns Some; qed");
prove_read_on_trie_backend(
trie_storage,
blocks.into_iter().map(|number| encode_cht_key(number)),
)
.map_err(ClientError::Execution)
}
/// Check CHT-based header proof.
pub fn check_proof<Header, Hasher>(
local_root: Header::Hash,
local_number: Header::Number,
remote_hash: Header::Hash,
remote_proof: StorageProof,
) -> ClientResult<()>
where
Header: HeaderT,
Hasher: hash_db::Hasher,
Hasher::Out: Ord + codec::Codec,
{
do_check_proof::<Header, Hasher, _>(
local_root,
local_number,
remote_hash,
move |local_root, local_cht_key| {
read_proof_check::<Hasher, _>(
local_root,
remote_proof,
::std::iter::once(local_cht_key),
)
.map(|mut map| map.remove(local_cht_key).expect("checked proof of local_cht_key; qed"))
.map_err(|e| ClientError::from(e))
},
)
}
/// Check CHT-based header proof on pre-created proving backend.
pub fn check_proof_on_proving_backend<Header, Hasher>(
local_root: Header::Hash,
local_number: Header::Number,
remote_hash: Header::Hash,
proving_backend: &TrieBackend<MemoryDB<Hasher>, Hasher>,
) -> ClientResult<()>
where
Header: HeaderT,
Hasher: hash_db::Hasher,
Hasher::Out: Ord + codec::Codec,
{
do_check_proof::<Header, Hasher, _>(
local_root,
local_number,
remote_hash,
|_, local_cht_key| {
read_proof_check_on_proving_backend::<Hasher>(proving_backend, local_cht_key)
.map_err(|e| ClientError::from(e))
},
)
}
/// Check CHT-based header proof using passed checker function.
fn do_check_proof<Header, Hasher, F>(
local_root: Header::Hash,
local_number: Header::Number,
remote_hash: Header::Hash,
checker: F,
) -> ClientResult<()>
where
Header: HeaderT,
Hasher: hash_db::Hasher,
Hasher::Out: Ord,
F: FnOnce(Hasher::Out, &[u8]) -> ClientResult<Option<Vec<u8>>>,
{
let root: Hasher::Out = convert_hash(&local_root);
let local_cht_key = encode_cht_key(local_number);
let local_cht_value = checker(root, &local_cht_key)?;
let local_cht_value = local_cht_value.ok_or_else(|| ClientError::InvalidCHTProof)?;
let local_hash =
decode_cht_value(&local_cht_value).ok_or_else(|| ClientError::InvalidCHTProof)?;
match &local_hash[..] == remote_hash.as_ref() {
true => Ok(()),
false => Err(ClientError::InvalidCHTProof.into()),
}
}
/// Group ordered blocks by CHT number and call functor with blocks of each group.
pub fn for_each_cht_group<Header, I, F, P>(
cht_size: Header::Number,
blocks: I,
mut functor: F,
mut functor_param: P,
) -> ClientResult<()>
where
Header: HeaderT,
I: IntoIterator<Item = Header::Number>,
F: FnMut(P, Header::Number, Vec<Header::Number>) -> ClientResult<P>,
{
let mut current_cht_num = None;
let mut current_cht_blocks = Vec::new();
for block in blocks {
let new_cht_num = match block_to_cht_number(cht_size, block) {
Some(new_cht_num) => new_cht_num,
None =>
return Err(ClientError::Backend(format!(
"Cannot compute CHT root for the block #{}",
block
))
.into()),
};
let advance_to_next_cht = current_cht_num.is_some() && current_cht_num != Some(new_cht_num);
if advance_to_next_cht {
let current_cht_num = current_cht_num.expect(
"advance_to_next_cht is true;
it is true only when current_cht_num is Some; qed",
);
assert!(
new_cht_num > current_cht_num,
"for_each_cht_group only supports ordered iterators"
);
functor_param =
functor(functor_param, current_cht_num, std::mem::take(&mut current_cht_blocks))?;
}
current_cht_blocks.push(block);
current_cht_num = Some(new_cht_num);
}
if let Some(current_cht_num) = current_cht_num {
functor(functor_param, current_cht_num, std::mem::take(&mut current_cht_blocks))?;
}
Ok(())
}
/// Build pairs for computing CHT.
fn build_pairs<Header, I>(
cht_size: Header::Number,
cht_num: Header::Number,
hashes: I,
) -> ClientResult<Vec<(Vec<u8>, Vec<u8>)>>
where
Header: HeaderT,
I: IntoIterator<Item = ClientResult<Option<Header::Hash>>>,
{
let start_num = start_number(cht_size, cht_num);
let mut pairs = Vec::new();
let mut hash_index = Header::Number::zero();
for hash in hashes.into_iter() {
let hash =
hash?.ok_or_else(|| ClientError::from(ClientError::MissingHashRequiredForCHT))?;
pairs.push((encode_cht_key(start_num + hash_index).to_vec(), encode_cht_value(hash)));
hash_index += Header::Number::one();
if hash_index == cht_size {
break
}
}
if hash_index == cht_size {
Ok(pairs)
} else {
Err(ClientError::MissingHashRequiredForCHT)
}
}
/// Get the starting block of a given CHT.
/// CHT 0 includes block 1...SIZE,
/// CHT 1 includes block SIZE + 1 ... 2*SIZE
/// More generally: CHT N includes block (1 + N*SIZE)...((N+1)*SIZE).
/// This is because the genesis hash is assumed to be known
/// and including it would be redundant.
pub fn start_number<N: AtLeast32Bit>(cht_size: N, cht_num: N) -> N {
(cht_num * cht_size) + N::one()
}
/// Get the ending block of a given CHT.
pub fn end_number<N: AtLeast32Bit>(cht_size: N, cht_num: N) -> N {
(cht_num + N::one()) * cht_size
}
/// Convert a block number to a CHT number.
/// Returns `None` for `block_num` == 0, `Some` otherwise.
pub fn block_to_cht_number<N: AtLeast32Bit>(cht_size: N, block_num: N) -> Option<N> {
if block_num == N::zero() {
None
} else {
Some((block_num - N::one()) / cht_size)
}
}
/// Convert header number into CHT key.
pub fn encode_cht_key<N: Encode>(number: N) -> Vec<u8> | /// Compute a CHT root from an iterator of block hashes. Fails if shorter than | random_line_split |
utils.go | err = getTableIndex(db, table, queryMaxRetry)
if err != nil {
return nil, errors.Trace(err)
}
if len(table.Columns) == 0 {
return nil, errors.Errorf("invalid table %s.%s", schema, name)
}
return table, nil
}
func getTableColumns(db *sql.DB, table *models.Table, maxRetry int) error {
if table.Schema == "" || table.Name == "" {
return errors.New("schema/table is empty")
}
query := fmt.Sprintf("SHOW COLUMNS FROM `%s`.`%s`", table.Schema, table.Name)
rows, err := querySQL(db, query, maxRetry)
if err != nil {
return errors.Trace(err)
}
defer rows.Close()
rowColumns, err := rows.Columns()
if err != nil {
return errors.Trace(err)
}
// Show an example.
/*
mysql> show columns from test.t;
+-------+---------+------+-----+---------+-------------------+
| Field | Type | Null | Key | Default | Extra |
+-------+---------+------+-----+---------+-------------------+
| a | int(11) | NO | PRI | NULL | |
| b | int(11) | NO | PRI | NULL | |
| c | int(11) | YES | MUL | NULL | |
| d | int(11) | YES | | NULL | |
| d | json | YES | | NULL | VIRTUAL GENERATED |
+-------+---------+------+-----+---------+-------------------+
*/
idx := 0
for rows.Next() {
data := make([]sql.RawBytes, len(rowColumns))
values := make([]interface{}, len(rowColumns))
for i := range values {
values[i] = &data[i]
}
err = rows.Scan(values...)
if err != nil {
return errors.Trace(err)
}
column := &models.Column{}
column.Idx = idx
column.Name = string(data[0])
column.Tp = string(data[1])
column.Key = string(data[3])
column.Extra = string(data[5])
bracketIdx := strings.Index(column.Tp, "(")
if bracketIdx > 0 {
column.SubTp = column.Tp[bracketIdx+1 : len(column.Tp)-1]
column.Tp = column.Tp[:bracketIdx]
}
if strings.ToLower(string(data[2])) == "no" {
column.NotNull = true
}
// Check whether column has unsigned flag.
if strings.Contains(strings.ToLower(string(data[1])), "unsigned") {
column.Unsigned = true
}
table.Columns = append(table.Columns, column)
idx++
}
if rows.Err() != nil {
return errors.Trace(rows.Err())
}
return nil
}
func getTableIndex(db *sql.DB, table *models.Table, maxRetry int) error {
if table.Schema == "" || table.Name == "" {
return errors.New("schema/table is empty")
}
query := fmt.Sprintf("SHOW INDEX FROM `%s`.`%s`", table.Schema, table.Name)
rows, err := querySQL(db, query, maxRetry)
if err != nil |
defer rows.Close()
rowColumns, err := rows.Columns()
if err != nil {
return errors.Trace(err)
}
// Show an example.
/*
mysql> show index from test.t;
+-------+------------+----------+--------------+-------------+-----------+-------------+----------+--------+------+------------+---------+---------------+
| Table | Non_unique | Key_name | Seq_in_index | Column_name | Collation | Cardinality | Sub_part | Packed | Null | Index_type | Comment | Index_comment |
+-------+------------+----------+--------------+-------------+-----------+-------------+----------+--------+------+------------+---------+---------------+
| t | 0 | PRIMARY | 1 | a | A | 0 | NULL | NULL | | BTREE | | |
| t | 0 | PRIMARY | 2 | b | A | 0 | NULL | NULL | | BTREE | | |
| t | 0 | ucd | 1 | c | A | 0 | NULL | NULL | YES | BTREE | | |
| t | 0 | ucd | 2 | d | A | 0 | NULL | NULL | YES | BTREE | | |
+-------+------------+----------+--------------+-------------+-----------+-------------+----------+--------+------+------------+---------+---------------+
*/
var columns = make(map[string][]string)
for rows.Next() {
data := make([]sql.RawBytes, len(rowColumns))
values := make([]interface{}, len(rowColumns))
for i := range values {
values[i] = &data[i]
}
err = rows.Scan(values...)
if err != nil {
return errors.Trace(err)
}
nonUnique := string(data[1])
if nonUnique == "0" {
keyName := strings.ToLower(string(data[2]))
columns[keyName] = append(columns[keyName], string(data[4]))
}
}
if rows.Err() != nil {
return errors.Trace(rows.Err())
}
table.IndexColumns = findColumns(table.Columns, columns)
return nil
}
func findColumn(columns []*models.Column, indexColumn string) *models.Column {
for _, column := range columns {
if column.Name == indexColumn {
return column
}
}
return nil
}
func findColumns(columns []*models.Column, indexColumns map[string][]string) map[string][]*models.Column {
result := make(map[string][]*models.Column)
for keyName, indexCols := range indexColumns {
cols := make([]*models.Column, 0, len(indexCols))
for _, name := range indexCols {
column := findColumn(columns, name)
if column != nil {
cols = append(cols, column)
}
}
result[keyName] = cols
}
return result
}
func findTables(db *sql.DB, schema string) ([]string, error) {
query := fmt.Sprintf("SHOW TABLES FROM `%s`", schema)
rows, err := querySQL(db, query, queryMaxRetry)
if err != nil {
return nil, errors.Trace(err)
}
defer rows.Close()
tables := make([]string, 0)
for rows.Next() {
var table string
err = rows.Scan(&table)
if err != nil {
return nil, errors.Trace(err)
}
tables = append(tables, table)
}
if rows.Err() != nil {
return nil, errors.Trace(rows.Err())
}
return tables, nil
}
func getMaxID(db *sql.DB, schema, table string) (int64, error) {
stmt := fmt.Sprintf("SELECT IFNULL(max(id), 0) FROM `%s`.`%s`", schema, table)
rows, err := db.Query(stmt)
if err != nil {
return 0, errors.Trace(err)
}
defer rows.Close()
var id int64
for rows.Next() {
if err := rows.Scan(&id); err != nil {
return 0, errors.Trace(err)
}
}
return id, nil
}
func getRandID(db *sql.DB, schema, table string) (int, error) {
stmt := fmt.Sprintf("SELECT IFNULL(id, 0) FROM `%s`.`%s` ORDER BY RAND() LIMIT 1", schema, table)
rows, err := db.Query(stmt)
if err != nil {
return 0, errors.Trace(err)
}
defer rows.Close()
var id int
for rows.Next() {
if err := rows.Scan(&id); err != nil {
return 0, errors.Trace(err)
}
}
return id, nil
}
func genRandomValue(column *models.Column) (interface{}, error) {
booleans := []string{"TRUE", "FALSE"}
upper := strings.ToUpper(column.Tp)
var value interface{}
switch upper {
case "TINYINT":
value = rand.Intn(tinyIntMax)
case "SMALLINT":
value = rand.Intn(smallIntMax)
case "INT":
value = rand.Int31()
case "INTUNSIGNED":
value = rand.Int31()
case "BOOLEAN":
value = booleans[rand.Intn(len(booleans))]
case "BIGINT":
value = rand.Int63()
case "BIGINTUNSIGNED":
value = rand.Int63()
case "FLOAT":
value = rand.Float32() * math.MaxFloat32
case "DOUBLE":
value = rand.ExpFloat64()
case "DOUBLEUNSIGNED":
value = rand.ExpFloat64()
case "DECIMAL":
value = strconv.FormatFloat(rand.ExpFloat64(), 'f', 5, 64)
case "DATETIME", "TIMESTAMP", "TIMESTAMPONUPDATE":
t := genRandomTime()
value = fmt.Sprintf("%.4d-%.2d-%.2d %.2d:%.2d | {
return errors.Trace(err)
} | conditional_block |
utils.go |
func querySQL(db *sql.DB, query string, maxRetry int) (*sql.Rows, error) {
// TODO: add retry mechanism
rows, err := db.Query(query)
if err != nil {
return nil, errors.Trace(err)
}
return rows, nil
}
func getTableFromDB(db *sql.DB, schema string, name string) (*models.Table, error) {
table := &models.Table{}
table.Schema = schema
table.Name = name
table.IndexColumns = make(map[string][]*models.Column)
err := getTableColumns(db, table, queryMaxRetry)
if err != nil {
return nil, errors.Trace(err)
}
err = getTableIndex(db, table, queryMaxRetry)
if err != nil {
return nil, errors.Trace(err)
}
if len(table.Columns) == 0 {
return nil, errors.Errorf("invalid table %s.%s", schema, name)
}
return table, nil
}
func getTableColumns(db *sql.DB, table *models.Table, maxRetry int) error {
if table.Schema == "" || table.Name == "" {
return errors.New("schema/table is empty")
}
query := fmt.Sprintf("SHOW COLUMNS FROM `%s`.`%s`", table.Schema, table.Name)
rows, err := querySQL(db, query, maxRetry)
if err != nil {
return errors.Trace(err)
}
defer rows.Close()
rowColumns, err := rows.Columns()
if err != nil {
return errors.Trace(err)
}
// Show an example.
/*
mysql> show columns from test.t;
+-------+---------+------+-----+---------+-------------------+
| Field | Type | Null | Key | Default | Extra |
+-------+---------+------+-----+---------+-------------------+
| a | int(11) | NO | PRI | NULL | |
| b | int(11) | NO | PRI | NULL | |
| c | int(11) | YES | MUL | NULL | |
| d | int(11) | YES | | NULL | |
| d | json | YES | | NULL | VIRTUAL GENERATED |
+-------+---------+------+-----+---------+-------------------+
*/
idx := 0
for rows.Next() {
data := make([]sql.RawBytes, len(rowColumns))
values := make([]interface{}, len(rowColumns))
for i := range values {
values[i] = &data[i]
}
err = rows.Scan(values...)
if err != nil {
return errors.Trace(err)
}
column := &models.Column{}
column.Idx = idx
column.Name = string(data[0])
column.Tp = string(data[1])
column.Key = string(data[3])
column.Extra = string(data[5])
bracketIdx := strings.Index(column.Tp, "(")
if bracketIdx > 0 {
column.SubTp = column.Tp[bracketIdx+1 : len(column.Tp)-1]
column.Tp = column.Tp[:bracketIdx]
}
if strings.ToLower(string(data[2])) == "no" {
column.NotNull = true
}
// Check whether column has unsigned flag.
if strings.Contains(strings.ToLower(string(data[1])), "unsigned") {
column.Unsigned = true
}
table.Columns = append(table.Columns, column)
idx++
}
if rows.Err() != nil {
return errors.Trace(rows.Err())
}
return nil
}
func getTableIndex(db *sql.DB, table *models.Table, maxRetry int) error {
if table.Schema == "" || table.Name == "" {
return errors.New("schema/table is empty")
}
query := fmt.Sprintf("SHOW INDEX FROM `%s`.`%s`", table.Schema, table.Name)
rows, err := querySQL(db, query, maxRetry)
if err != nil {
return errors.Trace(err)
}
defer rows.Close()
rowColumns, err := rows.Columns()
if err != nil {
return errors.Trace(err)
}
// Show an example.
/*
mysql> show index from test.t;
+-------+------------+----------+--------------+-------------+-----------+-------------+----------+--------+------+------------+---------+---------------+
| Table | Non_unique | Key_name | Seq_in_index | Column_name | Collation | Cardinality | Sub_part | Packed | Null | Index_type | Comment | Index_comment |
+-------+------------+----------+--------------+-------------+-----------+-------------+----------+--------+------+------------+---------+---------------+
| t | 0 | PRIMARY | 1 | a | A | 0 | NULL | NULL | | BTREE | | |
| t | 0 | PRIMARY | 2 | b | A | 0 | NULL | NULL | | BTREE | | |
| t | 0 | ucd | 1 | c | A | 0 | NULL | NULL | YES | BTREE | | |
| t | 0 | ucd | 2 | d | A | 0 | NULL | NULL | YES | BTREE | | |
+-------+------------+----------+--------------+-------------+-----------+-------------+----------+--------+------+------------+---------+---------------+
*/
var columns = make(map[string][]string)
for rows.Next() {
data := make([]sql.RawBytes, len(rowColumns))
values := make([]interface{}, len(rowColumns))
for i := range values {
values[i] = &data[i]
}
err = rows.Scan(values...)
if err != nil {
return errors.Trace(err)
}
nonUnique := string(data[1])
if nonUnique == "0" {
keyName := strings.ToLower(string(data[2]))
columns[keyName] = append(columns[keyName], string(data[4]))
}
}
if rows.Err() != nil {
return errors.Trace(rows.Err())
}
table.IndexColumns = findColumns(table.Columns, columns)
return nil
}
func findColumn(columns []*models.Column, indexColumn string) *models.Column {
for _, column := range columns {
if column.Name == indexColumn {
return column
}
}
return nil
}
func findColumns(columns []*models.Column, indexColumns map[string][]string) map[string][]*models.Column {
result := make(map[string][]*models.Column)
for keyName, indexCols := range indexColumns {
cols := make([]*models.Column, 0, len(indexCols))
for _, name := range indexCols {
column := findColumn(columns, name)
if column != nil {
cols = append(cols, column)
}
}
result[keyName] = cols
}
return result
}
func findTables(db *sql.DB, schema string) ([]string, error) {
query := fmt.Sprintf("SHOW TABLES FROM `%s`", schema)
rows, err := querySQL(db, query, queryMaxRetry)
if err != nil {
return nil, errors.Trace(err)
}
defer rows.Close()
tables := make([]string, 0)
for rows.Next() {
var table string
err = rows.Scan(&table)
if err != nil {
return nil, errors.Trace(err)
}
tables = append(tables, table)
}
if rows.Err() != nil {
return nil, errors.Trace(rows.Err())
}
return tables, nil
}
func getMaxID(db *sql.DB, schema, table string) (int64, error) {
stmt := fmt.Sprintf("SELECT IFNULL(max(id), 0) FROM `%s`.`%s`", schema, table)
rows, err := db.Query(stmt)
if err != nil {
return 0, errors.Trace(err)
}
defer rows.Close()
var id int64
for rows.Next() {
if err := rows.Scan(&id); err != nil {
return 0, errors.Trace(err)
}
}
return id, nil
}
func getRandID(db *sql.DB, schema, table string) (int, error) {
stmt := fmt.Sprintf("SELECT IFNULL(id, 0) FROM `%s`.`%s` ORDER BY RAND() LIMIT 1", schema, table)
rows, err := db.Query(stmt)
if err != nil {
return 0, errors.Trace(err)
}
defer rows.Close()
var id int
for rows.Next() {
if err := rows.Scan(&id); err != nil {
return 0, errors.Trace(err)
}
}
return id, nil
}
func genRandomValue(column *models.Column) (interface{}, error) {
booleans := []string{"TRUE", "FALSE"}
upper := strings.ToUpper(column.Tp)
var value interface{}
switch upper {
case "TINYINT":
value = rand.Intn(tinyIntMax)
case "SMALLINT":
value = rand.Intn(smallIntMax)
case "INT":
value = rand.Int31()
case "INTUNSIGNED":
value = rand.Int31()
case "BOOLEAN":
value = booleans[rand.Intn | {
return strings.Replace(name, "`", "``", -1)
} | identifier_body |
|
utils.go | (name string) string {
return strings.Replace(name, "`", "``", -1)
}
func querySQL(db *sql.DB, query string, maxRetry int) (*sql.Rows, error) {
// TODO: add retry mechanism
rows, err := db.Query(query)
if err != nil {
return nil, errors.Trace(err)
}
return rows, nil
}
func getTableFromDB(db *sql.DB, schema string, name string) (*models.Table, error) {
table := &models.Table{}
table.Schema = schema
table.Name = name
table.IndexColumns = make(map[string][]*models.Column)
err := getTableColumns(db, table, queryMaxRetry)
if err != nil {
return nil, errors.Trace(err)
}
err = getTableIndex(db, table, queryMaxRetry)
if err != nil {
return nil, errors.Trace(err)
}
if len(table.Columns) == 0 {
return nil, errors.Errorf("invalid table %s.%s", schema, name)
}
return table, nil
}
func getTableColumns(db *sql.DB, table *models.Table, maxRetry int) error {
if table.Schema == "" || table.Name == "" {
return errors.New("schema/table is empty")
}
query := fmt.Sprintf("SHOW COLUMNS FROM `%s`.`%s`", table.Schema, table.Name)
rows, err := querySQL(db, query, maxRetry)
if err != nil {
return errors.Trace(err)
}
defer rows.Close()
rowColumns, err := rows.Columns()
if err != nil {
return errors.Trace(err)
}
// Show an example.
/*
mysql> show columns from test.t;
+-------+---------+------+-----+---------+-------------------+
| Field | Type | Null | Key | Default | Extra |
+-------+---------+------+-----+---------+-------------------+
| a | int(11) | NO | PRI | NULL | |
| b | int(11) | NO | PRI | NULL | |
| c | int(11) | YES | MUL | NULL | |
| d | int(11) | YES | | NULL | |
| d | json | YES | | NULL | VIRTUAL GENERATED |
+-------+---------+------+-----+---------+-------------------+
*/
idx := 0
for rows.Next() {
data := make([]sql.RawBytes, len(rowColumns))
values := make([]interface{}, len(rowColumns))
for i := range values {
values[i] = &data[i]
}
err = rows.Scan(values...)
if err != nil {
return errors.Trace(err)
}
column := &models.Column{}
column.Idx = idx
column.Name = string(data[0])
column.Tp = string(data[1])
column.Key = string(data[3])
column.Extra = string(data[5])
bracketIdx := strings.Index(column.Tp, "(")
if bracketIdx > 0 {
column.SubTp = column.Tp[bracketIdx+1 : len(column.Tp)-1]
column.Tp = column.Tp[:bracketIdx]
}
if strings.ToLower(string(data[2])) == "no" {
column.NotNull = true
}
// Check whether column has unsigned flag.
if strings.Contains(strings.ToLower(string(data[1])), "unsigned") {
column.Unsigned = true
}
table.Columns = append(table.Columns, column)
idx++
}
if rows.Err() != nil {
return errors.Trace(rows.Err())
}
return nil
}
func getTableIndex(db *sql.DB, table *models.Table, maxRetry int) error {
if table.Schema == "" || table.Name == "" {
return errors.New("schema/table is empty")
}
query := fmt.Sprintf("SHOW INDEX FROM `%s`.`%s`", table.Schema, table.Name)
rows, err := querySQL(db, query, maxRetry)
if err != nil {
return errors.Trace(err)
}
defer rows.Close()
rowColumns, err := rows.Columns()
if err != nil {
return errors.Trace(err)
}
// Show an example.
/*
mysql> show index from test.t;
+-------+------------+----------+--------------+-------------+-----------+-------------+----------+--------+------+------------+---------+---------------+
| Table | Non_unique | Key_name | Seq_in_index | Column_name | Collation | Cardinality | Sub_part | Packed | Null | Index_type | Comment | Index_comment |
+-------+------------+----------+--------------+-------------+-----------+-------------+----------+--------+------+------------+---------+---------------+
| t | 0 | PRIMARY | 1 | a | A | 0 | NULL | NULL | | BTREE | | |
| t | 0 | PRIMARY | 2 | b | A | 0 | NULL | NULL | | BTREE | | |
| t | 0 | ucd | 1 | c | A | 0 | NULL | NULL | YES | BTREE | | |
| t | 0 | ucd | 2 | d | A | 0 | NULL | NULL | YES | BTREE | | |
+-------+------------+----------+--------------+-------------+-----------+-------------+----------+--------+------+------------+---------+---------------+
*/
var columns = make(map[string][]string)
for rows.Next() {
data := make([]sql.RawBytes, len(rowColumns))
values := make([]interface{}, len(rowColumns))
for i := range values {
values[i] = &data[i]
}
err = rows.Scan(values...)
if err != nil {
return errors.Trace(err)
}
nonUnique := string(data[1])
if nonUnique == "0" {
keyName := strings.ToLower(string(data[2]))
columns[keyName] = append(columns[keyName], string(data[4]))
}
}
if rows.Err() != nil {
return errors.Trace(rows.Err())
}
table.IndexColumns = findColumns(table.Columns, columns)
return nil
}
func findColumn(columns []*models.Column, indexColumn string) *models.Column {
for _, column := range columns {
if column.Name == indexColumn {
return column
}
}
return nil
}
func findColumns(columns []*models.Column, indexColumns map[string][]string) map[string][]*models.Column {
result := make(map[string][]*models.Column)
for keyName, indexCols := range indexColumns {
cols := make([]*models.Column, 0, len(indexCols))
for _, name := range indexCols {
column := findColumn(columns, name)
if column != nil {
cols = append(cols, column)
}
}
result[keyName] = cols
}
return result
}
func findTables(db *sql.DB, schema string) ([]string, error) {
query := fmt.Sprintf("SHOW TABLES FROM `%s`", schema)
rows, err := querySQL(db, query, queryMaxRetry)
if err != nil {
return nil, errors.Trace(err)
}
defer rows.Close()
tables := make([]string, 0)
for rows.Next() {
var table string
err = rows.Scan(&table)
if err != nil {
return nil, errors.Trace(err)
}
tables = append(tables, table)
}
if rows.Err() != nil {
return nil, errors.Trace(rows.Err())
}
return tables, nil
}
func getMaxID(db *sql.DB, schema, table string) (int64, error) {
stmt := fmt.Sprintf("SELECT IFNULL(max(id), 0) FROM `%s`.`%s`", schema, table)
rows, err := db.Query(stmt)
if err != nil {
return 0, errors.Trace(err)
}
defer rows.Close()
var id int64
for rows.Next() {
if err := rows.Scan(&id); err != nil {
return 0, errors.Trace(err)
}
}
return id, nil
}
func getRandID(db *sql.DB, schema, table string) (int, error) {
stmt := fmt.Sprintf("SELECT IFNULL(id, 0) FROM `%s`.`%s` ORDER BY RAND() LIMIT 1", schema, table)
rows, err := db.Query(stmt)
if err != nil {
return 0, errors.Trace(err)
}
defer rows.Close()
var id int
for rows.Next() {
if err := rows.Scan(&id); err != nil {
return 0, errors.Trace(err)
}
}
return id, nil
}
func genRandomValue(column *models.Column) (interface{}, error) {
booleans := []string{"TRUE", "FALSE"}
upper := strings.ToUpper(column.Tp)
var value interface{}
switch upper {
case "TINYINT":
value = rand.Intn(tinyIntMax)
case "SMALLINT":
value = rand.Intn(smallIntMax)
case "INT":
value = rand.Int31()
case "INTUNSIGNED":
value = rand.Int31()
case "BOOLEAN":
value = | escapeName | identifier_name |
|
utils.go | err = getTableIndex(db, table, queryMaxRetry)
if err != nil {
return nil, errors.Trace(err)
}
if len(table.Columns) == 0 {
return nil, errors.Errorf("invalid table %s.%s", schema, name)
}
return table, nil
}
func getTableColumns(db *sql.DB, table *models.Table, maxRetry int) error {
if table.Schema == "" || table.Name == "" {
return errors.New("schema/table is empty")
}
query := fmt.Sprintf("SHOW COLUMNS FROM `%s`.`%s`", table.Schema, table.Name)
rows, err := querySQL(db, query, maxRetry)
if err != nil {
return errors.Trace(err)
} | if err != nil {
return errors.Trace(err)
}
// Show an example.
/*
mysql> show columns from test.t;
+-------+---------+------+-----+---------+-------------------+
| Field | Type | Null | Key | Default | Extra |
+-------+---------+------+-----+---------+-------------------+
| a | int(11) | NO | PRI | NULL | |
| b | int(11) | NO | PRI | NULL | |
| c | int(11) | YES | MUL | NULL | |
| d | int(11) | YES | | NULL | |
| d | json | YES | | NULL | VIRTUAL GENERATED |
+-------+---------+------+-----+---------+-------------------+
*/
idx := 0
for rows.Next() {
data := make([]sql.RawBytes, len(rowColumns))
values := make([]interface{}, len(rowColumns))
for i := range values {
values[i] = &data[i]
}
err = rows.Scan(values...)
if err != nil {
return errors.Trace(err)
}
column := &models.Column{}
column.Idx = idx
column.Name = string(data[0])
column.Tp = string(data[1])
column.Key = string(data[3])
column.Extra = string(data[5])
bracketIdx := strings.Index(column.Tp, "(")
if bracketIdx > 0 {
column.SubTp = column.Tp[bracketIdx+1 : len(column.Tp)-1]
column.Tp = column.Tp[:bracketIdx]
}
if strings.ToLower(string(data[2])) == "no" {
column.NotNull = true
}
// Check whether column has unsigned flag.
if strings.Contains(strings.ToLower(string(data[1])), "unsigned") {
column.Unsigned = true
}
table.Columns = append(table.Columns, column)
idx++
}
if rows.Err() != nil {
return errors.Trace(rows.Err())
}
return nil
}
func getTableIndex(db *sql.DB, table *models.Table, maxRetry int) error {
if table.Schema == "" || table.Name == "" {
return errors.New("schema/table is empty")
}
query := fmt.Sprintf("SHOW INDEX FROM `%s`.`%s`", table.Schema, table.Name)
rows, err := querySQL(db, query, maxRetry)
if err != nil {
return errors.Trace(err)
}
defer rows.Close()
rowColumns, err := rows.Columns()
if err != nil {
return errors.Trace(err)
}
// Show an example.
/*
mysql> show index from test.t;
+-------+------------+----------+--------------+-------------+-----------+-------------+----------+--------+------+------------+---------+---------------+
| Table | Non_unique | Key_name | Seq_in_index | Column_name | Collation | Cardinality | Sub_part | Packed | Null | Index_type | Comment | Index_comment |
+-------+------------+----------+--------------+-------------+-----------+-------------+----------+--------+------+------------+---------+---------------+
| t | 0 | PRIMARY | 1 | a | A | 0 | NULL | NULL | | BTREE | | |
| t | 0 | PRIMARY | 2 | b | A | 0 | NULL | NULL | | BTREE | | |
| t | 0 | ucd | 1 | c | A | 0 | NULL | NULL | YES | BTREE | | |
| t | 0 | ucd | 2 | d | A | 0 | NULL | NULL | YES | BTREE | | |
+-------+------------+----------+--------------+-------------+-----------+-------------+----------+--------+------+------------+---------+---------------+
*/
var columns = make(map[string][]string)
for rows.Next() {
data := make([]sql.RawBytes, len(rowColumns))
values := make([]interface{}, len(rowColumns))
for i := range values {
values[i] = &data[i]
}
err = rows.Scan(values...)
if err != nil {
return errors.Trace(err)
}
nonUnique := string(data[1])
if nonUnique == "0" {
keyName := strings.ToLower(string(data[2]))
columns[keyName] = append(columns[keyName], string(data[4]))
}
}
if rows.Err() != nil {
return errors.Trace(rows.Err())
}
table.IndexColumns = findColumns(table.Columns, columns)
return nil
}
func findColumn(columns []*models.Column, indexColumn string) *models.Column {
for _, column := range columns {
if column.Name == indexColumn {
return column
}
}
return nil
}
func findColumns(columns []*models.Column, indexColumns map[string][]string) map[string][]*models.Column {
result := make(map[string][]*models.Column)
for keyName, indexCols := range indexColumns {
cols := make([]*models.Column, 0, len(indexCols))
for _, name := range indexCols {
column := findColumn(columns, name)
if column != nil {
cols = append(cols, column)
}
}
result[keyName] = cols
}
return result
}
func findTables(db *sql.DB, schema string) ([]string, error) {
query := fmt.Sprintf("SHOW TABLES FROM `%s`", schema)
rows, err := querySQL(db, query, queryMaxRetry)
if err != nil {
return nil, errors.Trace(err)
}
defer rows.Close()
tables := make([]string, 0)
for rows.Next() {
var table string
err = rows.Scan(&table)
if err != nil {
return nil, errors.Trace(err)
}
tables = append(tables, table)
}
if rows.Err() != nil {
return nil, errors.Trace(rows.Err())
}
return tables, nil
}
func getMaxID(db *sql.DB, schema, table string) (int64, error) {
stmt := fmt.Sprintf("SELECT IFNULL(max(id), 0) FROM `%s`.`%s`", schema, table)
rows, err := db.Query(stmt)
if err != nil {
return 0, errors.Trace(err)
}
defer rows.Close()
var id int64
for rows.Next() {
if err := rows.Scan(&id); err != nil {
return 0, errors.Trace(err)
}
}
return id, nil
}
func getRandID(db *sql.DB, schema, table string) (int, error) {
stmt := fmt.Sprintf("SELECT IFNULL(id, 0) FROM `%s`.`%s` ORDER BY RAND() LIMIT 1", schema, table)
rows, err := db.Query(stmt)
if err != nil {
return 0, errors.Trace(err)
}
defer rows.Close()
var id int
for rows.Next() {
if err := rows.Scan(&id); err != nil {
return 0, errors.Trace(err)
}
}
return id, nil
}
func genRandomValue(column *models.Column) (interface{}, error) {
booleans := []string{"TRUE", "FALSE"}
upper := strings.ToUpper(column.Tp)
var value interface{}
switch upper {
case "TINYINT":
value = rand.Intn(tinyIntMax)
case "SMALLINT":
value = rand.Intn(smallIntMax)
case "INT":
value = rand.Int31()
case "INTUNSIGNED":
value = rand.Int31()
case "BOOLEAN":
value = booleans[rand.Intn(len(booleans))]
case "BIGINT":
value = rand.Int63()
case "BIGINTUNSIGNED":
value = rand.Int63()
case "FLOAT":
value = rand.Float32() * math.MaxFloat32
case "DOUBLE":
value = rand.ExpFloat64()
case "DOUBLEUNSIGNED":
value = rand.ExpFloat64()
case "DECIMAL":
value = strconv.FormatFloat(rand.ExpFloat64(), 'f', 5, 64)
case "DATETIME", "TIMESTAMP", "TIMESTAMPONUPDATE":
t := genRandomTime()
value = fmt.Sprintf("%.4d-%.2d-%.2d %.2d:%.2d:%. | defer rows.Close()
rowColumns, err := rows.Columns() | random_line_split |
stimuli_LSL.py | filled contours
center_depth = obj_depths[np.where(obj_depths<1023)].mean() # take mean of non-white pixels (grayscale value < 1023)
cv2.circle(im_rgb, (cX, cY), 3, (255, 255, 255), -1)
depth_cm = int(100/((-0.00307*(center_depth))+3.33)) # convert pixel value to depth in cm, measurement not that accurate so only keep int value
cv2.putText(im_rgb, str(depth_cm)+' '+'cm', (cX - 20, cY - 80), cv2.FONT_HERSHEY_SIMPLEX, 0.5, (255, 255, 0), 2)
print 'pixel value of depth: ', center_depth, 'depth in cm: ', depth_cm, 'at coordinates ', 'X: ', cX,'Y: ', cY
midcont.append(cont) # 3d list of contours
cv2.drawContours(im_rgb, midcont, -1, (255,0,0), 3) # draw valid contours
im_rgb_scaled = cv2.resize(im_rgb,None,fx=screenX/640,fy=screenY/480,interpolation = cv2.INTER_CUBIC) # 640,480 are original dimensions of image
cv2.imwrite('segmented.jpg', im_rgb_scaled)
# cv2.imshow('depth', res) # openCV conflicts with psychopy, if imshow() is called, the stimuli cannot be shown
# k = cv2.waitKey(0) & 0xFF
if len(center_coords) != 0: # show flickering stimuli if we have a target
'''stimuli'''
#initialise variables
# method 1
exp_frames = 90 # show image for twice this number of frames ( see flip() )
screen_refresh_rate = 60 # refresh rate of monitor being used
# method2
waitdur = 2 # wait time before showing stimuli, in seconds
trialdur = 3 # duration of stimuli, in seconds
tenHz = [1, 1, 1, -1, -1, -1, 1, 1, 1, -1, -1, -1, 1, 1, 1, -1, -1, -1, 1, 1, 1, -1, -1, -1, 1, 1, 1, -1, -1, -1, 1, 1, 1, -1, -1, -1, 1, 1, 1, -1, -1, -1,1, 1, 1, -1, -1, -1, 1, 1, 1, -1, -1, -1, 1, 1, 1, -1, -1, -1]
twelveHz = [1, 1, 1, -1, -1, 1, 1, 1, -1, -1, 1, 1, 1, -1, -1, 1, 1, 1, -1, -1, 1, 1, 1, -1, -1, 1, 1, 1, -1, -1, 1, 1, 1, -1, -1, 1, 1, 1, -1, -1, 1, 1, 1, -1, -1, 1, 1, 1, -1, -1, 1, 1, 1, -1, -1, 1, 1, 1, -1, -1]
fifteenHz = [1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1]
freqs = [tenHz, twelveHz, fifteenHz]
# format contours for mask attribute in grating (to give grating the shape of the target)
for mask in obj_masks:
mask[mask==0] = -1
mask[mask==255] = 1 # 1 is black
obj_masks = obj_masks[:3] # have a maximum of 3 stimuli in image (frequency decoder trained on 3 frequencies)
center_coords = center_coords[:3]
# display image and flicker contours
# create a window
mywin = visual.Window((screenX,screenY),monitor="testMonitor",fullscr=True,units="deg")
grating_list = []
phases = [10,12,15] # frequencies of stimuli in fps
i = 0 # initialise index to assign frequencies to stimuli
# create some stimuli
img = visual.ImageStim(mywin, 'segmented.jpg') # show gratigs on top of this image
for mask in obj_masks:
grating = visual.GratingStim(win=mywin, mask=mask, units='pix', size=(screenX, -screenY), pos=[0,0], sf=3, tex='sqr') # mask is image of dim(640, 480) centered in the middle of the frame
# grating_list.append([grating,phases[i]]) # list of gratings and associated frequencies # method 1
grating_list.append([grating, freqs[i]]) # method 2
i += 1
fixation = visual.GratingStim(mywin, tex=None, mask='gauss', sf=0, size=0,name='fixation', autoLog=False) # empty frame to alternate with grating for flickering stimulus
'''receive data from eeg kit to use for CNN'''
s = socket(AF_INET,SOCK_STREAM)
host = '10.245.233.148'
port = 12396
s.connect((host, port))
rec = s.recv(12)
# show targets for 2 second without flickering, to allow user to choose
if rec == 'startstimuli':
# # method 1
# for frameN in range(60):
# img.draw()
# event.clearEvents()
# mywin.flip()
# # draw the stimuli and update the window
# stream = s.recv(10)
# print stream
# for frameN in range(exp_frames):
# img.draw() # background image
# for obj in grating_list: # draw valid contours to flicker
# frame_rate = obj[1]
# if int(frameN/(screen_refresh_rate/(frame_rate*4)))%2 != 0:
# fixation.draw()
# if int(frameN/(screen_refresh_rate/(frame_rate*4)))%2 == 0:
# obj[0].draw()
# if len(event.getKeys())>0: # press any key to exit
# break
# event.clearEvents()
# mywin.flip() # syncronizes for loop with refresh rate of screen
# method 2
mywin.flip() # syncronises loop with screen refresh rate
for seconds in range(int(waitdur*60)): # draw just image with targets
img.draw()
mywin.flip()
stream = s.recv(10)
print stream
for seconds in range(int(trialdur)): # flicker targets
for frameN in range(len(freqs[0])): # 60 frames (1 sec)
img.draw()
for obj in grating_list:
frame_f0 = obj[1] # 10, 12 or 15 hz
if frame_f0[frameN] == 1 :
obj[0].draw()
if frame_f0[frameN] == -1 :
fixation.draw()
event.clearEvents()
mywin.flip()
mywin.flip(clearBuffer=True)
# cleanup
mywin.close()
# receive data from EEG kit
sample_tot = []
while(True):
| sample = s.recv(104)
if len(sample) == 104:
sample = np.fromstring(sample) # recieve array of 13 elements (last 5 are metadata)
sample = sample[:-5]
sample_tot.append(sample)
if len(sample) == 0:
break | conditional_block |
|
stimuli_LSL.py | # edge detection
contour_list = []
canny = cv2.Canny(res, lower_threshold, upper_threshold, apertureSize=sobel_size) # finds edges, but not always closed contours (needed for cv2.findContours)
dilated = cv2.dilate(canny, np.ones((3, 3))) # makes sure canny edges are closed for findContours
_, contours, hierarchy = cv2.findContours(dilated,cv2.RETR_TREE,cv2.CHAIN_APPROX_SIMPLE) # finds contours around inner and outer canny edges, use hierarchy to select only inner ones
# cv2.drawContours(res, contours, -1, (255,255,0), 1) # draw all contours - debugging
# in hierarchy, if first element is -1 then contour has no other contour on same hierarchy level (is inner one), choose only inner contours (insde of canny edge line)
ind = 0 # initialise index for hierarchy
for c in contours:
if hierarchy[0][ind][0] == -1:
contour_list.append(c)
ind += 1
# select valid contours from inner most one
for cont in contour_list:
area = int(cv2.contourArea(cont))
if (area > area_min) and (area < area_max): # only pick contours between specified values (to avoid contours aruond objects too big or too small)
# compute center of contour
M = cv2.moments(cont)
cX = int(M["m10"] / M["m00"])
cY = int(M["m01"] / M["m00"])
# calculate min depth of all valid contours for if statement
center_neighbourhood = depth[cY-10:cY+10,cX-10:cX+10]
center_depth = center_neighbourhood.mean()
if center_depth < 1020: # don't calculate center of contours in background (most probably noise)
center_coords.append([cX,cY]) # add to list of coordinates of centers of contours
# calculate avg depth value of valid contours excluding white pixels
each_obj = np.zeros((480,640)) # blank image
cv2.drawContours(each_obj, [cont], -1, 255, -1) # fill valid contours on blank image
obj_masks.append(each_obj) # use for mask in flickering image
pts = np.where(each_obj == 255) # find coordinates of pixels in filled contours
obj_depths = depth[pts] # get grayscale value of pixels in filled contours
center_depth = obj_depths[np.where(obj_depths<1023)].mean() # take mean of non-white pixels (grayscale value < 1023)
cv2.circle(im_rgb, (cX, cY), 3, (255, 255, 255), -1)
depth_cm = int(100/((-0.00307*(center_depth))+3.33)) # convert pixel value to depth in cm, measurement not that accurate so only keep int value
cv2.putText(im_rgb, str(depth_cm)+' '+'cm', (cX - 20, cY - 80), cv2.FONT_HERSHEY_SIMPLEX, 0.5, (255, 255, 0), 2)
print 'pixel value of depth: ', center_depth, 'depth in cm: ', depth_cm, 'at coordinates ', 'X: ', cX,'Y: ', cY
midcont.append(cont) # 3d list of contours
cv2.drawContours(im_rgb, midcont, -1, (255,0,0), 3) # draw valid contours
im_rgb_scaled = cv2.resize(im_rgb,None,fx=screenX/640,fy=screenY/480,interpolation = cv2.INTER_CUBIC) # 640,480 are original dimensions of image
cv2.imwrite('segmented.jpg', im_rgb_scaled)
# cv2.imshow('depth', res) # openCV conflicts with psychopy, if imshow() is called, the stimuli cannot be shown
# k = cv2.waitKey(0) & 0xFF
if len(center_coords) != 0: # show flickering stimuli if we have a target
'''stimuli'''
#initialise variables
# method 1
exp_frames = 90 # show image for twice this number of frames ( see flip() )
screen_refresh_rate = 60 # refresh rate of monitor being used
# method2
waitdur = 2 # wait time before showing stimuli, in seconds
trialdur = 3 # duration of stimuli, in seconds
tenHz = [1, 1, 1, -1, -1, -1, 1, 1, 1, -1, -1, -1, 1, 1, 1, -1, -1, -1, 1, 1, 1, -1, -1, -1, 1, 1, 1, -1, -1, -1, 1, 1, 1, -1, -1, -1, 1, 1, 1, -1, -1, -1,1, 1, 1, -1, -1, -1, 1, 1, 1, -1, -1, -1, 1, 1, 1, -1, -1, -1]
twelveHz = [1, 1, 1, -1, -1, 1, 1, 1, -1, -1, 1, 1, 1, -1, -1, 1, 1, 1, -1, -1, 1, 1, 1, -1, -1, 1, 1, 1, -1, -1, 1, 1, 1, -1, -1, 1, 1, 1, -1, -1, 1, 1, 1, -1, -1, 1, 1, 1, -1, -1, 1, 1, 1, -1, -1, 1, 1, 1, -1, -1]
fifteenHz = [1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1]
freqs = [tenHz, twelveHz, fifteenHz]
# format contours for mask attribute in grating (to give grating the shape of the target)
for mask in obj_masks:
mask[mask==0] = -1
mask[mask==255] = 1 # 1 is black
obj_masks = obj_masks[:3] # have a maximum of 3 stimuli in image (frequency decoder trained on 3 frequencies)
center_coords = center_coords[:3]
# display image and flicker contours
# create a window
mywin = visual.Window((screenX,screenY),monitor="testMonitor",fullscr=True,units="deg")
grating_list = []
phases = [10,12,15] # frequencies of stimuli in fps
i = 0 # initialise index to assign frequencies to stimuli
# create some stimuli
img = visual.ImageStim(mywin, 'segmented.jpg') # show gratigs on top of this image
for mask in obj_masks:
grating = visual.GratingStim(win=mywin, mask=mask, units='pix', size=(screenX, -screenY), pos=[0,0], sf=3, tex='sqr') # mask is image of dim(640, 480) centered in the middle of the frame
# grating_list.append([grating,phases[i]]) # list of gratings and associated frequencies # method 1
grating_list.append([grating, freqs[i]]) # method 2
i += 1
fixation = visual.GratingStim(mywin, tex=None, mask='gauss', sf=0, size=0,name='fixation', autoLog=False) # empty frame to alternate with grating for flickering stimulus
'''receive data from eeg kit to use for CNN'''
s = socket(AF_INET,SOCK_STREAM)
host = '10.245.233.148'
port = 12396
s.connect((host, port))
| rec = s.recv(12)
| random_line_split |
|
descriptor.pb.go | i++
i = encodeVarintDescriptor(dAtA, i, uint64(len(m.MediaType)))
i += copy(dAtA[i:], m.MediaType)
}
if len(m.Digest) > 0 {
dAtA[i] = 0x12
i++
i = encodeVarintDescriptor(dAtA, i, uint64(len(m.Digest)))
i += copy(dAtA[i:], m.Digest)
}
if m.Size_ != 0 {
dAtA[i] = 0x18
i++
i = encodeVarintDescriptor(dAtA, i, uint64(m.Size_))
}
return i, nil
}
func encodeVarintDescriptor(dAtA []byte, offset int, v uint64) int {
for v >= 1<<7 {
dAtA[offset] = uint8(v&0x7f | 0x80)
v >>= 7
offset++
}
dAtA[offset] = uint8(v)
return offset + 1
}
func (m *Descriptor) Size() (n int) |
func sovDescriptor(x uint64) (n int) {
for {
n++
x >>= 7
if x == 0 {
break
}
}
return n
}
func sozDescriptor(x uint64) (n int) {
return sovDescriptor(uint64((x << 1) ^ uint64((int64(x) >> 63))))
}
func (this *Descriptor) String() string {
if this == nil {
return "nil"
}
s := strings.Join([]string{`&Descriptor{`,
`MediaType:` + fmt.Sprintf("%v", this.MediaType) + `,`,
`Digest:` + fmt.Sprintf("%v", this.Digest) + `,`,
`Size_:` + fmt.Sprintf("%v", this.Size_) + `,`,
`}`,
}, "")
return s
}
func valueToStringDescriptor(v interface{}) string {
rv := reflect.ValueOf(v)
if rv.IsNil() {
return "nil"
}
pv := reflect.Indirect(rv).Interface()
return fmt.Sprintf("*%v", pv)
}
func (m *Descriptor) Unmarshal(dAtA []byte) error {
l := len(dAtA)
iNdEx := 0
for iNdEx < l {
preIndex := iNdEx
var wire uint64
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return ErrIntOverflowDescriptor
}
if iNdEx >= l {
return io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
wire |= (uint64(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
fieldNum := int32(wire >> 3)
wireType := int(wire & 0x7)
if wireType == 4 {
return fmt.Errorf("proto: Descriptor: wiretype end group for non-group")
}
if fieldNum <= 0 {
return fmt.Errorf("proto: Descriptor: illegal tag %d (wire type %d)", fieldNum, wire)
}
switch fieldNum {
case 1:
if wireType != 2 {
return fmt.Errorf("proto: wrong wireType = %d for field MediaType", wireType)
}
var stringLen uint64
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return ErrIntOverflowDescriptor
}
if iNdEx >= l {
return io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
stringLen |= (uint64(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
intStringLen := int(stringLen)
if intStringLen < 0 {
return ErrInvalidLengthDescriptor
}
postIndex := iNdEx + intStringLen
if postIndex > l {
return io.ErrUnexpectedEOF
}
m.MediaType = string(dAtA[iNdEx:postIndex])
iNdEx = postIndex
case 2:
if wireType != 2 {
return fmt.Errorf("proto: wrong wireType = %d for field Digest", wireType)
}
var stringLen uint64
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return ErrIntOverflowDescriptor
}
if iNdEx >= l {
return io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
stringLen |= (uint64(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
intStringLen := int(stringLen)
if intStringLen < 0 {
return ErrInvalidLengthDescriptor
}
postIndex := iNdEx + intStringLen
if postIndex > l {
return io.ErrUnexpectedEOF
}
m.Digest = github_com_opencontainers_go_digest.Digest(dAtA[iNdEx:postIndex])
iNdEx = postIndex
case 3:
if wireType != 0 {
return fmt.Errorf("proto: wrong wireType = %d for field Size_", wireType)
}
m.Size_ = 0
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return ErrIntOverflowDescriptor
}
if iNdEx >= l {
return io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
m.Size_ |= (int64(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
default:
iNdEx = preIndex
skippy, err := skipDescriptor(dAtA[iNdEx:])
if err != nil {
return err
}
if skippy < 0 {
return ErrInvalidLengthDescriptor
}
if (iNdEx + skippy) > l {
return io.ErrUnexpectedEOF
}
iNdEx += skippy
}
}
if iNdEx > l {
return io.ErrUnexpectedEOF
}
return nil
}
func skipDescriptor(dAtA []byte) (n int, err error) {
l := len(dAtA)
iNdEx := 0
for iNdEx < l {
var wire uint64
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return 0, ErrIntOverflowDescriptor
}
if iNdEx >= l {
return 0, io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
wire |= (uint64(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
wireType := int(wire & 0x7)
switch wireType {
case 0:
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return 0, ErrIntOverflowDescriptor
}
if iNdEx >= l {
return 0, io.ErrUnexpectedEOF
}
iNdEx++
if dAtA[iNdEx-1] < 0x80 {
break
}
}
return iNdEx, nil
case 1:
iNdEx += 8
return iNdEx, nil
case 2:
var length int
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return 0, ErrIntOverflowDescriptor
}
if iNdEx >= l {
return 0, io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
length |= (int(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
iNdEx += length
if length < 0 {
return 0, ErrInvalidLengthDescriptor
}
return iNdEx, nil
case 3:
for {
var innerWire uint64
var start int = iNdEx
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return 0, ErrIntOverflowDescriptor
}
if iNdEx >= l {
return 0, io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
innerWire |= (uint64(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
innerWireType := int(innerWire & 0x7)
if innerWireType == 4 {
break
}
next, err := | {
var l int
_ = l
l = len(m.MediaType)
if l > 0 {
n += 1 + l + sovDescriptor(uint64(l))
}
l = len(m.Digest)
if l > 0 {
n += 1 + l + sovDescriptor(uint64(l))
}
if m.Size_ != 0 {
n += 1 + sovDescriptor(uint64(m.Size_))
}
return n
} | identifier_body |
descriptor.pb.go |
return dAtA[:n], nil
}
func (m *Descriptor) MarshalTo(dAtA []byte) (int, error) {
var i int
_ = i
var l int
_ = l
if len(m.MediaType) > 0 {
dAtA[i] = 0xa
i++
i = encodeVarintDescriptor(dAtA, i, uint64(len(m.MediaType)))
i += copy(dAtA[i:], m.MediaType)
}
if len(m.Digest) > 0 {
dAtA[i] = 0x12
i++
i = encodeVarintDescriptor(dAtA, i, uint64(len(m.Digest)))
i += copy(dAtA[i:], m.Digest)
}
if m.Size_ != 0 {
dAtA[i] = 0x18
i++
i = encodeVarintDescriptor(dAtA, i, uint64(m.Size_))
}
return i, nil
}
func encodeVarintDescriptor(dAtA []byte, offset int, v uint64) int {
for v >= 1<<7 {
dAtA[offset] = uint8(v&0x7f | 0x80)
v >>= 7
offset++
}
dAtA[offset] = uint8(v)
return offset + 1
}
func (m *Descriptor) Size() (n int) {
var l int
_ = l
l = len(m.MediaType)
if l > 0 {
n += 1 + l + sovDescriptor(uint64(l))
}
l = len(m.Digest)
if l > 0 {
n += 1 + l + sovDescriptor(uint64(l))
}
if m.Size_ != 0 {
n += 1 + sovDescriptor(uint64(m.Size_))
}
return n
}
func sovDescriptor(x uint64) (n int) {
for {
n++
x >>= 7
if x == 0 {
break
}
}
return n
}
func sozDescriptor(x uint64) (n int) {
return sovDescriptor(uint64((x << 1) ^ uint64((int64(x) >> 63))))
}
func (this *Descriptor) String() string {
if this == nil {
return "nil"
}
s := strings.Join([]string{`&Descriptor{`,
`MediaType:` + fmt.Sprintf("%v", this.MediaType) + `,`,
`Digest:` + fmt.Sprintf("%v", this.Digest) + `,`,
`Size_:` + fmt.Sprintf("%v", this.Size_) + `,`,
`}`,
}, "")
return s
}
func valueToStringDescriptor(v interface{}) string {
rv := reflect.ValueOf(v)
if rv.IsNil() {
return "nil"
}
pv := reflect.Indirect(rv).Interface()
return fmt.Sprintf("*%v", pv)
}
func (m *Descriptor) Unmarshal(dAtA []byte) error {
l := len(dAtA)
iNdEx := 0
for iNdEx < l {
preIndex := iNdEx
var wire uint64
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return ErrIntOverflowDescriptor
}
if iNdEx >= l {
return io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
wire |= (uint64(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
fieldNum := int32(wire >> 3)
wireType := int(wire & 0x7)
if wireType == 4 {
return fmt.Errorf("proto: Descriptor: wiretype end group for non-group")
}
if fieldNum <= 0 {
return fmt.Errorf("proto: Descriptor: illegal tag %d (wire type %d)", fieldNum, wire)
}
switch fieldNum {
case 1:
if wireType != 2 {
return fmt.Errorf("proto: wrong wireType = %d for field MediaType", wireType)
}
var stringLen uint64
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return ErrIntOverflowDescriptor
}
if iNdEx >= l {
return io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
stringLen |= (uint64(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
intStringLen := int(stringLen)
if intStringLen < 0 {
return ErrInvalidLengthDescriptor
}
postIndex := iNdEx + intStringLen
if postIndex > l {
return io.ErrUnexpectedEOF
}
m.MediaType = string(dAtA[iNdEx:postIndex])
iNdEx = postIndex
case 2:
if wireType != 2 {
return fmt.Errorf("proto: wrong wireType = %d for field Digest", wireType)
}
var stringLen uint64
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return ErrIntOverflowDescriptor
}
if iNdEx >= l {
return io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
stringLen |= (uint64(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
intStringLen := int(stringLen)
if intStringLen < 0 {
return ErrInvalidLengthDescriptor
}
postIndex := iNdEx + intStringLen
if postIndex > l {
return io.ErrUnexpectedEOF
}
m.Digest = github_com_opencontainers_go_digest.Digest(dAtA[iNdEx:postIndex])
iNdEx = postIndex
case 3:
if wireType != 0 {
return fmt.Errorf("proto: wrong wireType = %d for field Size_", wireType)
}
m.Size_ = 0
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return ErrIntOverflowDescriptor
}
if iNdEx >= l {
return io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
m.Size_ |= (int64(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
default:
iNdEx = preIndex
skippy, err := skipDescriptor(dAtA[iNdEx:])
if err != nil {
return err
}
if skippy < 0 {
return ErrInvalidLengthDescriptor
}
if (iNdEx + skippy) > l {
return io.ErrUnexpectedEOF
}
iNdEx += skippy
}
}
if iNdEx > l {
return io.ErrUnexpectedEOF
}
return nil
}
func skipDescriptor(dAtA []byte) (n int, err error) {
l := len(dAtA)
iNdEx := 0
for iNdEx < l {
var wire uint64
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return 0, ErrIntOverflowDescriptor
}
if iNdEx >= l {
return 0, io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
wire |= (uint64(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
wireType := int(wire & 0x7)
switch wireType {
case 0:
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return 0, ErrIntOverflowDescriptor
}
if iNdEx >= l {
return 0, io.ErrUnexpectedEOF
}
iNdEx++
if dAtA[iNdEx-1] < 0x80 {
break
}
}
return iNdEx, nil
case 1:
iNdEx += 8
return iNdEx, nil
case 2:
var length int
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return 0, ErrIntOverflowDescriptor
}
if iNdEx >= l {
return 0, io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
length |= (int(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
iNdEx += length
if length < 0 {
return 0, ErrInvalidLengthDescriptor
}
return iNdEx, nil
case 3:
for {
var innerWire uint64
var start int = iNdEx
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return 0, ErrIntOverflowDescriptor
}
if iNdEx >= l {
return 0, io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
inner | {
return nil, err
} | conditional_block |
|
descriptor.pb.go |
i++
i = encodeVarintDescriptor(dAtA, i, uint64(len(m.MediaType)))
i += copy(dAtA[i:], m.MediaType)
}
if len(m.Digest) > 0 {
dAtA[i] = 0x12
i++
i = encodeVarintDescriptor(dAtA, i, uint64(len(m.Digest)))
i += copy(dAtA[i:], m.Digest)
}
if m.Size_ != 0 {
dAtA[i] = 0x18
i++
i = encodeVarintDescriptor(dAtA, i, uint64(m.Size_))
}
return i, nil
}
func encodeVarintDescriptor(dAtA []byte, offset int, v uint64) int {
for v >= 1<<7 {
dAtA[offset] = uint8(v&0x7f | 0x80)
v >>= 7
offset++
}
dAtA[offset] = uint8(v)
return offset + 1
}
func (m *Descriptor) Size() (n int) {
var l int
_ = l
l = len(m.MediaType)
if l > 0 {
n += 1 + l + sovDescriptor(uint64(l))
}
l = len(m.Digest)
if l > 0 {
n += 1 + l + sovDescriptor(uint64(l))
}
if m.Size_ != 0 {
n += 1 + sovDescriptor(uint64(m.Size_))
}
return n
}
func sovDescriptor(x uint64) (n int) {
for {
n++
x >>= 7
if x == 0 {
break
}
}
return n
}
func sozDescriptor(x uint64) (n int) {
return sovDescriptor(uint64((x << 1) ^ uint64((int64(x) >> 63))))
}
func (this *Descriptor) String() string {
if this == nil {
return "nil"
}
s := strings.Join([]string{`&Descriptor{`,
`MediaType:` + fmt.Sprintf("%v", this.MediaType) + `,`,
`Digest:` + fmt.Sprintf("%v", this.Digest) + `,`,
`Size_:` + fmt.Sprintf("%v", this.Size_) + `,`,
`}`,
}, "")
return s
}
func valueToStringDescriptor(v interface{}) string {
rv := reflect.ValueOf(v)
if rv.IsNil() {
return "nil"
}
pv := reflect.Indirect(rv).Interface()
return fmt.Sprintf("*%v", pv)
}
func (m *Descriptor) Unmarshal(dAtA []byte) error {
l := len(dAtA)
iNdEx := 0
for iNdEx < l {
preIndex := iNdEx
var wire uint64
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return ErrIntOverflowDescriptor
}
if iNdEx >= l {
return io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
wire |= (uint64(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
fieldNum := int32(wire >> 3)
wireType := int(wire & 0x7)
if wireType == 4 {
return fmt.Errorf("proto: Descriptor: wiretype end group for non-group")
}
if fieldNum <= 0 {
return fmt.Errorf("proto: Descriptor: illegal tag %d (wire type %d)", fieldNum, wire)
}
switch fieldNum {
case 1:
if wireType != 2 {
return fmt.Errorf("proto: wrong wireType = %d for field MediaType", wireType)
}
var stringLen uint64
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return ErrIntOverflowDescriptor
}
if iNdEx >= l {
return io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
stringLen |= (uint64(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
intStringLen := int(stringLen)
if intStringLen < 0 {
return ErrInvalidLengthDescriptor
}
postIndex := iNdEx + intStringLen
if postIndex > l {
return io.ErrUnexpectedEOF
}
m.MediaType = string(dAtA[iNdEx:postIndex])
iNdEx = postIndex
case 2:
if wireType != 2 {
return fmt.Errorf("proto: wrong wireType = %d for field Digest", wireType)
}
var stringLen uint64
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return ErrIntOverflowDescriptor
}
if iNdEx >= l {
return io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
stringLen |= (uint64(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
intStringLen := int(stringLen)
if intStringLen < 0 {
return ErrInvalidLengthDescriptor
}
postIndex := iNdEx + intStringLen
if postIndex > l {
return io.ErrUnexpectedEOF
} | if wireType != 0 {
return fmt.Errorf("proto: wrong wireType = %d for field Size_", wireType)
}
m.Size_ = 0
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return ErrIntOverflowDescriptor
}
if iNdEx >= l {
return io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
m.Size_ |= (int64(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
default:
iNdEx = preIndex
skippy, err := skipDescriptor(dAtA[iNdEx:])
if err != nil {
return err
}
if skippy < 0 {
return ErrInvalidLengthDescriptor
}
if (iNdEx + skippy) > l {
return io.ErrUnexpectedEOF
}
iNdEx += skippy
}
}
if iNdEx > l {
return io.ErrUnexpectedEOF
}
return nil
}
func skipDescriptor(dAtA []byte) (n int, err error) {
l := len(dAtA)
iNdEx := 0
for iNdEx < l {
var wire uint64
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return 0, ErrIntOverflowDescriptor
}
if iNdEx >= l {
return 0, io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
wire |= (uint64(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
wireType := int(wire & 0x7)
switch wireType {
case 0:
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return 0, ErrIntOverflowDescriptor
}
if iNdEx >= l {
return 0, io.ErrUnexpectedEOF
}
iNdEx++
if dAtA[iNdEx-1] < 0x80 {
break
}
}
return iNdEx, nil
case 1:
iNdEx += 8
return iNdEx, nil
case 2:
var length int
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return 0, ErrIntOverflowDescriptor
}
if iNdEx >= l {
return 0, io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
length |= (int(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
iNdEx += length
if length < 0 {
return 0, ErrInvalidLengthDescriptor
}
return iNdEx, nil
case 3:
for {
var innerWire uint64
var start int = iNdEx
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return 0, ErrIntOverflowDescriptor
}
if iNdEx >= l {
return 0, io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
innerWire |= (uint64(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
innerWireType := int(innerWire & 0x7)
if innerWireType == 4 {
break
}
next, err := skip | m.Digest = github_com_opencontainers_go_digest.Digest(dAtA[iNdEx:postIndex])
iNdEx = postIndex
case 3: | random_line_split |
descriptor.pb.go | i++
i = encodeVarintDescriptor(dAtA, i, uint64(len(m.MediaType)))
i += copy(dAtA[i:], m.MediaType)
}
if len(m.Digest) > 0 {
dAtA[i] = 0x12
i++
i = encodeVarintDescriptor(dAtA, i, uint64(len(m.Digest)))
i += copy(dAtA[i:], m.Digest)
}
if m.Size_ != 0 {
dAtA[i] = 0x18
i++
i = encodeVarintDescriptor(dAtA, i, uint64(m.Size_))
}
return i, nil
}
func encodeVarintDescriptor(dAtA []byte, offset int, v uint64) int {
for v >= 1<<7 {
dAtA[offset] = uint8(v&0x7f | 0x80)
v >>= 7
offset++
}
dAtA[offset] = uint8(v)
return offset + 1
}
func (m *Descriptor) Size() (n int) {
var l int
_ = l
l = len(m.MediaType)
if l > 0 {
n += 1 + l + sovDescriptor(uint64(l))
}
l = len(m.Digest)
if l > 0 {
n += 1 + l + sovDescriptor(uint64(l))
}
if m.Size_ != 0 {
n += 1 + sovDescriptor(uint64(m.Size_))
}
return n
}
func sovDescriptor(x uint64) (n int) {
for {
n++
x >>= 7
if x == 0 {
break
}
}
return n
}
func sozDescriptor(x uint64) (n int) {
return sovDescriptor(uint64((x << 1) ^ uint64((int64(x) >> 63))))
}
func (this *Descriptor) String() string {
if this == nil {
return "nil"
}
s := strings.Join([]string{`&Descriptor{`,
`MediaType:` + fmt.Sprintf("%v", this.MediaType) + `,`,
`Digest:` + fmt.Sprintf("%v", this.Digest) + `,`,
`Size_:` + fmt.Sprintf("%v", this.Size_) + `,`,
`}`,
}, "")
return s
}
func valueToStringDescriptor(v interface{}) string {
rv := reflect.ValueOf(v)
if rv.IsNil() {
return "nil"
}
pv := reflect.Indirect(rv).Interface()
return fmt.Sprintf("*%v", pv)
}
func (m *Descriptor) Unmarshal(dAtA []byte) error {
l := len(dAtA)
iNdEx := 0
for iNdEx < l {
preIndex := iNdEx
var wire uint64
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return ErrIntOverflowDescriptor
}
if iNdEx >= l {
return io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
wire |= (uint64(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
fieldNum := int32(wire >> 3)
wireType := int(wire & 0x7)
if wireType == 4 {
return fmt.Errorf("proto: Descriptor: wiretype end group for non-group")
}
if fieldNum <= 0 {
return fmt.Errorf("proto: Descriptor: illegal tag %d (wire type %d)", fieldNum, wire)
}
switch fieldNum {
case 1:
if wireType != 2 {
return fmt.Errorf("proto: wrong wireType = %d for field MediaType", wireType)
}
var stringLen uint64
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return ErrIntOverflowDescriptor
}
if iNdEx >= l {
return io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
stringLen |= (uint64(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
intStringLen := int(stringLen)
if intStringLen < 0 {
return ErrInvalidLengthDescriptor
}
postIndex := iNdEx + intStringLen
if postIndex > l {
return io.ErrUnexpectedEOF
}
m.MediaType = string(dAtA[iNdEx:postIndex])
iNdEx = postIndex
case 2:
if wireType != 2 {
return fmt.Errorf("proto: wrong wireType = %d for field Digest", wireType)
}
var stringLen uint64
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return ErrIntOverflowDescriptor
}
if iNdEx >= l {
return io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
stringLen |= (uint64(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
intStringLen := int(stringLen)
if intStringLen < 0 {
return ErrInvalidLengthDescriptor
}
postIndex := iNdEx + intStringLen
if postIndex > l {
return io.ErrUnexpectedEOF
}
m.Digest = github_com_opencontainers_go_digest.Digest(dAtA[iNdEx:postIndex])
iNdEx = postIndex
case 3:
if wireType != 0 {
return fmt.Errorf("proto: wrong wireType = %d for field Size_", wireType)
}
m.Size_ = 0
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return ErrIntOverflowDescriptor
}
if iNdEx >= l {
return io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
m.Size_ |= (int64(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
default:
iNdEx = preIndex
skippy, err := skipDescriptor(dAtA[iNdEx:])
if err != nil {
return err
}
if skippy < 0 {
return ErrInvalidLengthDescriptor
}
if (iNdEx + skippy) > l {
return io.ErrUnexpectedEOF
}
iNdEx += skippy
}
}
if iNdEx > l {
return io.ErrUnexpectedEOF
}
return nil
}
func | (dAtA []byte) (n int, err error) {
l := len(dAtA)
iNdEx := 0
for iNdEx < l {
var wire uint64
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return 0, ErrIntOverflowDescriptor
}
if iNdEx >= l {
return 0, io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
wire |= (uint64(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
wireType := int(wire & 0x7)
switch wireType {
case 0:
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return 0, ErrIntOverflowDescriptor
}
if iNdEx >= l {
return 0, io.ErrUnexpectedEOF
}
iNdEx++
if dAtA[iNdEx-1] < 0x80 {
break
}
}
return iNdEx, nil
case 1:
iNdEx += 8
return iNdEx, nil
case 2:
var length int
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return 0, ErrIntOverflowDescriptor
}
if iNdEx >= l {
return 0, io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
length |= (int(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
iNdEx += length
if length < 0 {
return 0, ErrInvalidLengthDescriptor
}
return iNdEx, nil
case 3:
for {
var innerWire uint64
var start int = iNdEx
for shift := uint(0); ; shift += 7 {
if shift >= 64 {
return 0, ErrIntOverflowDescriptor
}
if iNdEx >= l {
return 0, io.ErrUnexpectedEOF
}
b := dAtA[iNdEx]
iNdEx++
innerWire |= (uint64(b) & 0x7F) << shift
if b < 0x80 {
break
}
}
innerWireType := int(innerWire & 0x7)
if innerWireType == 4 {
break
}
next, err := skip | skipDescriptor | identifier_name |
main.rs | "Tool for mopdifying system wide hosts file to simulate arbitrary DNS A and AAAA records.
Expects a hosts file at {:?} and a configuration in YAML format at {:?}. This
program is intended to be run by non-priviledged users with the help of setuid. It therefore has
some safety features.
Any modifications will not be persisted until the end of program execution. In the event of any
error, the original hosts file will not be modified.
The configuration defines a whitelist of hostnames that can be modified. This program will refuse
to modify any hostname not present in that list. It will also ensure that certain hostnames are
never modified:
- {:?}
- {:?}
- {:?}
- {:?}
- {:?}
- {:?} <- current hostname
The only exception is if the config variable `enable_dangerous_operations` is set to true. Then even
these reserved hostnames can be modified."##,
PATH_HOSTSFILE,
PATH_CONFIG,
config::RESERVED_LOCALHOST,
config::RESERVED_IP6_LOCALHOST,
config::RESERVED_IP6_LOOPBACK,
config::RESERVED_IP6_ALLNODES,
config::RESERVED_IP6_ALLROUTERS,
hostname
);
let app = app
// .before_help("PRE!!!")
// .after_help("POST!!!")
.about(str_about.as_ref());
opts::HostsArgs::from_clap(&app.get_matches())
};
if opts.generate_sample_config {
let mut out = stdout();
let mut sample = HostsmodConfig::default();
sample.whitelist.insert("somerandomhost.with.tld".into());
serde_yaml::to_writer(&mut out, &sample).expect("unable to write default config to stdout");
return;
}
let euid = users::get_effective_uid();
// dbg!(uid);
if euid != 0 {
eprintln!("not effectively root, forced dry-run mode");
opts.dry_run = true;
}
// dbg!(opts);
// open file
let mut file_hosts_orig = OpenOptions::new()
.read(true)
// .write(!opts.dry_run)
.write(false)
.truncate(false)
.create(false)
.open(PATH_HOSTSFILE)
.expect("unable to open hosts");
// let opt_file_hosts_new = if opts.dry_run {
// None
// } else {
// Some(
// OpenOptions::new()
// .write(true)
// .create_new(true)
// .open(PATH_HOSTSFILE_NEW)
// .expect("unable to open new hosts file for writing! Stale file from previous run?"),
// )
// };
let mut str_content = String::with_capacity(1024 * 8);
let len_content = file_hosts_orig
.read_to_string(&mut str_content)
.expect("unable to read hosts file as UTF-8 string");
let mut hosts_parts =
try_parse_hosts(&str_content).expect("unable to parse contents of hosts file");
trim_hosts_parts(&mut hosts_parts);
let hosts_parts_orig = hosts_parts.clone();
// eprintln!("PRE-actions: {:#?}", &hosts_parts);
let cfg: HostsmodConfig = {
// TODO: check config file ownership & access rights
let file_cfg = BufReader::new(File::open(PATH_CONFIG).expect("unable to open config file"));
serde_yaml::from_reader(file_cfg).expect("unable to parse configuration")
};
if opts.dry_run || opts.verbose {
if opts.verbose {
eprintln!("config: {:#?}", cfg);
}
println!("original contents:\n>>>\n{}<<<", str_content);
}
let mut found_pre = vec![false; DONT_TOUCH.len()];
if !cfg.enable_dangerous_operations {
for (dt, found) in DONT_TOUCH.iter().zip(found_pre.iter_mut()) {
let dt_host = if dt.hostname == RESERVED_HOSTNAME {
Cow::Borrowed(hostname)
} else {
Cow::Borrowed(dt.hostname.as_ref())
};
for part in &hosts_parts {
if part.matches_hostname(&dt_host) && part.matches_ip(&dt.ip) {
*found = true;
}
}
}
}
let found_pre = found_pre;
// execute actions
perform_actions(&mut opts, &mut hosts_parts, &cfg).expect("unable to modify hosts file");
if !opts.dry_run && hosts_parts == hosts_parts_orig {
if opts.verbose {
println!("no changes, not modifying hosts file");
}
return;
}
// remove redundant Empty elements
trim_hosts_parts(&mut hosts_parts);
{
let mut remove = false;
hosts_parts.retain(|item| match (item.is_empty(), remove) {
(true, true) => false,
(true, false) => {
remove = true;
true
}
(false, _) => {
remove = false;
true
}
});
}
// eprintln!("POST-actions: {:#?}", &hosts_parts);
// compare against DONT_TOUCH
let buf_generate = generate_hosts_file(len_content, &hosts_parts);
// eprintln!(">\n{}<", &buf_generate);
// safety checks
if !cfg.enable_dangerous_operations {
let mut found_post = vec![false; DONT_TOUCH.len()];
for (dt, found) in DONT_TOUCH.iter().zip(found_post.iter_mut()) {
let dt_host = if dt.hostname == RESERVED_HOSTNAME {
Cow::Borrowed(hostname)
} else {
Cow::Borrowed(dt.hostname.as_ref())
};
for part in &hosts_parts {
match (part.matches_hostname(&dt_host), part.matches_ip(&dt.ip)) {
(true, true) => {
*found = true;
}
(true, false) => {
if DONT_TOUCH
.iter()
.find(|dt_lookup| {
// eprint!("conflict: {:?} == {:?} ", part, dt_lookup);
let res = part.matches_hostname(&dt_lookup.hostname)
&& part.matches_ip(&dt_lookup.ip);
// eprintln!("{}", res);
res
})
.is_none()
{
panic!(
"untouchable entry {:?} {:?} was changed! {:?}",
dt.ip, dt_host, part
);
}
// *found = true;
}
(false, _) => {}
}
}
}
if found_post != found_pre {
dbg!(&found_pre);
dbg!(&found_post);
for (i, (pre, post)) in found_pre.iter().zip(found_post.iter()).enumerate() {
if pre != post |
}
panic!("found_post != found_pre");
}
}
if opts.dry_run || opts.verbose {
println!("generated:\n>>>\n{}<<<", &buf_generate);
}
if opts.dry_run {
println!("DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN");
println!("hosts file not modified");
return;
}
let mut file_hosts_new = OpenOptions::new()
.write(true)
.create_new(true)
.open(PATH_HOSTSFILE_NEW)
.expect("unable to open new hosts file for writing! Stale file from previous run?");
file_hosts_new
.write_all(buf_generate.as_bytes())
.expect("unable to write generated hosts file");
file_hosts_new
.set_len(buf_generate.as_bytes().len() as u64)
.expect("unable to truncate hosts file to right len");
file_hosts_new.flush().expect("unable to flush hosts file");
// close file handles
drop(file_hosts_new);
drop(file_hosts_orig);
rename(PATH_HOSTSFILE_NEW, PATH_HOSTSFILE).expect("unable to move new hosts file into place!");
}
fn trim_hosts_parts(hosts_parts: &mut Vec<HostsPart>) {
let trim = hosts_parts
.iter()
.rev()
.take_while(|part| part.is_empty())
.count();
hosts_parts.truncate(hosts_parts.len() - trim);
}
fn perform_actions(
opts: &mut opts::HostsArgs,
hosts: &mut Vec<HostsPart>,
config: &HostsmodConfig,
) -> Result<(), String> {
'loop_actions: for action in &opts.actions {
match action {
Action::Define(ip, host) => {
if !config.whitelist.contains(host) {
return Err(format!("HOST {:?} not whitelisted!", host));
}
// eprintln!("defining additionally...: {:?} += {:?}", ip, host);
let mut opt_insert = Some(hosts.len());
let mut host_found_v4 = false;
let mut host_found_v6 = false;
for (i, part) in hosts
.iter_mut()
.enumerate()
.filter(|(_i, p)| p.matches_ip(ip) || p.matches_hostname(host))
{
// eprintln!("matching entry: {:?}", part);
let matches_hostname = | {
eprintln!("Difference: {:?}", DONT_TOUCH[i])
} | conditional_block |
main.rs | file at {:?} and a configuration in YAML format at {:?}. This
program is intended to be run by non-priviledged users with the help of setuid. It therefore has
some safety features.
Any modifications will not be persisted until the end of program execution. In the event of any
error, the original hosts file will not be modified.
The configuration defines a whitelist of hostnames that can be modified. This program will refuse
to modify any hostname not present in that list. It will also ensure that certain hostnames are
never modified:
- {:?}
- {:?}
- {:?}
- {:?}
- {:?}
- {:?} <- current hostname
The only exception is if the config variable `enable_dangerous_operations` is set to true. Then even
these reserved hostnames can be modified."##,
PATH_HOSTSFILE,
PATH_CONFIG,
config::RESERVED_LOCALHOST,
config::RESERVED_IP6_LOCALHOST,
config::RESERVED_IP6_LOOPBACK,
config::RESERVED_IP6_ALLNODES,
config::RESERVED_IP6_ALLROUTERS,
hostname
);
let app = app
// .before_help("PRE!!!")
// .after_help("POST!!!")
.about(str_about.as_ref());
opts::HostsArgs::from_clap(&app.get_matches())
};
if opts.generate_sample_config {
let mut out = stdout();
let mut sample = HostsmodConfig::default();
sample.whitelist.insert("somerandomhost.with.tld".into());
serde_yaml::to_writer(&mut out, &sample).expect("unable to write default config to stdout");
return;
}
let euid = users::get_effective_uid();
// dbg!(uid);
if euid != 0 {
eprintln!("not effectively root, forced dry-run mode");
opts.dry_run = true;
}
// dbg!(opts);
// open file
let mut file_hosts_orig = OpenOptions::new()
.read(true)
// .write(!opts.dry_run)
.write(false)
.truncate(false)
.create(false)
.open(PATH_HOSTSFILE)
.expect("unable to open hosts");
// let opt_file_hosts_new = if opts.dry_run {
// None
// } else {
// Some(
// OpenOptions::new()
// .write(true)
// .create_new(true)
// .open(PATH_HOSTSFILE_NEW)
// .expect("unable to open new hosts file for writing! Stale file from previous run?"),
// )
// };
let mut str_content = String::with_capacity(1024 * 8);
let len_content = file_hosts_orig
.read_to_string(&mut str_content)
.expect("unable to read hosts file as UTF-8 string");
let mut hosts_parts =
try_parse_hosts(&str_content).expect("unable to parse contents of hosts file");
trim_hosts_parts(&mut hosts_parts);
let hosts_parts_orig = hosts_parts.clone();
// eprintln!("PRE-actions: {:#?}", &hosts_parts);
let cfg: HostsmodConfig = {
// TODO: check config file ownership & access rights
let file_cfg = BufReader::new(File::open(PATH_CONFIG).expect("unable to open config file"));
serde_yaml::from_reader(file_cfg).expect("unable to parse configuration")
};
if opts.dry_run || opts.verbose {
if opts.verbose {
eprintln!("config: {:#?}", cfg);
}
println!("original contents:\n>>>\n{}<<<", str_content);
}
let mut found_pre = vec![false; DONT_TOUCH.len()];
if !cfg.enable_dangerous_operations {
for (dt, found) in DONT_TOUCH.iter().zip(found_pre.iter_mut()) {
let dt_host = if dt.hostname == RESERVED_HOSTNAME {
Cow::Borrowed(hostname)
} else {
Cow::Borrowed(dt.hostname.as_ref())
};
for part in &hosts_parts {
if part.matches_hostname(&dt_host) && part.matches_ip(&dt.ip) {
*found = true;
}
}
}
}
let found_pre = found_pre;
// execute actions
perform_actions(&mut opts, &mut hosts_parts, &cfg).expect("unable to modify hosts file");
if !opts.dry_run && hosts_parts == hosts_parts_orig {
if opts.verbose {
println!("no changes, not modifying hosts file");
}
return;
}
// remove redundant Empty elements
trim_hosts_parts(&mut hosts_parts);
{
let mut remove = false;
hosts_parts.retain(|item| match (item.is_empty(), remove) {
(true, true) => false,
(true, false) => {
remove = true;
true
}
(false, _) => {
remove = false;
true
}
});
}
// eprintln!("POST-actions: {:#?}", &hosts_parts);
// compare against DONT_TOUCH
let buf_generate = generate_hosts_file(len_content, &hosts_parts);
// eprintln!(">\n{}<", &buf_generate);
// safety checks
if !cfg.enable_dangerous_operations {
let mut found_post = vec![false; DONT_TOUCH.len()];
for (dt, found) in DONT_TOUCH.iter().zip(found_post.iter_mut()) {
let dt_host = if dt.hostname == RESERVED_HOSTNAME {
Cow::Borrowed(hostname)
} else {
Cow::Borrowed(dt.hostname.as_ref())
};
for part in &hosts_parts {
match (part.matches_hostname(&dt_host), part.matches_ip(&dt.ip)) {
(true, true) => {
*found = true;
}
(true, false) => {
if DONT_TOUCH
.iter()
.find(|dt_lookup| {
// eprint!("conflict: {:?} == {:?} ", part, dt_lookup);
let res = part.matches_hostname(&dt_lookup.hostname)
&& part.matches_ip(&dt_lookup.ip);
// eprintln!("{}", res);
res
})
.is_none()
{
panic!(
"untouchable entry {:?} {:?} was changed! {:?}",
dt.ip, dt_host, part
);
}
// *found = true;
}
(false, _) => {}
}
}
}
if found_post != found_pre {
dbg!(&found_pre);
dbg!(&found_post);
for (i, (pre, post)) in found_pre.iter().zip(found_post.iter()).enumerate() {
if pre != post {
eprintln!("Difference: {:?}", DONT_TOUCH[i])
}
}
panic!("found_post != found_pre");
}
}
if opts.dry_run || opts.verbose {
println!("generated:\n>>>\n{}<<<", &buf_generate);
}
if opts.dry_run {
println!("DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN");
println!("hosts file not modified");
return;
}
let mut file_hosts_new = OpenOptions::new()
.write(true)
.create_new(true)
.open(PATH_HOSTSFILE_NEW)
.expect("unable to open new hosts file for writing! Stale file from previous run?");
file_hosts_new
.write_all(buf_generate.as_bytes())
.expect("unable to write generated hosts file");
file_hosts_new
.set_len(buf_generate.as_bytes().len() as u64)
.expect("unable to truncate hosts file to right len");
file_hosts_new.flush().expect("unable to flush hosts file");
// close file handles
drop(file_hosts_new);
drop(file_hosts_orig);
rename(PATH_HOSTSFILE_NEW, PATH_HOSTSFILE).expect("unable to move new hosts file into place!");
}
fn trim_hosts_parts(hosts_parts: &mut Vec<HostsPart>) {
let trim = hosts_parts
.iter()
.rev()
.take_while(|part| part.is_empty())
.count();
hosts_parts.truncate(hosts_parts.len() - trim);
}
fn perform_actions(
opts: &mut opts::HostsArgs,
hosts: &mut Vec<HostsPart>,
config: &HostsmodConfig,
) -> Result<(), String> | {
'loop_actions: for action in &opts.actions {
match action {
Action::Define(ip, host) => {
if !config.whitelist.contains(host) {
return Err(format!("HOST {:?} not whitelisted!", host));
}
// eprintln!("defining additionally...: {:?} += {:?}", ip, host);
let mut opt_insert = Some(hosts.len());
let mut host_found_v4 = false;
let mut host_found_v6 = false;
for (i, part) in hosts
.iter_mut()
.enumerate()
.filter(|(_i, p)| p.matches_ip(ip) || p.matches_hostname(host))
{
// eprintln!("matching entry: {:?}", part);
let matches_hostname = part.matches_hostname(host);
if part.matches_ip(ip) && matches_hostname {
// eprintln!("already defined, NOP"); | identifier_body |
|
main.rs | ##"Tool for mopdifying system wide hosts file to simulate arbitrary DNS A and AAAA records.
Expects a hosts file at {:?} and a configuration in YAML format at {:?}. This
program is intended to be run by non-priviledged users with the help of setuid. It therefore has
some safety features.
Any modifications will not be persisted until the end of program execution. In the event of any
error, the original hosts file will not be modified.
The configuration defines a whitelist of hostnames that can be modified. This program will refuse
to modify any hostname not present in that list. It will also ensure that certain hostnames are
never modified:
- {:?}
- {:?}
- {:?}
- {:?}
- {:?}
- {:?} <- current hostname
The only exception is if the config variable `enable_dangerous_operations` is set to true. Then even
these reserved hostnames can be modified."##,
PATH_HOSTSFILE,
PATH_CONFIG,
config::RESERVED_LOCALHOST,
config::RESERVED_IP6_LOCALHOST,
config::RESERVED_IP6_LOOPBACK,
config::RESERVED_IP6_ALLNODES,
config::RESERVED_IP6_ALLROUTERS,
hostname
);
let app = app
// .before_help("PRE!!!")
// .after_help("POST!!!")
.about(str_about.as_ref());
opts::HostsArgs::from_clap(&app.get_matches())
};
if opts.generate_sample_config {
let mut out = stdout();
let mut sample = HostsmodConfig::default();
sample.whitelist.insert("somerandomhost.with.tld".into());
serde_yaml::to_writer(&mut out, &sample).expect("unable to write default config to stdout");
return;
}
let euid = users::get_effective_uid();
// dbg!(uid);
if euid != 0 {
eprintln!("not effectively root, forced dry-run mode");
opts.dry_run = true;
}
// dbg!(opts);
// open file
let mut file_hosts_orig = OpenOptions::new()
.read(true)
// .write(!opts.dry_run)
.write(false)
.truncate(false)
.create(false)
.open(PATH_HOSTSFILE)
.expect("unable to open hosts");
// let opt_file_hosts_new = if opts.dry_run {
// None
// } else {
// Some(
// OpenOptions::new()
// .write(true)
// .create_new(true)
// .open(PATH_HOSTSFILE_NEW)
// .expect("unable to open new hosts file for writing! Stale file from previous run?"),
// )
// };
let mut str_content = String::with_capacity(1024 * 8);
let len_content = file_hosts_orig
.read_to_string(&mut str_content)
.expect("unable to read hosts file as UTF-8 string");
let mut hosts_parts =
try_parse_hosts(&str_content).expect("unable to parse contents of hosts file");
trim_hosts_parts(&mut hosts_parts);
let hosts_parts_orig = hosts_parts.clone();
// eprintln!("PRE-actions: {:#?}", &hosts_parts);
let cfg: HostsmodConfig = {
// TODO: check config file ownership & access rights
let file_cfg = BufReader::new(File::open(PATH_CONFIG).expect("unable to open config file"));
serde_yaml::from_reader(file_cfg).expect("unable to parse configuration")
};
if opts.dry_run || opts.verbose {
if opts.verbose {
eprintln!("config: {:#?}", cfg);
}
println!("original contents:\n>>>\n{}<<<", str_content);
}
let mut found_pre = vec![false; DONT_TOUCH.len()];
if !cfg.enable_dangerous_operations {
for (dt, found) in DONT_TOUCH.iter().zip(found_pre.iter_mut()) {
let dt_host = if dt.hostname == RESERVED_HOSTNAME {
Cow::Borrowed(hostname)
} else {
Cow::Borrowed(dt.hostname.as_ref())
};
for part in &hosts_parts {
if part.matches_hostname(&dt_host) && part.matches_ip(&dt.ip) {
*found = true;
}
}
}
}
let found_pre = found_pre;
// execute actions
perform_actions(&mut opts, &mut hosts_parts, &cfg).expect("unable to modify hosts file");
if !opts.dry_run && hosts_parts == hosts_parts_orig {
if opts.verbose {
println!("no changes, not modifying hosts file");
}
return;
}
// remove redundant Empty elements
trim_hosts_parts(&mut hosts_parts);
{
let mut remove = false;
hosts_parts.retain(|item| match (item.is_empty(), remove) {
(true, true) => false,
(true, false) => {
remove = true;
true
}
(false, _) => {
remove = false;
true
}
});
}
// eprintln!("POST-actions: {:#?}", &hosts_parts);
// compare against DONT_TOUCH
let buf_generate = generate_hosts_file(len_content, &hosts_parts);
// eprintln!(">\n{}<", &buf_generate);
// safety checks
if !cfg.enable_dangerous_operations {
let mut found_post = vec![false; DONT_TOUCH.len()];
for (dt, found) in DONT_TOUCH.iter().zip(found_post.iter_mut()) {
let dt_host = if dt.hostname == RESERVED_HOSTNAME {
Cow::Borrowed(hostname)
} else {
Cow::Borrowed(dt.hostname.as_ref())
};
for part in &hosts_parts {
match (part.matches_hostname(&dt_host), part.matches_ip(&dt.ip)) {
(true, true) => {
*found = true;
}
(true, false) => {
if DONT_TOUCH
.iter()
.find(|dt_lookup| {
// eprint!("conflict: {:?} == {:?} ", part, dt_lookup);
let res = part.matches_hostname(&dt_lookup.hostname)
&& part.matches_ip(&dt_lookup.ip);
// eprintln!("{}", res);
res
})
.is_none()
{
panic!(
"untouchable entry {:?} {:?} was changed! {:?}",
dt.ip, dt_host, part
);
}
// *found = true;
}
(false, _) => {}
}
}
}
if found_post != found_pre {
dbg!(&found_pre);
dbg!(&found_post);
for (i, (pre, post)) in found_pre.iter().zip(found_post.iter()).enumerate() {
if pre != post {
eprintln!("Difference: {:?}", DONT_TOUCH[i])
}
}
panic!("found_post != found_pre");
}
}
if opts.dry_run || opts.verbose {
println!("generated:\n>>>\n{}<<<", &buf_generate);
}
if opts.dry_run {
println!("DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN");
println!("hosts file not modified");
return;
}
let mut file_hosts_new = OpenOptions::new()
.write(true)
.create_new(true)
.open(PATH_HOSTSFILE_NEW)
.expect("unable to open new hosts file for writing! Stale file from previous run?");
file_hosts_new
.write_all(buf_generate.as_bytes())
.expect("unable to write generated hosts file");
file_hosts_new
.set_len(buf_generate.as_bytes().len() as u64)
.expect("unable to truncate hosts file to right len");
file_hosts_new.flush().expect("unable to flush hosts file");
// close file handles
drop(file_hosts_new);
drop(file_hosts_orig);
rename(PATH_HOSTSFILE_NEW, PATH_HOSTSFILE).expect("unable to move new hosts file into place!");
}
fn trim_hosts_parts(hosts_parts: &mut Vec<HostsPart>) {
let trim = hosts_parts
.iter()
.rev()
.take_while(|part| part.is_empty())
.count();
hosts_parts.truncate(hosts_parts.len() - trim);
}
fn perform_actions(
opts: &mut opts::HostsArgs,
hosts: &mut Vec<HostsPart>,
config: &HostsmodConfig,
) -> Result<(), String> {
'loop_actions: for action in &opts.actions {
match action {
Action::Define(ip, host) => {
if !config.whitelist.contains(host) {
return Err(format!("HOST {:?} not whitelisted!", host));
}
// eprintln!("defining additionally...: {:?} += {:?}", ip, host);
let mut opt_insert = Some(hosts.len());
let mut host_found_v4 = false;
let mut host_found_v6 = false;
for (i, part) in hosts
.iter_mut() | // eprintln!("matching entry: {:?}", part);
let matches_hostname = part | .enumerate()
.filter(|(_i, p)| p.matches_ip(ip) || p.matches_hostname(host))
{ | random_line_split |
main.rs | dt.hostname == RESERVED_HOSTNAME {
Cow::Borrowed(hostname)
} else {
Cow::Borrowed(dt.hostname.as_ref())
};
for part in &hosts_parts {
if part.matches_hostname(&dt_host) && part.matches_ip(&dt.ip) {
*found = true;
}
}
}
}
let found_pre = found_pre;
// execute actions
perform_actions(&mut opts, &mut hosts_parts, &cfg).expect("unable to modify hosts file");
if !opts.dry_run && hosts_parts == hosts_parts_orig {
if opts.verbose {
println!("no changes, not modifying hosts file");
}
return;
}
// remove redundant Empty elements
trim_hosts_parts(&mut hosts_parts);
{
let mut remove = false;
hosts_parts.retain(|item| match (item.is_empty(), remove) {
(true, true) => false,
(true, false) => {
remove = true;
true
}
(false, _) => {
remove = false;
true
}
});
}
// eprintln!("POST-actions: {:#?}", &hosts_parts);
// compare against DONT_TOUCH
let buf_generate = generate_hosts_file(len_content, &hosts_parts);
// eprintln!(">\n{}<", &buf_generate);
// safety checks
if !cfg.enable_dangerous_operations {
let mut found_post = vec![false; DONT_TOUCH.len()];
for (dt, found) in DONT_TOUCH.iter().zip(found_post.iter_mut()) {
let dt_host = if dt.hostname == RESERVED_HOSTNAME {
Cow::Borrowed(hostname)
} else {
Cow::Borrowed(dt.hostname.as_ref())
};
for part in &hosts_parts {
match (part.matches_hostname(&dt_host), part.matches_ip(&dt.ip)) {
(true, true) => {
*found = true;
}
(true, false) => {
if DONT_TOUCH
.iter()
.find(|dt_lookup| {
// eprint!("conflict: {:?} == {:?} ", part, dt_lookup);
let res = part.matches_hostname(&dt_lookup.hostname)
&& part.matches_ip(&dt_lookup.ip);
// eprintln!("{}", res);
res
})
.is_none()
{
panic!(
"untouchable entry {:?} {:?} was changed! {:?}",
dt.ip, dt_host, part
);
}
// *found = true;
}
(false, _) => {}
}
}
}
if found_post != found_pre {
dbg!(&found_pre);
dbg!(&found_post);
for (i, (pre, post)) in found_pre.iter().zip(found_post.iter()).enumerate() {
if pre != post {
eprintln!("Difference: {:?}", DONT_TOUCH[i])
}
}
panic!("found_post != found_pre");
}
}
if opts.dry_run || opts.verbose {
println!("generated:\n>>>\n{}<<<", &buf_generate);
}
if opts.dry_run {
println!("DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN DRY-RUN");
println!("hosts file not modified");
return;
}
let mut file_hosts_new = OpenOptions::new()
.write(true)
.create_new(true)
.open(PATH_HOSTSFILE_NEW)
.expect("unable to open new hosts file for writing! Stale file from previous run?");
file_hosts_new
.write_all(buf_generate.as_bytes())
.expect("unable to write generated hosts file");
file_hosts_new
.set_len(buf_generate.as_bytes().len() as u64)
.expect("unable to truncate hosts file to right len");
file_hosts_new.flush().expect("unable to flush hosts file");
// close file handles
drop(file_hosts_new);
drop(file_hosts_orig);
rename(PATH_HOSTSFILE_NEW, PATH_HOSTSFILE).expect("unable to move new hosts file into place!");
}
fn trim_hosts_parts(hosts_parts: &mut Vec<HostsPart>) {
let trim = hosts_parts
.iter()
.rev()
.take_while(|part| part.is_empty())
.count();
hosts_parts.truncate(hosts_parts.len() - trim);
}
fn perform_actions(
opts: &mut opts::HostsArgs,
hosts: &mut Vec<HostsPart>,
config: &HostsmodConfig,
) -> Result<(), String> {
'loop_actions: for action in &opts.actions {
match action {
Action::Define(ip, host) => {
if !config.whitelist.contains(host) {
return Err(format!("HOST {:?} not whitelisted!", host));
}
// eprintln!("defining additionally...: {:?} += {:?}", ip, host);
let mut opt_insert = Some(hosts.len());
let mut host_found_v4 = false;
let mut host_found_v6 = false;
for (i, part) in hosts
.iter_mut()
.enumerate()
.filter(|(_i, p)| p.matches_ip(ip) || p.matches_hostname(host))
{
// eprintln!("matching entry: {:?}", part);
let matches_hostname = part.matches_hostname(host);
if part.matches_ip(ip) && matches_hostname {
// eprintln!("already defined, NOP");
//opt_insert = None;
continue 'loop_actions;
}
if matches_hostname {
match part.get_family() {
Some(HostsPartFamily::IPv4) => {
if host_found_v4 || ip.is_ipv4() {
return Err(format!(
"duplicate entry for host {:?} {:?}",
host,
HostsPartFamily::IPv4
));
}
host_found_v4 = true;
}
Some(HostsPartFamily::IPv6) => {
if host_found_v6 || ip.is_ipv6() {
return Err(format!(
"duplicate entry for host {:?} {:?}",
host,
HostsPartFamily::IPv6
));
}
host_found_v6 = true;
}
None => {}
};
}
if opt_insert.is_some() {
opt_insert = Some(i + 1);
}
}
if let Some(insert) = opt_insert {
let insert = min(insert, hosts.len());
hosts.insert(
insert,
HostsPart::Entry(ip.clone(), vec![Cow::Owned(host.clone())], None),
);
}
}
Action::DefineExclusive(ip, host) => {
if !config.whitelist.contains(host) {
return Err(format!("HOST {:?} not whitelisted!", host));
}
// eprintln!("defining exclusively...: {:?} += {:?}", ip, host);
let mut vec_remove = vec![];
for (i, _part) in hosts
.iter()
.enumerate()
.filter(|(_i, p)| p.matches_hostname(host))
{
// eprintln!("matching entry: {:?}", part);
// if part.matches_ip(ip) && part.matches_hostname(host) {
// eprintln!("already defined, NOP");
// return;
// }
// insert = i + 1;
vec_remove.push(i);
}
for remove in vec_remove.iter().rev() {
hosts.remove(*remove);
}
let insert = vec_remove.into_iter().min().unwrap_or(hosts.len());
hosts.insert(
insert,
HostsPart::Entry(ip.clone(), vec![Cow::Owned(host.clone())], None),
);
}
Action::Remove(host) => {
if !config.whitelist.contains(host) {
return Err(format!("HOST {:?} not whitelisted!", host));
}
let mut vec_remove = vec![];
let mut vec_insert = vec![];
let mut offset_remove = 0;
for (i, part) in hosts
.iter()
.enumerate()
.filter(|(_i, p)| p.matches_hostname(host))
{
match part {
HostsPart::Entry(ip, hosts, opt_comment) => {
// eprintln!("matching entry: {:?}", (&ip, &hosts, &opt_comment));
if hosts.len() > 1 {
let mut hosts_filtered = hosts.clone();
hosts_filtered.retain(|ent| ent != host);
vec_insert.push((
i,
HostsPart::Entry(
ip.clone(),
hosts_filtered,
opt_comment.clone(),
),
));
offset_remove += 1;
}
vec_remove.push(offset_remove + i);
// for h in hosts {
// if h == host {
// }
// }
}
_ => {}
}
}
// dbg!(&vec_insert);
for (idx, part) in vec_insert {
hosts.insert(idx, part);
}
// dbg!(&vec_remove);
// unimplemented!();
for remove in vec_remove.iter().rev() {
hosts.remove(*remove);
}
}
}
}
Ok(())
}
fn generate_hosts_file(len_content: usize, parsed: &Vec<HostsPart>) -> String {
let mut buf_generate = String::with_capacity(len_content);
// eprintln!("rendering: {:?}", parsed);
fn | render_entry | identifier_name |
|
game.rs | Game speed in addition to player's speed
pub jump_timeout: f64, // Count down to allow the next jump
pub rotate_cam: bool, // Allow rotation of camera or not
pub bullets: i64, // The number of bullets left
pub recharge: f64, // Bullets recharge time
pub fps: f64, // Real fps of game
pub last_frame: Instant, // Moment of the last draw
}
pub enum Turn { Left, Right, None, }
// Configurable game's contansts.
// A tuple presents a range of something.
#[derive(Serialize, Deserialize, Clone)]
pub struct GameConfig {
pub title: String,
pub screen_size: ::Pixel,
pub ups: u64, // Update per second
pub max_fps: u64,
pub tunel_size: [f64; 3],
pub player_size: [f64; 3],
pub player_speed: (f64, f64), // min and max player speed
pub player_turn_speed: f64,
pub bot_size: [(f64, f64); 3], // Range of bot's size
pub bot_speed: (f64, f64),
pub bot_turn_speed: (f64, f64),
pub divider_size: [f64; 2],
pub camera_height: f64, // Height of camera (from player)
pub camera_distance: f64, // Distance from camera to player
pub decor_distance: f64, // Distance between each decoration
pub sprint_factor: f64,
pub spawn_time: (f64, f64),
pub game_sprint: f64, // The increase of game_speed
pub game_max_speed: f64,
pub player_jump_v: f64,
pub player_jump_a: f64,
pub jump_turn_decrease: f64,
pub jump_timeout: f64,
pub mouse_speed: f64,
pub trueshot_distance: f64,
pub bullet_stock: i64, // Number of bullets
pub recharge_time: f64,
pub bullet_len: f64,
pub bullet_speed: f64,
pub zoom_in: bool, // If true, zoom-in while on stare mode
}
impl Game {
pub fn new(config: GameConfig) -> Game {
let mut window: PistonWindow = WindowSettings::new(
config.title.clone(), [config.screen_size.w, config.screen_size.h])
.exit_on_esc(true).build()
.expect("Cannot create window.");
window.set_ups(config.ups);
window.set_max_fps(config.max_fps);
window.set_capture_cursor(true);
let glyphs = Glyphs::new("resources/Ubuntu-R.ttf", window.factory.clone())
.expect("Unable to load font.");
let bot_rules = BoxRules {
size: config.bot_size,
position: [(0., config.tunel_size[0]), (0., 0.), (config.tunel_size[2], config.tunel_size[2])],
speed: config.bot_speed,
turn_speed: config.bot_turn_speed,
color: vec![RED, ORANGE, VIOLET, GREEN, PALE],
jump_turn_decrease: config.jump_turn_decrease,
};
let world = World::new(&config);
let camera = Game::new_camera(&config, &world.player);
let state = State {
turn: Turn::None,
sprint: false,
spawn: 0.,
ended: false,
game_speed: 0.,
jump_timeout: 0.,
rotate_cam: false,
bullets: config.bullet_stock,
recharge: 0.,
fps: 0.,
last_frame: Instant::now(),
};
let ellipse = Ellipse {
color: BLACK.alpha(0.).into(),
border: Some(ellipse::Border {
color: RED.alpha(0.5).into(),
radius: 1.,
}),
resolution: 16,
};
Game {
config: config,
world: world,
window: window,
bot_rules: bot_rules,
camera: camera,
state: state,
glyphs: RefCell::new(glyphs),
ellipse: RefCell::new(ellipse),
}
}
fn new_camera<T: Car>(config: &GameConfig, player: &T) -> Camera {
Camera::new(
config.screen_size.clone(),
vec3(0., config.camera_height, -config.camera_distance) + player.pos()
)
}
// Re-calculate fps
fn update_fps(&mut self) {
let d = self.state.last_frame.elapsed();
self.state.last_frame = Instant::now();
self.state.fps = 1. / (d.as_secs() as f64 + 1e-9*d.subsec_nanos() as f64);
}
pub fn run(&mut self) {
while let Some(e) = self.window.next() {
match e {
Input::Press(key) => self.press(key),
Input::Release(key) => self.release(key),
Input::Render(_) => {
self.update_fps();
self.draw(&e);
},
Input::Update(args) => self.update(args.dt),
Input::Move(Motion::MouseRelative(a, b)) => self.mouse_move(a as f64, b as f64),
_ => {}
}
if self.state.ended {
break;
}
}
}
fn mouse_move(&mut self, x: f64, y: f64) {
if self.state.rotate_cam {
self.camera.rotate(x*self.config.mouse_speed, y*self.config.mouse_speed, self.world.player.position);
}
}
fn press(&mut self, key: Button) {
match key {
Button::Keyboard(Key::A) => self.state.turn = Turn::Left,
Button::Keyboard(Key::D) => self.state.turn = Turn::Right,
Button::Keyboard(Key::W) => self.state.sprint = true,
Button::Keyboard(Key::Space) => if self.state.jump_timeout <= 0. {
self.state.jump_timeout = self.config.jump_timeout;
self.world.player.jump();
},
Button::Mouse(MouseButton::Right) => {
if self.config.zoom_in {
self.camera.zoom_in();
}
self.state.rotate_cam = true;
},
Button::Mouse(MouseButton::Left) => if self.state.rotate_cam && self.state.bullets > 0 {
let mut pos = self.world.player.position;
pos.y += self.world.player.size.y;
let mut d = vec3(0., 0., self.config.trueshot_distance + self.config.camera_distance);
d = self.camera.c * d.magnitude2() / d.dot(self.camera.c);
d = self.camera.eye + d - pos;
d = d * self.config.bullet_speed / d.magnitude();
self.world.add_bullet(pos, d, self.config.bullet_len);
self.state.bullets -= 1;
if self.state.bullets <= 0 {
self.state.recharge = self.config.recharge_time;
}
},
_ => (),
}
}
fn release(&mut self, key: Button) |
fn draw(&mut self, e: &Input) {
// Return a horizontal bar
macro_rules! bar {
($curr: expr, $full: expr) => {
[0.,
15.0,
self.config.screen_size.w as f64/2.*$curr/$full,
20.0,]
};
}
let jump_bar = bar!(self.state.jump_timeout, self.config.jump_timeout);
let recharge_bar = bar!(self.state.recharge, self.config.recharge_time);
let bullets_bar = bar!(self.state.bullets as f64, self.config.bullet_stock as f64);
// Closure in `draw_2d` requires unique access to `self`,
// so we use RefCell to hack it.
let mut glyphs = self.glyphs.borrow_mut();
let fps = format!("{:.3}", self.state.fps);
let lines = self.world.render(&self.camera);
self.window.draw_2d(e, |c, g| {
clear(BLACK.into(), g);
for (l, color) in lines {
line(color.into(), 1., convert(l), c.transform, g);
}
rectangle(BLUE.alpha(0.4).into(), jump_bar, c.transform, g);
rectangle(RED.alpha(0.4).into(), recharge_bar, c.transform, g);
rectangle(GREEN.alpha(0.4).into(), bullets_bar, c.transform, g);
text(WHITE.into(), 10 | {
match key {
Button::Keyboard(Key::A) => if let Turn::Left = self.state.turn {
self.state.turn = Turn::None;
},
Button::Keyboard(Key::D) => if let Turn::Right = self.state.turn {
self.state.turn = Turn::None;
},
Button::Keyboard(Key::W) => self.state.sprint = false,
Button::Mouse(MouseButton::Right) => {
self.state.rotate_cam = false;
self.camera = Game::new_camera(&self.config, &self.world.player);
},
_ => (),
}
} | identifier_body |
game.rs | Game speed in addition to player's speed
pub jump_timeout: f64, // Count down to allow the next jump
pub rotate_cam: bool, // Allow rotation of camera or not
pub bullets: i64, // The number of bullets left
pub recharge: f64, // Bullets recharge time
pub fps: f64, // Real fps of game
pub last_frame: Instant, // Moment of the last draw
}
pub enum Turn { Left, Right, None, }
// Configurable game's contansts.
// A tuple presents a range of something.
#[derive(Serialize, Deserialize, Clone)]
pub struct GameConfig {
pub title: String,
pub screen_size: ::Pixel,
pub ups: u64, // Update per second
pub max_fps: u64,
pub tunel_size: [f64; 3],
pub player_size: [f64; 3],
pub player_speed: (f64, f64), // min and max player speed
pub player_turn_speed: f64,
pub bot_size: [(f64, f64); 3], // Range of bot's size
pub bot_speed: (f64, f64),
pub bot_turn_speed: (f64, f64),
pub divider_size: [f64; 2],
pub camera_height: f64, // Height of camera (from player)
pub camera_distance: f64, // Distance from camera to player
pub decor_distance: f64, // Distance between each decoration
pub sprint_factor: f64,
pub spawn_time: (f64, f64),
pub game_sprint: f64, // The increase of game_speed
pub game_max_speed: f64,
pub player_jump_v: f64,
pub player_jump_a: f64,
pub jump_turn_decrease: f64,
pub jump_timeout: f64,
pub mouse_speed: f64,
pub trueshot_distance: f64,
pub bullet_stock: i64, // Number of bullets
pub recharge_time: f64,
pub bullet_len: f64,
pub bullet_speed: f64,
pub zoom_in: bool, // If true, zoom-in while on stare mode
}
impl Game {
pub fn new(config: GameConfig) -> Game {
let mut window: PistonWindow = WindowSettings::new(
config.title.clone(), [config.screen_size.w, config.screen_size.h])
.exit_on_esc(true).build()
.expect("Cannot create window.");
window.set_ups(config.ups);
window.set_max_fps(config.max_fps);
window.set_capture_cursor(true);
let glyphs = Glyphs::new("resources/Ubuntu-R.ttf", window.factory.clone())
.expect("Unable to load font.");
let bot_rules = BoxRules {
size: config.bot_size,
position: [(0., config.tunel_size[0]), (0., 0.), (config.tunel_size[2], config.tunel_size[2])],
speed: config.bot_speed,
turn_speed: config.bot_turn_speed,
color: vec![RED, ORANGE, VIOLET, GREEN, PALE],
jump_turn_decrease: config.jump_turn_decrease,
};
let world = World::new(&config);
let camera = Game::new_camera(&config, &world.player);
let state = State {
turn: Turn::None,
sprint: false,
spawn: 0.,
ended: false,
game_speed: 0.,
jump_timeout: 0.,
rotate_cam: false,
bullets: config.bullet_stock,
recharge: 0.,
fps: 0.,
last_frame: Instant::now(),
};
let ellipse = Ellipse {
color: BLACK.alpha(0.).into(),
border: Some(ellipse::Border {
color: RED.alpha(0.5).into(),
radius: 1.,
}),
resolution: 16,
};
Game {
config: config,
world: world,
window: window,
bot_rules: bot_rules,
camera: camera,
state: state,
glyphs: RefCell::new(glyphs),
ellipse: RefCell::new(ellipse),
}
}
fn | <T: Car>(config: &GameConfig, player: &T) -> Camera {
Camera::new(
config.screen_size.clone(),
vec3(0., config.camera_height, -config.camera_distance) + player.pos()
)
}
// Re-calculate fps
fn update_fps(&mut self) {
let d = self.state.last_frame.elapsed();
self.state.last_frame = Instant::now();
self.state.fps = 1. / (d.as_secs() as f64 + 1e-9*d.subsec_nanos() as f64);
}
pub fn run(&mut self) {
while let Some(e) = self.window.next() {
match e {
Input::Press(key) => self.press(key),
Input::Release(key) => self.release(key),
Input::Render(_) => {
self.update_fps();
self.draw(&e);
},
Input::Update(args) => self.update(args.dt),
Input::Move(Motion::MouseRelative(a, b)) => self.mouse_move(a as f64, b as f64),
_ => {}
}
if self.state.ended {
break;
}
}
}
fn mouse_move(&mut self, x: f64, y: f64) {
if self.state.rotate_cam {
self.camera.rotate(x*self.config.mouse_speed, y*self.config.mouse_speed, self.world.player.position);
}
}
fn press(&mut self, key: Button) {
match key {
Button::Keyboard(Key::A) => self.state.turn = Turn::Left,
Button::Keyboard(Key::D) => self.state.turn = Turn::Right,
Button::Keyboard(Key::W) => self.state.sprint = true,
Button::Keyboard(Key::Space) => if self.state.jump_timeout <= 0. {
self.state.jump_timeout = self.config.jump_timeout;
self.world.player.jump();
},
Button::Mouse(MouseButton::Right) => {
if self.config.zoom_in {
self.camera.zoom_in();
}
self.state.rotate_cam = true;
},
Button::Mouse(MouseButton::Left) => if self.state.rotate_cam && self.state.bullets > 0 {
let mut pos = self.world.player.position;
pos.y += self.world.player.size.y;
let mut d = vec3(0., 0., self.config.trueshot_distance + self.config.camera_distance);
d = self.camera.c * d.magnitude2() / d.dot(self.camera.c);
d = self.camera.eye + d - pos;
d = d * self.config.bullet_speed / d.magnitude();
self.world.add_bullet(pos, d, self.config.bullet_len);
self.state.bullets -= 1;
if self.state.bullets <= 0 {
self.state.recharge = self.config.recharge_time;
}
},
_ => (),
}
}
fn release(&mut self, key: Button) {
match key {
Button::Keyboard(Key::A) => if let Turn::Left = self.state.turn {
self.state.turn = Turn::None;
},
Button::Keyboard(Key::D) => if let Turn::Right = self.state.turn {
self.state.turn = Turn::None;
},
Button::Keyboard(Key::W) => self.state.sprint = false,
Button::Mouse(MouseButton::Right) => {
self.state.rotate_cam = false;
self.camera = Game::new_camera(&self.config, &self.world.player);
},
_ => (),
}
}
fn draw(&mut self, e: &Input) {
// Return a horizontal bar
macro_rules! bar {
($curr: expr, $full: expr) => {
[0.,
15.0,
self.config.screen_size.w as f64/2.*$curr/$full,
20.0,]
};
}
let jump_bar = bar!(self.state.jump_timeout, self.config.jump_timeout);
let recharge_bar = bar!(self.state.recharge, self.config.recharge_time);
let bullets_bar = bar!(self.state.bullets as f64, self.config.bullet_stock as f64);
// Closure in `draw_2d` requires unique access to `self`,
// so we use RefCell to hack it.
let mut glyphs = self.glyphs.borrow_mut();
let fps = format!("{:.3}", self.state.fps);
let lines = self.world.render(&self.camera);
self.window.draw_2d(e, |c, g| {
clear(BLACK.into(), g);
for (l, color) in lines {
line(color.into(), 1., convert(l), c.transform, g);
}
rectangle(BLUE.alpha(0.4).into(), jump_bar, c.transform, g);
rectangle(RED.alpha(0.4).into(), recharge_bar, c.transform, g);
rectangle(GREEN.alpha(0.4).into(), bullets_bar, c.transform, g);
text(WHITE.into(), 10, | new_camera | identifier_name |
game.rs | : f64, // Distance between each decoration
pub sprint_factor: f64,
pub spawn_time: (f64, f64),
pub game_sprint: f64, // The increase of game_speed
pub game_max_speed: f64,
pub player_jump_v: f64,
pub player_jump_a: f64,
pub jump_turn_decrease: f64,
pub jump_timeout: f64,
pub mouse_speed: f64,
pub trueshot_distance: f64,
pub bullet_stock: i64, // Number of bullets
pub recharge_time: f64,
pub bullet_len: f64,
pub bullet_speed: f64,
pub zoom_in: bool, // If true, zoom-in while on stare mode
}
impl Game {
pub fn new(config: GameConfig) -> Game {
let mut window: PistonWindow = WindowSettings::new(
config.title.clone(), [config.screen_size.w, config.screen_size.h])
.exit_on_esc(true).build()
.expect("Cannot create window.");
window.set_ups(config.ups);
window.set_max_fps(config.max_fps);
window.set_capture_cursor(true);
let glyphs = Glyphs::new("resources/Ubuntu-R.ttf", window.factory.clone())
.expect("Unable to load font.");
let bot_rules = BoxRules {
size: config.bot_size,
position: [(0., config.tunel_size[0]), (0., 0.), (config.tunel_size[2], config.tunel_size[2])],
speed: config.bot_speed,
turn_speed: config.bot_turn_speed,
color: vec![RED, ORANGE, VIOLET, GREEN, PALE],
jump_turn_decrease: config.jump_turn_decrease,
};
let world = World::new(&config);
let camera = Game::new_camera(&config, &world.player);
let state = State {
turn: Turn::None,
sprint: false,
spawn: 0.,
ended: false,
game_speed: 0.,
jump_timeout: 0.,
rotate_cam: false,
bullets: config.bullet_stock,
recharge: 0.,
fps: 0.,
last_frame: Instant::now(),
};
let ellipse = Ellipse {
color: BLACK.alpha(0.).into(),
border: Some(ellipse::Border {
color: RED.alpha(0.5).into(),
radius: 1.,
}),
resolution: 16,
};
Game {
config: config,
world: world,
window: window,
bot_rules: bot_rules,
camera: camera,
state: state,
glyphs: RefCell::new(glyphs),
ellipse: RefCell::new(ellipse),
}
}
fn new_camera<T: Car>(config: &GameConfig, player: &T) -> Camera {
Camera::new(
config.screen_size.clone(),
vec3(0., config.camera_height, -config.camera_distance) + player.pos()
)
}
// Re-calculate fps
fn update_fps(&mut self) {
let d = self.state.last_frame.elapsed();
self.state.last_frame = Instant::now();
self.state.fps = 1. / (d.as_secs() as f64 + 1e-9*d.subsec_nanos() as f64);
}
pub fn run(&mut self) {
while let Some(e) = self.window.next() {
match e {
Input::Press(key) => self.press(key),
Input::Release(key) => self.release(key),
Input::Render(_) => {
self.update_fps();
self.draw(&e);
},
Input::Update(args) => self.update(args.dt),
Input::Move(Motion::MouseRelative(a, b)) => self.mouse_move(a as f64, b as f64),
_ => {}
}
if self.state.ended {
break;
}
}
}
fn mouse_move(&mut self, x: f64, y: f64) {
if self.state.rotate_cam {
self.camera.rotate(x*self.config.mouse_speed, y*self.config.mouse_speed, self.world.player.position);
}
}
fn press(&mut self, key: Button) {
match key {
Button::Keyboard(Key::A) => self.state.turn = Turn::Left,
Button::Keyboard(Key::D) => self.state.turn = Turn::Right,
Button::Keyboard(Key::W) => self.state.sprint = true,
Button::Keyboard(Key::Space) => if self.state.jump_timeout <= 0. {
self.state.jump_timeout = self.config.jump_timeout;
self.world.player.jump();
},
Button::Mouse(MouseButton::Right) => {
if self.config.zoom_in {
self.camera.zoom_in();
}
self.state.rotate_cam = true;
},
Button::Mouse(MouseButton::Left) => if self.state.rotate_cam && self.state.bullets > 0 {
let mut pos = self.world.player.position;
pos.y += self.world.player.size.y;
let mut d = vec3(0., 0., self.config.trueshot_distance + self.config.camera_distance);
d = self.camera.c * d.magnitude2() / d.dot(self.camera.c);
d = self.camera.eye + d - pos;
d = d * self.config.bullet_speed / d.magnitude();
self.world.add_bullet(pos, d, self.config.bullet_len);
self.state.bullets -= 1;
if self.state.bullets <= 0 {
self.state.recharge = self.config.recharge_time;
}
},
_ => (),
}
}
fn release(&mut self, key: Button) {
match key {
Button::Keyboard(Key::A) => if let Turn::Left = self.state.turn {
self.state.turn = Turn::None;
},
Button::Keyboard(Key::D) => if let Turn::Right = self.state.turn {
self.state.turn = Turn::None;
},
Button::Keyboard(Key::W) => self.state.sprint = false,
Button::Mouse(MouseButton::Right) => {
self.state.rotate_cam = false;
self.camera = Game::new_camera(&self.config, &self.world.player);
},
_ => (),
}
}
fn draw(&mut self, e: &Input) {
// Return a horizontal bar
macro_rules! bar {
($curr: expr, $full: expr) => {
[0.,
15.0,
self.config.screen_size.w as f64/2.*$curr/$full,
20.0,]
};
}
let jump_bar = bar!(self.state.jump_timeout, self.config.jump_timeout);
let recharge_bar = bar!(self.state.recharge, self.config.recharge_time);
let bullets_bar = bar!(self.state.bullets as f64, self.config.bullet_stock as f64);
// Closure in `draw_2d` requires unique access to `self`,
// so we use RefCell to hack it.
let mut glyphs = self.glyphs.borrow_mut();
let fps = format!("{:.3}", self.state.fps);
let lines = self.world.render(&self.camera);
self.window.draw_2d(e, |c, g| {
clear(BLACK.into(), g);
for (l, color) in lines {
line(color.into(), 1., convert(l), c.transform, g);
}
rectangle(BLUE.alpha(0.4).into(), jump_bar, c.transform, g);
rectangle(RED.alpha(0.4).into(), recharge_bar, c.transform, g);
rectangle(GREEN.alpha(0.4).into(), bullets_bar, c.transform, g);
text(WHITE.into(), 10, &fps, glyphs.deref_mut(), c.transform.trans(0., 10.), g);
});
if self.state.rotate_cam {
let w = 20.;
let x = self.config.screen_size.w as f64 /2. - w/2.;
let y = self.config.screen_size.h as f64 /2. - w/2.;
let ellipse = self.ellipse.borrow();
self.window.draw_2d(e, |c, g| {
ellipse.draw([x, y, w, w], &c.draw_state, c.transform, g);
rectangle(RED.into(), [x+w/2.-1., y+w/2.-1., 2., 2.], c.transform, g);
});
}
}
// `dt` stands for delta, duration since the last update
fn update(&mut self, dt: f64) {
// Re-calculate delta according to fps
let dt = if self.state.fps != 0. { 1./self.state.fps}
else { dt };
let old = self.world.player.position;
if self.state.bullets <= 0 {
self.state.recharge -= dt;
if self.state.recharge < 0. {
self.state.bullets = self.config.bullet_stock;
}
}
self.state.jump_timeout -= dt;
if self.state.game_speed < self.config.game_max_speed {
self.state.game_speed += dt*self.config.game_sprint; | }
if self.state.sprint {
if self.world.player.speed < self.config.player_speed.1 { | random_line_split |
|
graph.rs | : usize,
}
pub struct FrameGraph {
node_arena: RwLock<Arena<ConnectedNode>>,
edges_arena: RwLock<Arena<ConnectedEdges>>,
end_node: Index,
output_map: Vec<Vec<Option<FrameNodeValue>>>,
levels_map: MultiMap<usize, TraversedGraphNode>,
traversed_node_cache: HashMap<Index, usize>,
}
impl FrameGraph {
/// Creates a new empty graph.
pub fn new<T: FrameGraphNode>(end_node: T) -> Self {
let mut node_arena = Arena::with_capacity(MAX_LAYERS * MAX_NODES_PER_LAYER);
let end_node = node_arena.insert(ConnectedNode {
name: "EndNode".to_string(),
node: Arc::new(end_node),
inputs: [None; MAX_INPUT_OUTPUTS_PER_NODE],
});
Self {
node_arena: RwLock::new(node_arena),
edges_arena: RwLock::new(Arena::with_capacity(
MAX_LAYERS * MAX_INPUT_OUTPUTS_PER_NODE * MAX_NODES_PER_LAYER,
)),
output_map: vec![vec![None; MAX_NODES_PER_LAYER * MAX_INPUT_OUTPUTS_PER_NODE]; MAX_LAYERS],
levels_map: MultiMap::with_capacity(MAX_LAYERS),
traversed_node_cache: HashMap::with_capacity(
MAX_LAYERS * MAX_INPUT_OUTPUTS_PER_NODE * MAX_NODES_PER_LAYER,
),
end_node,
}
}
/// Returns the end node.
pub fn get_end_node(&self) -> Index {
self.end_node
}
/// Resets the frame graph by removing all nodes and sets up a new end node.
pub fn reset(&mut self) {
let mut nodes = self.node_arena.write();
let end_node_impl = nodes.get(self.end_node).unwrap().node.clone();
nodes.clear();
self.traversed_node_cache.clear();
self.edges_arena.write().clear();
self.end_node = nodes.insert(ConnectedNode {
name: "EndNode".to_string(),
node: end_node_impl,
inputs: [None; MAX_INPUT_OUTPUTS_PER_NODE],
});
}
/// Add a new node into the graph.
pub fn add_node<T: FrameGraphNode>(&self, node: T, name: &str) -> Index {
self.node_arena.write().insert(ConnectedNode {
name: name.to_string(),
node: Arc::new(node),
inputs: [None; MAX_INPUT_OUTPUTS_PER_NODE],
})
}
/// Connects one nodes output to another nodes input.
pub fn connect(
&self,
source: Index,
source_output: usize,
destination: Index,
destination_input: usize,
) -> Result<(), GraphConnectError> {
// Validate connection parameters.
if destination_input >= MAX_INPUT_OUTPUTS_PER_NODE {
return Err(GraphConnectError::MaximumInputsReached);
};
if source_output >= MAX_INPUT_OUTPUTS_PER_NODE {
return Err(GraphConnectError::MaximumInputsReached);
};
let mut edges = self.edges_arena.write();
let mut nodes = self.node_arena.write();
let destination_node = nodes
.get_mut(destination)
.ok_or(GraphConnectError::InvalidDestination)?;
// Target input is already connected.
if destination_node.inputs[destination_input].is_some() {
return Err(GraphConnectError::AlreadyConnected);
}
// Target input is empty so simply create the connection.
let edge = edges.insert(ConnectedEdges {
owner_node_index: source,
output_index: source_output,
});
destination_node.inputs[destination_input] = Some(edge);
Ok(())
}
fn traverse_node(
cache: &mut HashMap<Index, usize>,
levels_map: &mut MultiMap<usize, TraversedGraphNode>,
nodes: &RwLockReadGuard<Arena<ConnectedNode>>,
edges: &RwLockReadGuard<Arena<ConnectedEdges>>,
node_index: Index,
level: usize,
) {
//Build traverse node with input/output mapping info.
let mut traversed_node = TraversedGraphNode {
index: node_index,
inputs: [None; MAX_INPUT_OUTPUTS_PER_NODE],
};
// Remove from dependencies from all levels lower
let mut has_retained = false;
for l in level..0 {
// Remove previous traversed node from level.
let vec = levels_map.get_vec_mut(&l).unwrap();
let before_len = vec.len();
vec.retain(|x| x.index != node_index);
if before_len != vec.len() {
has_retained = true;
}
}
// Update all inputs that still reference kicked out node.
if has_retained {
for l in level..0 {
let vec = levels_map.get_vec_mut(&l).unwrap();
for node in vec {
for input in &mut node.inputs {
if let Some((nlevel, _, index)) = input {
if index == &node_index {
*nlevel = level;
}
}
}
}
}
}
// Loop through all inputs
let next_level = level + 1;
let node = nodes.get(node_index).unwrap();
for (input_index, input) in node.inputs.iter().enumerate() {
if let Some(input) = input {
let edge = edges.get(*input).unwrap();
let inner_node = edge.owner_node_index;
traversed_node.inputs[input_index] = Some((next_level, edge.output_index, inner_node));
Self::traverse_node(cache, levels_map, nodes, edges, inner_node, next_level);
}
}
// Store traversed node at level.
//let traversed_index = levels_map.get_vec(&level).map(|x| x.len()).unwrap_or(0);
//cache.insert(node_index, traversed_index);
// TODO: Due to retaining this index breaks currently :'(
levels_map.insert(level, traversed_node);
}
/// Executes the graph using the given scheduler.
pub fn execute<T: DeviceHost>(
&mut self,
sc_frame: Arc<wgpu::SwapChainFrame>,
device_host: &'static T,
pool: &ThreadPool,
) {
{
{
optick::event!("FrameGraph::traverse");
// Gain read access to nodes and connections.
let nodes = self.node_arena.read();
let edges = self.edges_arena.read();
// Start traversing from end.
self.levels_map.clear();
Self::traverse_node(
&mut self.traversed_node_cache,
&mut self.levels_map,
&nodes,
&edges,
self.end_node,
0,
);
}
let cache = &mut self.traversed_node_cache;
// Create async executer.
let mut local_pool = futures::executor::LocalPool::new();
let local_spawner = local_pool.spawner();
// Execute in levels order
let mut all_levels = self.levels_map.keys().cloned().collect::<Vec<_>>();
all_levels.sort_unstable();
let max_levels = all_levels.len();
for level in all_levels.into_iter().rev() { | optick::tag!("level", level as u32);
// Get rid of duplicated nodes.
let mut nodes_in_level = self.levels_map.get_vec_mut(&level).unwrap().clone();
nodes_in_level.sort_unstable_by_key(|x| x.index);
nodes_in_level.dedup_by_key(|x| x.index);
// Build cache for this level
for (index, node) in nodes_in_level.iter().enumerate() {
cache.insert(node.index, index);
}
// Get chunks
let nodes = self.node_arena.read();
let read_nodes = nodes_in_level
.iter()
.map(|node| (nodes.get(node.index).unwrap(), node.inputs))
.collect::<Vec<_>>();
let mut empty = [Vec::with_capacity(0)];
#[allow(clippy::type_complexity)]
let (outputs, previous_outputs): (
&mut [Vec<Option<FrameNodeValue>>],
&mut [Vec<Option<FrameNodeValue>>],
) = if level == (max_levels - 1) {
(&mut self.output_map, &mut empty)
} else {
self.output_map.split_at_mut(level + 1)
};
let outputs_per_node = outputs[outputs.len() - 1]
.chunks_mut(MAX_INPUT_OUTPUTS_PER_NODE)
.enumerate()
.collect::<Vec<_>>();
// Execute
let encoder_outputs = pool.install(|| {
read_nodes
.par_iter()
.zip(outputs_per_node)
.enumerate()
.map(|(_i, ((node, inputs), (_oi, outputs)))| {
optick::event!("FrameGraph::node");
// Prepare node execution
optick::tag!("name", node.name);
let node_trait = node.node.clone();
let label = format!("NodeCommandEncoder_{}", node.name);
// Map outputs -> inputs.
/*
for (idx, input) in inputs.iter().enumerate() {
if let Some((target_level, output_index, node_index)) = input {
let i = cache.get(&node_index).unwrap();
println!(
"Mapping input #{} to level = {} ({}) and index = {} ({}, {})",
idx,
target_level,
previous_outputs.len() - (target_level - level),
i * MAX_INPUT_OUTPUTS_PER_NODE + output_index | optick::event!("FrameGraph::execute_level"); | random_line_split |
graph.rs | : usize,
}
pub struct | {
node_arena: RwLock<Arena<ConnectedNode>>,
edges_arena: RwLock<Arena<ConnectedEdges>>,
end_node: Index,
output_map: Vec<Vec<Option<FrameNodeValue>>>,
levels_map: MultiMap<usize, TraversedGraphNode>,
traversed_node_cache: HashMap<Index, usize>,
}
impl FrameGraph {
/// Creates a new empty graph.
pub fn new<T: FrameGraphNode>(end_node: T) -> Self {
let mut node_arena = Arena::with_capacity(MAX_LAYERS * MAX_NODES_PER_LAYER);
let end_node = node_arena.insert(ConnectedNode {
name: "EndNode".to_string(),
node: Arc::new(end_node),
inputs: [None; MAX_INPUT_OUTPUTS_PER_NODE],
});
Self {
node_arena: RwLock::new(node_arena),
edges_arena: RwLock::new(Arena::with_capacity(
MAX_LAYERS * MAX_INPUT_OUTPUTS_PER_NODE * MAX_NODES_PER_LAYER,
)),
output_map: vec![vec![None; MAX_NODES_PER_LAYER * MAX_INPUT_OUTPUTS_PER_NODE]; MAX_LAYERS],
levels_map: MultiMap::with_capacity(MAX_LAYERS),
traversed_node_cache: HashMap::with_capacity(
MAX_LAYERS * MAX_INPUT_OUTPUTS_PER_NODE * MAX_NODES_PER_LAYER,
),
end_node,
}
}
/// Returns the end node.
pub fn get_end_node(&self) -> Index {
self.end_node
}
/// Resets the frame graph by removing all nodes and sets up a new end node.
pub fn reset(&mut self) {
let mut nodes = self.node_arena.write();
let end_node_impl = nodes.get(self.end_node).unwrap().node.clone();
nodes.clear();
self.traversed_node_cache.clear();
self.edges_arena.write().clear();
self.end_node = nodes.insert(ConnectedNode {
name: "EndNode".to_string(),
node: end_node_impl,
inputs: [None; MAX_INPUT_OUTPUTS_PER_NODE],
});
}
/// Add a new node into the graph.
pub fn add_node<T: FrameGraphNode>(&self, node: T, name: &str) -> Index {
self.node_arena.write().insert(ConnectedNode {
name: name.to_string(),
node: Arc::new(node),
inputs: [None; MAX_INPUT_OUTPUTS_PER_NODE],
})
}
/// Connects one nodes output to another nodes input.
pub fn connect(
&self,
source: Index,
source_output: usize,
destination: Index,
destination_input: usize,
) -> Result<(), GraphConnectError> {
// Validate connection parameters.
if destination_input >= MAX_INPUT_OUTPUTS_PER_NODE {
return Err(GraphConnectError::MaximumInputsReached);
};
if source_output >= MAX_INPUT_OUTPUTS_PER_NODE {
return Err(GraphConnectError::MaximumInputsReached);
};
let mut edges = self.edges_arena.write();
let mut nodes = self.node_arena.write();
let destination_node = nodes
.get_mut(destination)
.ok_or(GraphConnectError::InvalidDestination)?;
// Target input is already connected.
if destination_node.inputs[destination_input].is_some() {
return Err(GraphConnectError::AlreadyConnected);
}
// Target input is empty so simply create the connection.
let edge = edges.insert(ConnectedEdges {
owner_node_index: source,
output_index: source_output,
});
destination_node.inputs[destination_input] = Some(edge);
Ok(())
}
fn traverse_node(
cache: &mut HashMap<Index, usize>,
levels_map: &mut MultiMap<usize, TraversedGraphNode>,
nodes: &RwLockReadGuard<Arena<ConnectedNode>>,
edges: &RwLockReadGuard<Arena<ConnectedEdges>>,
node_index: Index,
level: usize,
) {
//Build traverse node with input/output mapping info.
let mut traversed_node = TraversedGraphNode {
index: node_index,
inputs: [None; MAX_INPUT_OUTPUTS_PER_NODE],
};
// Remove from dependencies from all levels lower
let mut has_retained = false;
for l in level..0 {
// Remove previous traversed node from level.
let vec = levels_map.get_vec_mut(&l).unwrap();
let before_len = vec.len();
vec.retain(|x| x.index != node_index);
if before_len != vec.len() {
has_retained = true;
}
}
// Update all inputs that still reference kicked out node.
if has_retained {
for l in level..0 {
let vec = levels_map.get_vec_mut(&l).unwrap();
for node in vec {
for input in &mut node.inputs {
if let Some((nlevel, _, index)) = input {
if index == &node_index {
*nlevel = level;
}
}
}
}
}
}
// Loop through all inputs
let next_level = level + 1;
let node = nodes.get(node_index).unwrap();
for (input_index, input) in node.inputs.iter().enumerate() {
if let Some(input) = input {
let edge = edges.get(*input).unwrap();
let inner_node = edge.owner_node_index;
traversed_node.inputs[input_index] = Some((next_level, edge.output_index, inner_node));
Self::traverse_node(cache, levels_map, nodes, edges, inner_node, next_level);
}
}
// Store traversed node at level.
//let traversed_index = levels_map.get_vec(&level).map(|x| x.len()).unwrap_or(0);
//cache.insert(node_index, traversed_index);
// TODO: Due to retaining this index breaks currently :'(
levels_map.insert(level, traversed_node);
}
/// Executes the graph using the given scheduler.
pub fn execute<T: DeviceHost>(
&mut self,
sc_frame: Arc<wgpu::SwapChainFrame>,
device_host: &'static T,
pool: &ThreadPool,
) {
{
{
optick::event!("FrameGraph::traverse");
// Gain read access to nodes and connections.
let nodes = self.node_arena.read();
let edges = self.edges_arena.read();
// Start traversing from end.
self.levels_map.clear();
Self::traverse_node(
&mut self.traversed_node_cache,
&mut self.levels_map,
&nodes,
&edges,
self.end_node,
0,
);
}
let cache = &mut self.traversed_node_cache;
// Create async executer.
let mut local_pool = futures::executor::LocalPool::new();
let local_spawner = local_pool.spawner();
// Execute in levels order
let mut all_levels = self.levels_map.keys().cloned().collect::<Vec<_>>();
all_levels.sort_unstable();
let max_levels = all_levels.len();
for level in all_levels.into_iter().rev() {
optick::event!("FrameGraph::execute_level");
optick::tag!("level", level as u32);
// Get rid of duplicated nodes.
let mut nodes_in_level = self.levels_map.get_vec_mut(&level).unwrap().clone();
nodes_in_level.sort_unstable_by_key(|x| x.index);
nodes_in_level.dedup_by_key(|x| x.index);
// Build cache for this level
for (index, node) in nodes_in_level.iter().enumerate() {
cache.insert(node.index, index);
}
// Get chunks
let nodes = self.node_arena.read();
let read_nodes = nodes_in_level
.iter()
.map(|node| (nodes.get(node.index).unwrap(), node.inputs))
.collect::<Vec<_>>();
let mut empty = [Vec::with_capacity(0)];
#[allow(clippy::type_complexity)]
let (outputs, previous_outputs): (
&mut [Vec<Option<FrameNodeValue>>],
&mut [Vec<Option<FrameNodeValue>>],
) = if level == (max_levels - 1) {
(&mut self.output_map, &mut empty)
} else {
self.output_map.split_at_mut(level + 1)
};
let outputs_per_node = outputs[outputs.len() - 1]
.chunks_mut(MAX_INPUT_OUTPUTS_PER_NODE)
.enumerate()
.collect::<Vec<_>>();
// Execute
let encoder_outputs = pool.install(|| {
read_nodes
.par_iter()
.zip(outputs_per_node)
.enumerate()
.map(|(_i, ((node, inputs), (_oi, outputs)))| {
optick::event!("FrameGraph::node");
// Prepare node execution
optick::tag!("name", node.name);
let node_trait = node.node.clone();
let label = format!("NodeCommandEncoder_{}", node.name);
// Map outputs -> inputs.
/*
for (idx, input) in inputs.iter().enumerate() {
if let Some((target_level, output_index, node_index)) = input {
let i = cache.get(&node_index).unwrap();
println!(
"Mapping input #{} to level = {} ({}) and index = {} ({}, {})",
idx,
target_level,
previous_outputs.len() - (target_level - level),
i * MAX_INPUT_OUTPUTS_PER_NODE + output | FrameGraph | identifier_name |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.