text
stringlengths
57
2.88k
labels
sequencelengths
6
6
Title: A Language for Function Signature Representations, Abstract: Recent work by (Richardson and Kuhn, 2017a,b; Richardson et al., 2018) looks at semantic parser induction and question answering in the domain of source code libraries and APIs. In this brief note, we formalize the representations being learned in these studies and introduce a simple domain specific language and a systematic translation from this language to first-order logic. By recasting the target representations in terms of classical logic, we aim to broaden the applicability of existing code datasets for investigating more complex natural language understanding and reasoning problems in the software domain.
[ 1, 0, 0, 0, 0, 0 ]
Title: Link Adaptation for Wireless Video Communication Systems, Abstract: This PhD thesis considers the performance evaluation and enhancement of video communication over wireless channels. The system model considers hybrid automatic repeat request (HARQ) with Chase combining and turbo product codes (TPC). The thesis proposes algorithms and techniques to optimize the throughput, transmission power and complexity of HARQ-based wireless video communication. A semi-analytical solution is developed to model the performance of delay-constrained HARQ systems. The semi-analytical and Monte Carlo simulation results reveal that significant complexity reduction can be achieved by noting that the coding gain advantage of the soft over hard decoding is reduced when Chase combining is used, and it actually vanishes completely for particular codes. Moreover, the thesis proposes a novel power optimization algorithm that achieves a significant power saving of up to 80%. Joint throughput maximization and complexity reduction is considered as well. A CRC (cyclic redundancy check)-free HARQ is proposed to improve the system throughput when short packets are transmitted. In addition, the computational complexity/delay is reduced when the packets transmitted are long. Finally, a content-aware and occupancy-based HARQ scheme is proposed to ensure minimum video quality distortion with continuous playback.
[ 1, 0, 0, 0, 0, 0 ]
Title: On the Complexity of Polytopes in $LI(2)$, Abstract: In this paper we consider polytopes given by systems of $n$ inequalities in $d$ variables, where every inequality has at most two variables with nonzero coefficient. We denote this family by $LI(2)$. We show that despite of the easy algebraic structure, polytopes in $LI(2)$ can have high complexity. We construct a polytope in $LI(2)$, whose number of vertices is almost the number of vertices of the dual cyclic polytope, the difference is a multiplicative factor of depending on $d$ and in particular independent of $n$. Moreover we show that the dual cyclic polytope can not be realized in $LI(2)$.
[ 1, 0, 0, 0, 0, 0 ]
Title: Dynamic Sensitivity Study of MEMS Capacitive Acceleration Transducer Based on Analytical Squeeze Film Damping and Mechanical Thermoelasticity Approaches, Abstract: The dynamic behavior of a capacitive micro-electro-mechanical (MEMS) accelerometer is evaluated by using a theoretical approach which makes use of a squeeze film damping (SFD) model and ideal gas approach. The study investigates the performance of the device as a function of the temperature, from 228 K to 398 K, and pressure, from 20 to 1000 Pa, observing the damping gas trapped inside de mechanical transducer. Thermoelastic properties of the silicon bulk are considered for the entire range of temperature. The damping gases considered are Air, Helium and Argon. The global behavior of the system is evaluated considering the electro-mechanical sensitivity (SEM) as the main figure of merit in frequency domain. The results show the behavior of the main mechanism losses of SFD, as well as the dynamic sensitivity of the MEMS transducer system, and are in good agreement with experimental dynamic results behavior.
[ 0, 1, 0, 0, 0, 0 ]
Title: Nature of the electromagnetic force between classical magnetic dipoles, Abstract: The Lorentz force law of classical electrodynamics states that the force F exerted by the magnetic induction B on a particle of charge q moving with velocity V is given by F=qVxB. Since this force is orthogonal to the direction of motion, the magnetic field is said to be incapable of performing mechanical work. Yet there is no denying that a permanent magnet can readily perform mechanical work by pushing/pulling on another permanent magnet -- or by attracting pieces of magnetizable material such as scrap iron or iron filings. We explain this apparent contradiction by examining the magnetic Lorentz force acting on an Amperian current loop, which is the model for a magnetic dipole. We then extend the discussion by analyzing the Einstein-Laub model of magnetic dipoles in the presence of external magnetic fields.
[ 0, 1, 0, 0, 0, 0 ]
Title: Tractability of $\mathbb{L}_2$-approximation in hybrid function spaces, Abstract: We consider multivariate $\mathbb{L}_2$-approximation in reproducing kernel Hilbert spaces which are tensor products of weighted Walsh spaces and weighted Korobov spaces. We study the minimal worst-case error $e^{\mathbb{L}_2-\mathrm{app},\Lambda}(N,d)$ of all algorithms that use $N$ information evaluations from the class $\Lambda$ in the $d$-dimensional case. The two classes $\Lambda$ considered in this paper are the class $\Lambda^{\rm all}$ consisting of all linear functionals and the class $\Lambda^{\rm std}$ consisting only of function evaluations. The focus lies on the dependence of $e^{\mathbb{L}_2-\mathrm{app},\Lambda}(N,d)$ on the dimension $d$. The main results are conditions for weak, polynomial, and strong polynomial tractability.
[ 0, 0, 1, 0, 0, 0 ]
Title: The Linearity of the Mitchell Order, Abstract: We show from a weak comparison principle (the Ultrapower Axiom) that the Mitchell order is linear on certain kinds of ultrafilters: normal ultrafilters, Dodd solid ultrafilters, and assuming GCH, generalized normal ultrafilters. In the process we prove a generalization of Solovay's lemma to singular cardinals.
[ 0, 0, 1, 0, 0, 0 ]
Title: Atomic Order in Non-Equilibrium Silicon-Germanium-Tin Semiconductors, Abstract: The precise knowledge of the atomic order in monocrystalline alloys is fundamental to understand and predict their physical properties. With this perspective, we utilized laser-assisted atom probe tomography to investigate the three-dimensional distribution of atoms in non-equilibrium epitaxial Sn-rich group IV SiGeSn ternary semiconductors. Different atom probe statistical analysis tools including frequency distribution analysis, partial radial distribution functions, and nearest neighbor analysis were employed in order to evaluate and compare the behavior of the three elements to their spatial distributions in an ideal solid solution. This atomistic-level analysis provided clear evidence of an unexpected repulsive interaction between Sn and Si leading to the deviation of Si atoms from the theoretical random distribution. This departure from an ideal solid solution is supported by first principal calculations and attributed to the tendency of the system to reduce its mixing enthalpy throughout the layer-by-layer growth process.
[ 0, 1, 0, 0, 0, 0 ]
Title: On the Laws of Large Numbers in Possibility Theory, Abstract: In this paper we obtain some possibilistic variants of the probabilistic laws of large numbers, different from those obtained by other authors, but very natural extensions of the corresponding ones in probability theory. Our results are based on the possibility measure and on the "maxitive" definitions for possibility expectation and possibility variance. Also, we show that in this frame, the weak form of the law of large numbers, implies the strong law of large numbers.
[ 0, 0, 1, 0, 0, 0 ]
Title: Finite Sample Inference for Targeted Learning, Abstract: The Highly-Adaptive-Lasso(HAL)-TMLE is an efficient estimator of a pathwise differentiable parameter in a statistical model that at minimal (and possibly only) assumes that the sectional variation norm of the true nuisance parameters are finite. It relies on an initial estimator (HAL-MLE) of the nuisance parameters by minimizing the empirical risk over the parameter space under the constraint that sectional variation norm is bounded by a constant, where this constant can be selected with cross-validation. In the formulation of the HALMLE this sectional variation norm corresponds with the sum of absolute value of coefficients for an indicator basis. Due to its reliance on machine learning, statistical inference for the TMLE has been based on its normal limit distribution, thereby potentially ignoring a large second order remainder in finite samples. In this article, we present four methods for construction of a finite sample 0.95-confidence interval that use the nonparametric bootstrap to estimate the finite sample distribution of the HAL-TMLE or a conservative distribution dominating the true finite sample distribution. We prove that it consistently estimates the optimal normal limit distribution, while its approximation error is driven by the performance of the bootstrap for a well behaved empirical process. We demonstrate our general inferential methods for 1) nonparametric estimation of the average treatment effect based on observing on each unit a covariate vector, binary treatment, and outcome, and for 2) nonparametric estimation of the integral of the square of the multivariate density of the data distribution.
[ 0, 0, 1, 1, 0, 0 ]
Title: A Short and Elementary Proof of the Two-sidedness of the Matrix-Inverse, Abstract: An elementary proof of the two-sidedness of the matrix-inverse is given using only linear independence and the reduced row-echelon form of a matrix. In addition, it is shown that a matrix is invertible if and only if it is row-equivalent to the identity matrix without appealing to elementary matrices. This proof underscores the importance of a basis and provides a proof of the invertible matrix theorem.
[ 0, 0, 1, 0, 0, 0 ]
Title: Rich-Club Ordering and the Dyadic Effect: Two Interrelated Phenomena, Abstract: Rich-club ordering and the dyadic effect are two phenomena observed in complex networks that are based on the presence of certain substructures composed of specific nodes. Rich-club ordering represents the tendency of highly connected and important elements to form tight communities with other central elements. The dyadic effect denotes the tendency of nodes that share a common property to be much more interconnected than expected. In this study, we consider the interrelation between these two phenomena, which until now have always been studied separately. We contribute with a new formulation of the rich-club measures in terms of the dyadic effect. Moreover, we introduce certain measures related to the analysis of the dyadic effect, which are useful in confirming the presence and relevance of rich-clubs in complex networks. In addition, certain computational experiences show the usefulness of the introduced quantities with regard to different classes of real networks.
[ 1, 1, 0, 0, 0, 0 ]
Title: The global dust modelling framework THEMIS (The Heterogeneous dust Evolution Model for Interstellar Solids), Abstract: Here we introduce the interstellar dust modelling framework THEMIS (The Heterogeneous dust Evolution Model for Interstellar Solids), which takes a global view of dust and its evolution in response to the local conditions in interstellar media. This approach is built upon a core model that was developed to explain the dust extinction and emission in the diffuse interstellar medium. The model was then further developed to self-consistently include the effects of dust evolution in the transition to denser regions. The THEMIS approach is under continuous development and currently we are extending the framework to explore the implications of dust evolution in HII regions and the photon-dominated regions associated with star formation. We provide links to the THEMIS, DustEM and DustPedia websites where more information about the model, its input data and applications can be found.
[ 0, 1, 0, 0, 0, 0 ]
Title: Variable Selection Methods for Model-based Clustering, Abstract: Model-based clustering is a popular approach for clustering multivariate data which has seen applications in numerous fields. Nowadays, high-dimensional data are more and more common and the model-based clustering approach has adapted to deal with the increasing dimensionality. In particular, the development of variable selection techniques has received a lot of attention and research effort in recent years. Even for small size problems, variable selection has been advocated to facilitate the interpretation of the clustering results. This review provides a summary of the methods developed for variable selection in model-based clustering. Existing R packages implementing the different methods are indicated and illustrated in application to two data analysis examples.
[ 0, 0, 0, 1, 0, 0 ]
Title: Synergies between Asteroseismology and Exoplanetary Science, Abstract: Over the past decade asteroseismology has become a powerful method to systematically characterize host stars and dynamical architectures of exoplanet systems. In this contribution I review current key synergies between asteroseismology and exoplanetary science such as the precise determination of planet radii and ages, the measurement of orbital eccentricities, stellar obliquities and their impact on hot Jupiter formation theories, and the importance of asteroseismology on spectroscopic analyses of exoplanet hosts. I also give an outlook on future synergies such as the characterization of sub-Neptune-size planets orbiting solar-type stars, the study of planet populations orbiting evolved stars, and the determination of ages of intermediate-mass stars hosting directly imaged planets.
[ 0, 1, 0, 0, 0, 0 ]
Title: The stellar contents and star formation in the NGC 7538 region, Abstract: Deep optical photometric data on the NGC 7538 region were collected and combined with archival data sets from $Chandra$, 2MASS and {\it Spitzer} surveys in order to generate a new catalog of young stellar objects (YSOs) including those not showing IR excess emission. This new catalog is complete down to 0.8 M$_\odot$. The nature of the YSOs associated with the NGC 7538 region and their spatial distribution are used to study the star formation process and the resultant mass function (MF) in the region. Out of the 419 YSOs, $\sim$91\% have ages between 0.1 to 2.5 Myr and $\sim$86\% have masses between 0.5 to 3.5 M$_\odot$, as derived by spectral energy distribution fitting analysis. Around 24\%, 62\% and 2\% of these YSOs are classified to be the Class I, Class II and Class III sources, respectively. The X-ray activity in the Class I, Class II and Class III objects is not significantly different from each other. This result implies that the enhanced X-ray surface flux due to the increase in the rotation rate may be compensated by the decrease in the stellar surface area during the pre-main sequence evolution. Our analysis shows that the O3V type high mass star `IRS 6' might have triggered the formation of young low mass stars up to a radial distance of 3 pc. The MF shows a turn-off at around 1.5 M$_\odot$ and the value of its slope `$\Gamma$' in the mass range $1.5 <$M/M$_\odot < 6$ comes out to be $-1.76\pm0.24$, which is steeper than the Salpeter value.
[ 0, 1, 0, 0, 0, 0 ]
Title: Ambiguity set and learning via Bregman and Wasserstein, Abstract: Construction of ambiguity set in robust optimization relies on the choice of divergences between probability distributions. In distribution learning, choosing appropriate probability distributions based on observed data is critical for approximating the true distribution. To improve the performance of machine learning models, there has recently been interest in designing objective functions based on Lp-Wasserstein distance rather than the classical Kullback-Leibler (KL) divergence. In this paper, we derive concentration and asymptotic results using Bregman divergence. We propose a novel asymmetric statistical divergence called Wasserstein-Bregman divergence as a generalization of L2-Wasserstein distance. We discuss how these results can be applied to the construction of ambiguity set in robust optimization.
[ 1, 0, 0, 1, 0, 0 ]
Title: 1-bit Massive MU-MIMO Precoding in VLSI, Abstract: Massive multiuser (MU) multiple-input multiple-output (MIMO) will be a core technology in fifth-generation (5G) wireless systems as it offers significant improvements in spectral efficiency compared to existing multi-antenna technologies. The presence of hundreds of antenna elements at the base station (BS), however, results in excessively high hardware costs and power consumption, and requires high interconnect throughput between the baseband-processing unit and the radio unit. Massive MU-MIMO that uses low-resolution analog-to-digital and digital-to-analog converters (DACs) has the potential to address all these issues. In this paper, we focus on downlink precoding for massive MU-MIMO systems with 1-bit DACs at the BS. The objective is to design precoders that simultaneously mitigate multi-user interference (MUI) and quantization artifacts. We propose two nonlinear 1-bit precoding algorithms and corresponding very-large scale integration (VLSI) designs. Our algorithms rely on biconvex relaxation, which enables the design of efficient 1-bit precoding algorithms that achieve superior error-rate performance compared to that of linear precoding algorithms followed by quantization. To showcase the efficacy of our algorithms, we design VLSI architectures that enable efficient 1-bit precoding for massive MU-MIMO systems in which hundreds of antennas serve tens of user equipments. We present corresponding field-programmable gate array (FPGA) implementations to demonstrate that 1-bit precoding enables reliable and high-rate downlink data transmission in practical systems.
[ 1, 0, 0, 0, 0, 0 ]
Title: Step evolution in two-dimensional diblock copolymer films, Abstract: The formation and dynamics of free-surface structures, such as steps or terraces and their interplay with the phase separation in the bulk are key features of diblock copolymer films. We present a phase-field model with an obstacle potential which follows naturally from derivations of the Ohta-Kawasaki energy functional via self-consistent field theory. The free surface of the film is incorporated into the phase-field model by including a third phase for the void. The resulting model and its sharp interface limit are shown to capture the energetics of films with steps in two dimensions. For this model, we then develop a numerical approach that is capable of resolving the long-time complex free-surface structures that arise in diblock copolymer films.
[ 0, 1, 0, 0, 0, 0 ]
Title: The univalence axiom in cubical sets, Abstract: In this note we show that Voevodsky's univalence axiom holds in the model of type theory based on symmetric cubical sets. We will also discuss Swan's construction of the identity type in this variation of cubical sets. This proves that we have a model of type theory supporting dependent products, dependent sums, univalent universes, and identity types with the usual judgmental equality, and this model is formulated in a constructive metatheory.
[ 1, 0, 1, 0, 0, 0 ]
Title: Assessing the Privacy Cost in Centralized Event-Based Demand Response for Microgrids, Abstract: Demand response (DR) programs have emerged as a potential key enabling ingredient in the context of smart grid (SG). Nevertheless, the rising concerns over privacy issues raised by customers subscribed to these programs constitute a major threat towards their effective deployment and utilization. This has driven extensive research to resolve the hindrance confronted, resulting in a number of methods being proposed for preserving customers' privacy. While these methods provide stringent privacy guarantees, only limited attention has been paid to their computational efficiency and performance quality. Under the paradigm of differential privacy, this paper initiates a systematic empirical study on quantifying the trade-off between privacy and optimality in centralized DR systems for maximizing cumulative customer utility. Aiming to elucidate the factors governing this trade-off, we analyze the cost of privacy in terms of the effect incurred on the objective value of the DR optimization problem when applying the employed privacy-preserving strategy based on Laplace mechanism. The theoretical results derived from the analysis are complemented with empirical findings, corroborated extensively by simulations on a 4-bus MG system with up to thousands of customers. By evaluating the impact of privacy, this pilot study serves DR practitioners when considering the social and economic implications of deploying privacy-preserving DR programs in practice. Moreover, it stimulates further research on exploring more efficient approaches with bounded performance guarantees for optimizing energy procurement of MGs without infringing the privacy of customers on demand side.
[ 1, 0, 1, 0, 0, 0 ]
Title: Threat Modeling Data Analysis in Socio-technical Systems, Abstract: Our decision-making processes are becoming more data driven, based on data from multiple sources, of different types, processed by a variety of technologies. As technology becomes more relevant for decision processes, the more likely they are to be subjects of attacks aimed at disrupting their execution or changing their outcome. With the increasing complexity and dependencies on technical components, such attempts grow more sophisticated and their impact will be more severe. This is especially important in scenarios with shared goals, which had to be previously agreed to, or decisions with broad social impact. We need to think about our decisions-making and underlying data analysis processes in a systemic way to correctly evaluate benefits and risks of specific solutions and to design them to be resistant to attacks. To reach these goals, we can apply experiences from threat modeling analysis used in software security. We will need to adapt these practices to new types of threats, protecting different assets and operating in socio-technical systems. With these changes, threat modeling can become a foundation for implementing detailed technical, organizational or legal mitigations and making our decisions more reliable and trustworthy.
[ 1, 0, 0, 0, 0, 0 ]
Title: Nonlinear Instability of Half-Solitons on Star Graphs, Abstract: We consider a half-soliton stationary state of the nonlinear Schrodinger equation with the power nonlinearity on a star graph consisting of N edges and a single vertex. For the subcritical power nonlinearity, the half-soliton state is a degenerate critical point of the action functional under the mass constraint such that the second variation is nonnegative. By using normal forms, we prove that the degenerate critical point is a nonlinear saddle point, for which the small perturbations to the half-soliton state grow slowly in time resulting in the nonlinear instability of the half-soliton state. The result holds for any $N \geq 3$ and arbitrary subcritical power nonlinearity. It gives a precise dynamical characterization of the previous result of Adami {\em et al.}, where the half-soliton state was shown to be a saddle point of the action functional under the mass constraint for $N = 3$ and for cubic nonlinearity.
[ 0, 1, 1, 0, 0, 0 ]
Title: Demographics and discussion influence views on algorithmic fairness, Abstract: The field of algorithmic fairness has highlighted ethical questions which may not have purely technical answers. For example, different algorithmic fairness constraints are often impossible to satisfy simultaneously, and choosing between them requires value judgments about which people may disagree. Achieving consensus on algorithmic fairness will be difficult unless we understand why people disagree in the first place. Here we use a series of surveys to investigate how two factors affect disagreement: demographics and discussion. First, we study whether disagreement on algorithmic fairness questions is caused partially by differences in demographic backgrounds. This is a question of interest because computer science is demographically non-representative. If beliefs about algorithmic fairness correlate with demographics, and algorithm designers are demographically non-representative, decisions made about algorithmic fairness may not reflect the will of the population as a whole. We show, using surveys of three separate populations, that there are gender differences in beliefs about algorithmic fairness. For example, women are less likely to favor including gender as a feature in an algorithm which recommends courses to students if doing so would make female students less likely to be recommended science courses. Second, we investigate whether people's views on algorithmic fairness can be changed by discussion and show, using longitudinal surveys of students in two computer science classes, that they can.
[ 1, 0, 0, 0, 0, 0 ]
Title: The stifness of the supranuclear equation of state (once again), Abstract: We revisit the present status of the stiffness of the supranuclear equations of state, particularly the solutions that increase the stiffness in the presence of hyperons, the putative transition to a quark matter phase and the robustness of massive compact star observations.
[ 0, 1, 0, 0, 0, 0 ]
Title: Stable Desynchronization for Wireless Sensor Networks: (III) Stability Analysis, Abstract: In this paper, we use dynamical systems to analyze stability of desynchronization algorithms at equilibrium. We start by illustrating the equilibrium of a dynamic systems and formalizing force components and time phases. Then, we use Linear Approximation to obtain Jaconian (J) matrixes which are used to find the eigenvalues. Next, we employ the Hirst and Macey theorem and Gershgorins theorem to find the bounds of those eigenvalues. Finally, if the number of nodes (n) is within such bounds, the systems are stable at equilibrium. (This paper is the last part of the series Stable Desynchronization for Wireless Sensor Networks - (I) Concepts and Algorithms (II) Performance Evaluation (III) Stability Analysis)
[ 1, 0, 0, 0, 0, 0 ]
Title: Optimizing Adiabatic Quantum Program Compilation using a Graph-Theoretic Framework, Abstract: Adiabatic quantum computing has evolved in recent years from a theoretical field into an immensely practical area, a change partially sparked by D-Wave System's quantum annealing hardware. These multimillion-dollar quantum annealers offer the potential to solve optimization problems millions of times faster than classical heuristics, prompting researchers at Google, NASA and Lockheed Martin to study how these computers can be applied to complex real-world problems such as NASA rover missions. Unfortunately, compiling (embedding) an optimization problem into the annealing hardware is itself a difficult optimization problem and a major bottleneck currently preventing widespread adoption. Additionally, while finding a single embedding is difficult, no generalized method is known for tuning embeddings to use minimal hardware resources. To address these barriers, we introduce a graph-theoretic framework for developing structured embedding algorithms. Using this framework, we introduce a biclique virtual hardware layer to provide a simplified interface to the physical hardware. Additionally, we exploit bipartite structure in quantum programs using odd cycle transversal (OCT) decompositions. By coupling an OCT-based embedding algorithm with new, generalized reduction methods, we develop a new baseline for embedding a wide range of optimization problems into fault-free D-Wave annealing hardware. To encourage the reuse and extension of these techniques, we provide an implementation of the framework and embedding algorithms.
[ 1, 0, 0, 0, 0, 0 ]
Title: Analysis of pedestrian behaviors through non-invasive Bluetooth monitoring, Abstract: This paper analyzes pedestrians' behavioral patterns in the pedestrianized shopping environment in the historical center of Barcelona, Spain. We employ a Bluetooth detection technique to capture a large-scale dataset of pedestrians' behavior over a one-month period, including during a key sales period. We focused on comparing particular behaviors before, during, and after the discount sales by analyzing this large-scale dataset, which is different but complementary to the conventionally used small-scale samples. Our results uncover pedestrians actively exploring a wider area of the district during a discount period compared to weekdays, giving rise to strong underlying mobility patterns.
[ 1, 1, 0, 0, 0, 0 ]
Title: Open Gromov-Witten theory without Obstruction, Abstract: We define Open Gromov-Witten invariants counting psudoholomorphic curves with boundary of a fixed Euler characteristic. There is not obstruction in the construction of the invariant.
[ 0, 0, 1, 0, 0, 0 ]
Title: Resilience of Complex Networks, Abstract: This article determines and characterizes the minimal number of actuators needed to ensure structural controllability of a linear system under structural alterations that can severe the connection between any two states. We assume that initially the system is structurally controllable with respect to a given set of controls, and propose an efficient system-synthesis mechanism to find the minimal number of additional actuators required for resilience of the system w.r.t such structural changes. The effectiveness of this approach is demonstrated by using standard IEEE power networks.
[ 1, 0, 1, 0, 0, 0 ]
Title: Forward Thinking: Building and Training Neural Networks One Layer at a Time, Abstract: We present a general framework for training deep neural networks without backpropagation. This substantially decreases training time and also allows for construction of deep networks with many sorts of learners, including networks whose layers are defined by functions that are not easily differentiated, like decision trees. The main idea is that layers can be trained one at a time, and once they are trained, the input data are mapped forward through the layer to create a new learning problem. The process is repeated, transforming the data through multiple layers, one at a time, rendering a new data set, which is expected to be better behaved, and on which a final output layer can achieve good performance. We call this forward thinking and demonstrate a proof of concept by achieving state-of-the-art accuracy on the MNIST dataset for convolutional neural networks. We also provide a general mathematical formulation of forward thinking that allows for other types of deep learning problems to be considered.
[ 1, 0, 0, 1, 0, 0 ]
Title: Non-negative Tensor Factorization for Human Behavioral Pattern Mining in Online Games, Abstract: Multiplayer online battle arena has become a popular game genre. It also received increasing attention from our research community because they provide a wealth of information about human interactions and behaviors. A major problem is extracting meaningful patterns of activity from this type of data, in a way that is also easy to interpret. Here, we propose to exploit tensor decomposition techniques, and in particular Non-negative Tensor Factorization, to discover hidden correlated behavioral patterns of play in a popular game: League of Legends. We first collect the entire gaming history of a group of about one thousand players, totaling roughly $100K$ matches. By applying our methodological framework, we then separate players into groups that exhibit similar features and playing strategies, as well as similar temporal trajectories, i.e., behavioral progressions over the course of their gaming history: this will allow us to investigate how players learn and improve their skills.
[ 1, 1, 0, 0, 0, 0 ]
Title: Graph-Cut RANSAC, Abstract: A novel method for robust estimation, called Graph-Cut RANSAC, GC-RANSAC in short, is introduced. To separate inliers and outliers, it runs the graph-cut algorithm in the local optimization (LO) step which is applied when a so-far-the-best model is found. The proposed LO step is conceptually simple, easy to implement, globally optimal and efficient. GC-RANSAC is shown experimentally, both on synthesized tests and real image pairs, to be more geometrically accurate than state-of-the-art methods on a range of problems, e.g. line fitting, homography, affine transformation, fundamental and essential matrix estimation. It runs in real-time for many problems at a speed approximately equal to that of the less accurate alternatives (in milliseconds on standard CPU).
[ 1, 0, 0, 0, 0, 0 ]
Title: Mixed measurements and the detection of phase synchronization in networks, Abstract: Multivariate singular spectrum analysis (M-SSA), with a varimax rotation of eigenvectors, was recently proposed to provide detailed information about phase synchronization in networks of nonlinear oscillators without any a priori need for phase estimation. The discriminatory power of M-SSA is often enhanced by using only the time series of the variable that provides the best observability of the node dynamics. In practice, however, diverse factors could prevent one to have access to this variable in some nodes and other variables should be used, resulting in a mixed set of variables. In the present work, the impact of this mixed measurement approach on the M-SSA is numerically investigated in networks of Rössler systems and cord oscillators. The results are threefold. First, a node measured by a poor variable, in terms of observability, becomes virtually invisible to the technique. Second, a side effect of using a poor variable is that the characterization of phase synchronization clustering of the {\it other}\, nodes is hindered by a small amount. This suggests that, given a network, synchronization analysis with M-SSA could be more reliable by not measuring those nodes that are accessible only through poor variables. Third, global phase synchronization could be detected even using only poor variables, given enough of them are measured. These insights could be useful in defining measurement strategies for both experimental design and real world applications for use with M-SSA.
[ 0, 1, 0, 0, 0, 0 ]
Title: A Unified Neural Network Approach for Estimating Travel Time and Distance for a Taxi Trip, Abstract: In building intelligent transportation systems such as taxi or rideshare services, accurate prediction of travel time and distance is crucial for customer experience and resource management. Using the NYC taxi dataset, which contains taxi trips data collected from GPS-enabled taxis [23], this paper investigates the use of deep neural networks to jointly predict taxi trip time and distance. We propose a model, called ST-NN (Spatio-Temporal Neural Network), which first predicts the travel distance between an origin and a destination GPS coordinate, then combines this prediction with the time of day to predict the travel time. The beauty of ST-NN is that it uses only the raw trips data without requiring further feature engineering and provides a joint estimate of travel time and distance. We compare the performance of ST-NN to that of state-of-the-art travel time estimation methods, and we observe that the proposed approach generalizes better than state-of-the-art methods. We show that ST-NN approach significantly reduces the mean absolute error for both predicted travel time and distance, about 17% for travel time prediction. We also observe that the proposed approach is more robust to outliers present in the dataset by testing the performance of ST-NN on the datasets with and without outliers.
[ 1, 0, 0, 1, 0, 0 ]
Title: Lion and man in non-metric spaces, Abstract: A lion and a man move continuously in a space $X$. The aim of the lion is to capture his prey while the man wants to escape forever. Which of them has a strategy? This question has been studied for different metric domains. In this article we consider the case of general topological spaces.
[ 0, 0, 1, 0, 0, 0 ]
Title: Banach Algebra of Complex Bounded Radon Measures on Homogeneous Space, Abstract: Let $ H $ be a compact subgroup of a locally compact group $G$. In this paper we define a convolution on $ M(G/H) $, the space of all complex bounded Radon measures on the homogeneous space G/H. Then we prove that the measure space $ M(G/H, *) $ is a non-unital Banach algebra that possesses an approximate identity. Finally, it is shown that the Banach algebra $ M(G/H, *) $ is not involutive and also $ L^1(G/H, *) $ is a two-sided ideal of it.
[ 0, 0, 1, 0, 0, 0 ]
Title: A Simple Fusion of Deep and Shallow Learning for Acoustic Scene Classification, Abstract: In the past, Acoustic Scene Classification systems have been based on hand crafting audio features that are input to a classifier. Nowadays, the common trend is to adopt data driven techniques, e.g., deep learning, where audio representations are learned from data. In this paper, we propose a system that consists of a simple fusion of two methods of the aforementioned types: a deep learning approach where log-scaled mel-spectrograms are input to a convolutional neural network, and a feature engineering approach, where a collection of hand-crafted features is input to a gradient boosting machine. We first show that both methods provide complementary information to some extent. Then, we use a simple late fusion strategy to combine both methods. We report classification accuracy of each method individually and the combined system on the TUT Acoustic Scenes 2017 dataset. The proposed fused system outperforms each of the individual methods and attains a classification accuracy of 72.8% on the evaluation set, improving the baseline system by 11.8%.
[ 0, 0, 0, 1, 0, 0 ]
Title: Online Convex Optimization with Unconstrained Domains and Losses, Abstract: We propose an online convex optimization algorithm (RescaledExp) that achieves optimal regret in the unconstrained setting without prior knowledge of any bounds on the loss functions. We prove a lower bound showing an exponential separation between the regret of existing algorithms that require a known bound on the loss functions and any algorithm that does not require such knowledge. RescaledExp matches this lower bound asymptotically in the number of iterations. RescaledExp is naturally hyperparameter-free and we demonstrate empirically that it matches prior optimization algorithms that require hyperparameter optimization.
[ 1, 0, 0, 1, 0, 0 ]
Title: A Variational Feature Encoding Method of 3D Object for Probabilistic Semantic SLAM, Abstract: This paper presents a feature encoding method of complex 3D objects for high-level semantic features. Recent approaches to object recognition methods become important for semantic simultaneous localization and mapping (SLAM). However, there is a lack of consideration of the probabilistic observation model for 3D objects, as the shape of a 3D object basically follows a complex probability distribution. Furthermore, since the mobile robot equipped with a range sensor observes only a single view, much information of the object shape is discarded. These limitations are the major obstacles to semantic SLAM and view-independent loop closure using 3D object shapes as features. In order to enable the numerical analysis for the Bayesian inference, we approximate the true observation model of 3D objects to tractable distributions. Since the observation likelihood can be obtained from the generative model, we formulate the true generative model for 3D object with the Bayesian networks. To capture these complex distributions, we apply a variational auto-encoder. To analyze the approximated distributions and encoded features, we perform classification with maximum likelihood estimation and shape retrieval.
[ 1, 0, 0, 0, 0, 0 ]
Title: Dimensionality reduction methods for molecular simulations, Abstract: Molecular simulations produce very high-dimensional data-sets with millions of data points. As analysis methods are often unable to cope with so many dimensions, it is common to use dimensionality reduction and clustering methods to reach a reduced representation of the data. Yet these methods often fail to capture the most important features necessary for the construction of a Markov model. Here we demonstrate the results of various dimensionality reduction methods on two simulation data-sets, one of protein folding and another of protein-ligand binding. The methods tested include a k-means clustering variant, a non-linear auto encoder, principal component analysis and tICA. The dimension-reduced data is then used to estimate the implied timescales of the slowest process by a Markov state model analysis to assess the quality of the projection. The projected dimensions learned from the data are visualized to demonstrate which conformations the various methods choose to represent the molecular process.
[ 1, 0, 0, 1, 0, 0 ]
Title: Exploration--Exploitation in MDPs with Options, Abstract: While a large body of empirical results show that temporally-extended actions and options may significantly affect the learning performance of an agent, the theoretical understanding of how and when options can be beneficial in online reinforcement learning is relatively limited. In this paper, we derive an upper and lower bound on the regret of a variant of UCRL using options. While we first analyze the algorithm in the general case of semi-Markov decision processes (SMDPs), we show how these results can be translated to the specific case of MDPs with options and we illustrate simple scenarios in which the regret of learning with options can be \textit{provably} much smaller than the regret suffered when learning with primitive actions.
[ 1, 0, 0, 1, 0, 0 ]
Title: Pressure induced spin crossover in disordered α-LiFeO2, Abstract: Structural, magnetic and electrical-transport properties of {\alpha}-LiFeO2, crystallizing in the rock salt structure with random distribution of Li and Fe ions, have been studied by synchrotron X-ray diffraction, 57Fe Mössbauer spectroscopy and electrical resistance measurements at pressures up to 100 GPa using diamond anvil cells. It was found that the crystal structure is stable at least to 82 GPa, though a significant change in compressibility has been observed above 50 GPa. The changes in the structural properties are found to be on a par with a sluggish Fe3+ high- to low-spin (HS-LS) transition (S=5/2 to S=1/2) starting at 50 GPa and not completed even at ~100 GPa. The HS-LS transition is accompanied by an appreciable resistance decrease remaining a semiconductor up to 115 GPa and is not expected to be metallic even at about 200 GPa. The observed feature of the pressure-induced HS-LS transition is not an ordinary behavior of ferric oxides at high pressures. The effect of Fe3+ nearest and next nearest neighbors on the features of the spin crossover is discussed.
[ 0, 1, 0, 0, 0, 0 ]
Title: Central Limit Theorem for empirical transportation cost in general dimension, Abstract: We consider the problem of optimal transportation with quadratic cost between a empirical measure and a general target probability on R d , with d $\ge$ 1. We provide new results on the uniqueness and stability of the associated optimal transportation potentials , namely, the minimizers in the dual formulation of the optimal transportation problem. As a consequence, we show that a CLT holds for the empirical transportation cost under mild moment and smoothness requirements. The limiting distributions are Gaussian and admit a simple description in terms of the optimal transportation potentials.
[ 0, 0, 1, 1, 0, 0 ]
Title: Characterization of temperatures associated to Schrödinger operators with initial data in Morrey spaces, Abstract: Let $\mathcal{L}$ be a Schrödinger operator of the form $\mathcal{L} = -\Delta+V$ acting on $L^2(\mathbb R^n)$ where the nonnegative potential $V$ belongs to the reverse Hölder class $B_q$ for some $q\geq n.$ Let $L^{p,\lambda}(\mathbb{R}^{n})$, $0\le \lambda<n$ denote the Morrey space on $\mathbb{R}^{n}$. In this paper, we will show that a function $f\in L^{2,\lambda}(\mathbb{R}^{n})$ is the trace of the solution of ${\mathbb L}u=u_{t}+{\mathcal{L}}u=0, u(x,0)= f(x),$ where $u$ satisfies a Carleson-type condition \begin{eqnarray*} \sup_{x_B, r_B} r_B^{-\lambda}\int_0^{r_B^2}\int_{B(x_B, r_B)} |\nabla u(x,t)|^2 {dx dt} \leq C <\infty. \end{eqnarray*} Conversely, this Carleson-type condition characterizes all the ${\mathbb L}$-carolic functions whose traces belong to the Morrey space $L^{2,\lambda}(\mathbb{R}^{n})$ for all $0\le \lambda<n$. This result extends the analogous characterization founded by Fabes and Neri for the classical BMO space of John and Nirenberg.
[ 0, 0, 1, 0, 0, 0 ]
Title: Dynamic-sensitive cooperation in the presence of multiple strategy updating rules, Abstract: The importance of microscopic details on cooperation level is an intensively studied aspect of evolutionary game theory. Interestingly, these details become crucial on heterogeneous populations where individuals may possess diverse traits. By introducing a coevolutionary model in which not only strategies but also individual dynamical features may evolve we revealed that the formerly established conclusion is not necessarily true when different updating rules are on stage. In particular, we apply two strategy updating rules, imitation and Death-Birth rule, which allow local selection in a spatial system. Our observation highlights that the microscopic feature of dynamics, like the level of learning activity, could be a fundamental factor even if all players share the same trait uniformly.
[ 0, 0, 0, 0, 1, 0 ]
Title: Inference in Graphical Models via Semidefinite Programming Hierarchies, Abstract: Maximum A posteriori Probability (MAP) inference in graphical models amounts to solving a graph-structured combinatorial optimization problem. Popular inference algorithms such as belief propagation (BP) and generalized belief propagation (GBP) are intimately related to linear programming (LP) relaxation within the Sherali-Adams hierarchy. Despite the popularity of these algorithms, it is well understood that the Sum-of-Squares (SOS) hierarchy based on semidefinite programming (SDP) can provide superior guarantees. Unfortunately, SOS relaxations for a graph with $n$ vertices require solving an SDP with $n^{\Theta(d)}$ variables where $d$ is the degree in the hierarchy. In practice, for $d\ge 4$, this approach does not scale beyond a few tens of variables. In this paper, we propose binary SDP relaxations for MAP inference using the SOS hierarchy with two innovations focused on computational efficiency. Firstly, in analogy to BP and its variants, we only introduce decision variables corresponding to contiguous regions in the graphical model. Secondly, we solve the resulting SDP using a non-convex Burer-Monteiro style method, and develop a sequential rounding procedure. We demonstrate that the resulting algorithm can solve problems with tens of thousands of variables within minutes, and outperforms BP and GBP on practical problems such as image denoising and Ising spin glasses. Finally, for specific graph types, we establish a sufficient condition for the tightness of the proposed partial SOS relaxation.
[ 1, 0, 0, 1, 0, 0 ]
Title: A Hybrid Approach using Ontology Similarity and Fuzzy Logic for Semantic Question Answering, Abstract: One of the challenges in information retrieval is providing accurate answers to a user's question often expressed as uncertainty words. Most answers are based on a Syntactic approach rather than a Semantic analysis of the query. In this paper, our objective is to present a hybrid approach for a Semantic question answering retrieval system using Ontology Similarity and Fuzzy logic. We use a Fuzzy Co-clustering algorithm to retrieve the collection of documents based on Ontology Similarity. The Fuzzy Scale uses Fuzzy type-1 for documents and Fuzzy type-2 for words to prioritize answers. The objective of this work is to provide retrieval system with more accurate answers than non-fuzzy Semantic Ontology approach.
[ 1, 0, 0, 0, 0, 0 ]
Title: Scalable multimodal convolutional networks for brain tumour segmentation, Abstract: Brain tumour segmentation plays a key role in computer-assisted surgery. Deep neural networks have increased the accuracy of automatic segmentation significantly, however these models tend to generalise poorly to different imaging modalities than those for which they have been designed, thereby limiting their applications. For example, a network architecture initially designed for brain parcellation of monomodal T1 MRI can not be easily translated into an efficient tumour segmentation network that jointly utilises T1, T1c, Flair and T2 MRI. To tackle this, we propose a novel scalable multimodal deep learning architecture using new nested structures that explicitly leverage deep features within or across modalities. This aims at making the early layers of the architecture structured and sparse so that the final architecture becomes scalable to the number of modalities. We evaluate the scalable architecture for brain tumour segmentation and give evidence of its regularisation effect compared to the conventional concatenation approach.
[ 1, 0, 0, 0, 0, 0 ]
Title: Fermionic projected entangled-pair states and topological phases, Abstract: We study fermionic matrix product operator algebras and identify the associated algebraic data. Using this algebraic data we construct fermionic tensor network states in two dimensions that have non-trivial symmetry-protected or intrinsic topological order. The tensor network states allow us to relate physical properties of the topological phases to the underlying algebraic data. We illustrate this by calculating defect properties and modular matrices of supercohomology phases. Our formalism also captures Majorana defects as we show explicitly for a class of $\mathbb{Z}_2$ symmetry-protected and intrinsic topological phases. The tensor networks states presented here are well-suited for numerical applications and hence open up new possibilities for studying interacting fermionic topological phases.
[ 0, 1, 0, 0, 0, 0 ]
Title: Exact collisional moments for plasma fluid theories, Abstract: The velocity-space moments of the often troublesome nonlinear Landau collision operator are expressed exactly in terms of multi-index Hermite-polynomial moments of the distribution functions. The collisional moments are shown to be generated by derivatives of two well-known functions, namely the Rosenbluth-MacDonald-Judd-Trubnikov potentials for a Gaussian distribution. The resulting formula has a nonlinear dependency on the relative mean flow of the colliding species normalised to the root-mean-square of the corresponding thermal velocities, and a bilinear dependency on densities and higher-order velocity moments of the distribution functions, with no restriction on temperature, flow or mass ratio of the species. The result can be applied to both the classic transport theory of plasmas, that relies on the Chapman-Enskog method, as well as to deriving collisional fluid equations that follow Grad's moment approach. As an illustrative example, we provide the collisional ten-moment equations with exact conservation laws for momentum- and energy-transfer rate.
[ 0, 1, 1, 0, 0, 0 ]
Title: What is Unique in Individual Gait Patterns? Understanding and Interpreting Deep Learning in Gait Analysis, Abstract: Machine learning (ML) techniques such as (deep) artificial neural networks (DNN) are solving very successfully a plethora of tasks and provide new predictive models for complex physical, chemical, biological and social systems. However, in most cases this comes with the disadvantage of acting as a black box, rarely providing information about what made them arrive at a particular prediction. This black box aspect of ML techniques can be problematic especially in medical diagnoses, so far hampering a clinical acceptance. The present paper studies the uniqueness of individual gait patterns in clinical biomechanics using DNNs. By attributing portions of the model predictions back to the input variables (ground reaction forces and full-body joint angles), the Layer-Wise Relevance Propagation (LRP) technique reliably demonstrates which variables at what time windows of the gait cycle are most relevant for the characterisation of gait patterns from a certain individual. By measuring the timeresolved contribution of each input variable to the prediction of ML techniques such as DNNs, our method describes the first general framework that enables to understand and interpret non-linear ML methods in (biomechanical) gait analysis and thereby supplies a powerful tool for analysis, diagnosis and treatment of human gait.
[ 0, 0, 0, 1, 0, 0 ]
Title: On the ground state of spiking network activity in mammalian cortex, Abstract: Electrophysiological recordings of spiking activity are limited to a small number of neurons. This spatial subsampling has hindered characterizing even most basic properties of collective spiking in cortical networks. In particular, two contradictory hypotheses prevailed for over a decade: the first proposed an asynchronous irregular state, the second a critical state. While distinguishing them is straightforward in models, we show that in experiments classical approaches fail to correctly infer network dynamics because of subsampling. Deploying a novel, subsampling-invariant estimator, we find that in vivo dynamics do not comply with either hypothesis, but instead occupy a narrow "reverberating" state consistently across multiple mammalian species and cortical areas. A generic model tuned to this reverberating state predicts single neuron, pairwise, and population properties. With these predictions we first validate the model and then deduce network properties that are challenging to obtain experimentally, like the network timescale and strength of cortical input.
[ 0, 0, 0, 1, 1, 0 ]
Title: Inferring Information Flow in Spike-train Data Sets using a Trial-Shuffle Method, Abstract: Understanding information processing in the brain requires the ability to determine the functional connectivity between the different regions of the brain. We present a method using transfer entropy to extract this flow of information between brain regions from spike-train data commonly obtained in neurological experiments. Transfer entropy is a statistical measure based in information theory that attempts to quantify the information flow from one process to another, and has been applied to find connectivity in simulated spike-train data. Due to statistical error in the estimator, inferring functional connectivity requires a method for determining significance in the transfer entropy values. We discuss the issues with numerical estimation of transfer entropy and resulting challenges in determining significance before presenting the trial-shuffle method as a viable option. The trial-shuffle method, for spike-train data that is split into multiple trials, determines significant transfer entropy values independently for each individual pair of neurons by comparing to a created baseline distribution using a rigorous statistical test. This is in contrast to either globally comparing all neuron transfer entropy values or comparing pairwise values to a single baseline value. In establishing the viability of this method by comparison to several alternative approaches in the literature, we find evidence that preserving the inter-spike-interval timing is important. We then use the trial-shuffle method to investigate information flow within a model network as we vary model parameters. This includes investigating the global flow of information within a connectivity network divided into two well-connected subnetworks, going beyond local transfer of information between pairs of neurons.
[ 0, 0, 0, 0, 1, 0 ]
Title: Correlated Components Analysis - Extracting Reliable Dimensions in Multivariate Data, Abstract: How does one find dimensions in multivariate data that are reliably expressed across repetitions? For example, in a brain imaging study one may want to identify combinations of neural signals that are reliably expressed across multiple trials or subjects. For a behavioral assessment with multiple ratings, one may want to identify an aggregate score that is reliably reproduced across raters. Correlated Components Analysis (CorrCA) addresses this problem by identifying components that are maximally correlated between repetitions (e.g. trials, subjects, raters). Here we formalize this as the maximization of the ratio of between-repetition to within-repetition covariance. We show that this criterion maximizes repeat-reliability, defined as mean over variance across repeats, and that it leads to CorrCA or to multi-set Canonical Correlation Analysis, depending on the constraints. Surprisingly, we also find that CorrCA is equivalent to Linear Discriminant Analysis for zero-mean signals, which provides an unexpected link between classic concepts of multivariate analysis. We present an exact parametric test of statistical significance based on the F-statistic for normally distributed independent samples, and present and validate shuffle statistics for the case of dependent samples. Regularization and extension to non-linear mappings using kernels are also presented. The algorithms are demonstrated on a series of data analysis applications, and we provide all code and data required to reproduce the results.
[ 0, 0, 0, 1, 0, 0 ]
Title: Defining and estimating stochastic rate change in a dynamic general insurance portfolio, Abstract: Rate change calculations in the literature involve deterministic methods that measure the change in premium for a given policy. The definition of rate change as a statistical parameter is proposed to address the stochastic nature of the premium charged for a policy. It promotes the idea that rate change is a property of an asymptotic population to be estimated, not just a property to measure or monitor in the sample of observed policies that are written. Various models and techniques are given for estimating this stochastic rate change and quantifying the uncertainty in the estimates. The use of matched sampling is emphasized for rate change estimation, as it adjusts for changes in policy characteristics by directly searching for similar policies across policy years. This avoids any of the assumptions and recipes that are required to re-rate policies in years where they were not written, as is common with deterministic methods. Such procedures can be subjective or implausible if the structure of rating algorithms change or there are complex and heterogeneous exposure bases and coverages. The methods discussed are applied to a motor premium database. The application includes the use of a genetic algorithm with parallel computations to automatically optimize the matched sampling.
[ 0, 0, 0, 0, 0, 1 ]
Title: Teacher Improves Learning by Selecting a Training Subset, Abstract: We call a learner super-teachable if a teacher can trim down an iid training set while making the learner learn even better. We provide sharp super-teaching guarantees on two learners: the maximum likelihood estimator for the mean of a Gaussian, and the large margin classifier in 1D. For general learners, we provide a mixed-integer nonlinear programming-based algorithm to find a super teaching set. Empirical experiments show that our algorithm is able to find good super-teaching sets for both regression and classification problems.
[ 0, 0, 0, 1, 0, 0 ]
Title: Photospheric Emission of Gamma-Ray Bursts, Abstract: We review the physics of GRB production by relativistic jets that start highly opaque near the central source and then expand to transparency. We discuss dissipative and radiative processes in the jet and how radiative transfer shapes the observed nonthermal spectrum released at the photosphere. A comparison of recent detailed models with observations gives estimates for important parameters of GRB jets, such as the Lorentz factor and magnetization. We also discuss predictions for GRB polarization and neutrino emission.
[ 0, 1, 0, 0, 0, 0 ]
Title: GPU acceleration and performance of the particle-beam-dynamics code Elegant, Abstract: Elegant is an accelerator physics and particle-beam dynamics code widely used for modeling and design of a variety of high-energy particle accelerators and accelerator-based systems. In this paper we discuss a recently developed version of the code that can take advantage of CUDA-enabled graphics processing units (GPUs) to achieve significantly improved performance for a large class of simulations that are important in practice. The GPU version is largely defined by a framework that simplifies implementations of the fundamental kernel types that are used by Elegant: particle operations, reductions, particle loss, histograms, array convolutions and random number generation. Accelerated performance on the Titan Cray XK-7 supercomputer is approximately 6-10 times better with the GPU than all the CPU cores associated with the same node count. In addition to performance, the maintainability of the GPU-accelerated version of the code was considered a key design objective. Accuracy with respect to the CPU implementation is also a core consideration. Four different methods are used to ensure that the accelerated code faithfully reproduces the CPU results.
[ 0, 1, 0, 0, 0, 0 ]
Title: Short-distance breakdown of the Higgs mechanism and the robustness of the BCS theory for charged superconductors, Abstract: Through the Higgs mechanism, the long-range Coulomb interaction eliminates the low-energy Goldstone phase mode in superconductors and transfers spectral weight all the way up to the plasma frequency. Here we show that the Higgs mechanism breaks down for length scales shorter than the superconducting coherence length while it stays intact, even at high energies, in the long-wavelength limit. This effect is a consequence of the composite nature of the Higgs field of superconductivity and the broken Lorentz invariance in a solid. Most importantly, the breakdown of the Higgs mechanism inside the superconducting coherence volume is crucial to ensure the stability of the BCS mean-field theory in the weak-coupling limit. We also show that changes in the gap equation due to plasmon-induced fluctuations can lead to significant corrections to the mean-field theory and reveal that changes in the density-fluctuation spectrum of a superconductor are not limited to the vicinity of the gap.
[ 0, 1, 0, 0, 0, 0 ]
Title: The Massey's method for sport rating: a network science perspective, Abstract: We revisit the Massey's method for rating and ranking in sports and contextualize it as a general centrality measure in network science.
[ 1, 1, 0, 0, 0, 0 ]
Title: Pharmacokinetics Simulations for Studying Correlates of Prevention Efficacy of Passive HIV-1 Antibody Prophylaxis in the Antibody Mediated Prevention (AMP) Study, Abstract: A key objective in two phase 2b AMP clinical trials of VRC01 is to evaluate whether drug concentration over time, as estimated by non-linear mixed effects pharmacokinetics (PK) models, is associated with HIV infection rate. We conducted a simulation study of marker sampling designs, and evaluated the effect of study adherence and sub-cohort sample size on PK model estimates in multiple-dose studies. With m=120, even under low adherence (about half of study visits missing per participant), reasonably unbiased and consistent estimates of most fixed and random effect terms were obtained. Coarsened marker sampling schedules were also studied.
[ 0, 0, 0, 1, 1, 0 ]
Title: Magellan/M2FS Spectroscopy of Galaxy Clusters: Stellar Population Model and Application to Abell 267, Abstract: We report the results of a pilot program to use the Magellan/M2FS spectrograph to survey the galactic populations and internal kinematics of galaxy clusters. For this initial study, we present spectroscopic measurements for $223$ quiescent galaxies observed along the line of sight to the galaxy cluster Abell 267 ($z\sim0.23$). We develop a Bayesian method for modeling the integrated light from each galaxy as a simple stellar population, with free parameters that specify redshift ($v_\mathrm{los}/c$) and characteristic age, metallicity ($\mathrm{[Fe/H]}$), alpha-abundance ($[\alpha/\mathrm{Fe}]$), and internal velocity dispersion ($\sigma_\mathrm{int}$) for individual galaxies. Parameter estimates derived from our 1.5-hour observation of A267 have median random errors of $\sigma_{v_\mathrm{los}}=20\ \mathrm{km\ s^{-1}}$, $\sigma_{\mathrm{Age}}=1.2\ \mathrm{Gyr}$, $\sigma_{\mathrm{[Fe/H]}}=0.11\ \mathrm{dex}$, $\sigma_{[\alpha/\mathrm{Fe}]}=0.07\ \mathrm{dex}$, and $\sigma_{\sigma_\mathrm{int}}=20\ \mathrm{km\ s^{-1}}$. In a companion paper, we use these results to model the structure and internal kinematics of A267.
[ 0, 1, 0, 0, 0, 0 ]
Title: Parametric Decay Instability and Dissipation of Low-frequency Alfvén Waves in Low-beta Turbulent Plasmas, Abstract: Evolution of the parametric decay instability (PDI) of a circularly polarized Aflvén wave in a turbulent low-beta plasma background is investigated using 3D hybrid simulations. It is shown that the turbulence reduces the growth rate of PDI as compared to the linear theory predictions, but PDI can still exist. Interestingly, the damping rate of ion acoustic mode (as the product of PDI) is also reduced as compared to the linear Vlasov predictions. Nonetheless, significant heating of ions in the direction parallel to the background magnetic field is observed due to resonant Landau damping of the ion acoustic waves. In low-beta turbulent plasmas, PDI can provide an important channel for energy dissipation of low-frequency Alfvén waves at a scale much larger than the ion kinetic scales, different from the traditional turbulence dissipation models.
[ 0, 1, 0, 0, 0, 0 ]
Title: Generalised model-independent characterisation of strong gravitational lenses II: Transformation matrix between multiple images, Abstract: (shortened) We determine the transformation matrix T that maps multiple images with resolved features onto one another and that is based on a Taylor-expanded lensing potential close to a point on the critical curve within our model-independent lens characterisation approach. From T, the same information about the critical curve at fold and cusp points is derived as determined by the quadrupole moment of the individual images as observables. In addition, we read off the relative parities between the images, so that the parity of all images is determined, when one is known. We compare all retrievable ratios of potential derivatives to the actual ones and to those obtained by using the quadrupole moment as observable for two and three image configurations generated by a galaxy-cluster scale singular isothermal ellipse. We conclude that using the quadrupole moments as observables, the properties of the critical curve at the cusp points are retrieved to higher accuracy, at the fold points to lower accuracy, and the ratios of second order potential derivatives to comparable accuracy. We show that the approach using ratios of convergences and reduced shear is equivalent to ours close to the critical curve but yields more accurate results and is more robust because it does not require a special coordinate system like the approach using potential derivatives. T is determined by mapping manually assigned reference points in the images onto each other. If the assignment of reference points is subject to measurement uncertainties under noise, we find that the confidence intervals of the lens parameters can be as large as the values, when the uncertainties are larger than one pixel. Observed multiple images with resolved features are more extended than unresolved ones, so that higher order moments should be taken into account to improve the reconstruction.
[ 0, 1, 0, 0, 0, 0 ]
Title: Properties of interaction networks, structure coefficients, and benefit-to-cost ratios, Abstract: In structured populations the spatial arrangement of cooperators and defectors on the interaction graph together with the structure of the graph itself determines the game dynamics and particularly whether or not fixation of cooperation (or defection) is favored. For a single cooperator (and a single defector) and a network described by a regular graph the question of fixation can be addressed by a single parameter, the structure coefficient. As this quantity is generic for any regular graph, we may call it the generic structure coefficient. For two and more cooperators (or several defectors) fixation properties can also be assigned by structure coefficients. These structure coefficients, however, depend on the arrangement of cooperators and defectors which we may interpret as a configuration of the game. Moreover, the coefficients are specific for a given interaction network modeled as regular graph, which is why we may call them specific structure coefficients. In this paper, we study how specific structure coefficients vary over interaction graphs and link the distributions obtained over different graphs to spectral properties of interaction networks. We also discuss implications for the benefit-to-cost ratios of donation games.
[ 0, 0, 0, 0, 1, 0 ]
Title: Optimized Household Demand Management with Local Solar PV Generation, Abstract: Demand Side Management (DSM) strategies are of-ten associated with the objectives of smoothing the load curve and reducing peak load. Although the future of demand side manage-ment is technically dependent on remote and automatic control of residential loads, the end-users play a significant role by shifting the use of appliances to the off-peak hours when they are exposed to Day-ahead market price. This paper proposes an optimum so-lution to the problem of scheduling of household demand side management in the presence of PV generation under a set of tech-nical constraints such as dynamic electricity pricing and voltage deviation. The proposed solution is implemented based on the Clonal Selection Algorithm (CSA). This solution is evaluated through a set of scenarios and simulation results show that the proposed approach results in the reduction of electricity bills and the import of energy from the grid.
[ 1, 0, 0, 0, 0, 0 ]
Title: Interaction-Based Distributed Learning in Cyber-Physical and Social Networks, Abstract: In this paper we consider a network scenario in which agents can evaluate each other according to a score graph that models some physical or social interaction. The goal is to design a distributed protocol, run by the agents, allowing them to learn their unknown state among a finite set of possible values. We propose a Bayesian framework in which scores and states are associated to probabilistic events with unknown parameters and hyperparameters respectively. We prove that each agent can learn its state by means of a local Bayesian classifier and a (centralized) Maximum-Likelihood (ML) estimator of the parameter-hyperparameter that combines plain ML and Empirical Bayes approaches. By using tools from graphical models, which allow us to gain insight on conditional dependences of scores and states, we provide two relaxed probabilistic models that ultimately lead to ML parameter-hyperparameter estimators amenable to distributed computation. In order to highlight the appropriateness of the proposed relaxations, we demonstrate the distributed estimators on a machine-to-machine testing set-up for anomaly detection and on a social interaction set-up for user profiling.
[ 1, 0, 1, 1, 0, 0 ]
Title: EC3: Combining Clustering and Classification for Ensemble Learning, Abstract: Classification and clustering algorithms have been proved to be successful individually in different contexts. Both of them have their own advantages and limitations. For instance, although classification algorithms are more powerful than clustering methods in predicting class labels of objects, they do not perform well when there is a lack of sufficient manually labeled reliable data. On the other hand, although clustering algorithms do not produce label information for objects, they provide supplementary constraints (e.g., if two objects are clustered together, it is more likely that the same label is assigned to both of them) that one can leverage for label prediction of a set of unknown objects. Therefore, systematic utilization of both these types of algorithms together can lead to better prediction performance. In this paper, We propose a novel algorithm, called EC3 that merges classification and clustering together in order to support both binary and multi-class classification. EC3 is based on a principled combination of multiple classification and multiple clustering methods using an optimization function. We theoretically show the convexity and optimality of the problem and solve it by block coordinate descent method. We additionally propose iEC3, a variant of EC3 that handles imbalanced training data. We perform an extensive experimental analysis by comparing EC3 and iEC3 with 14 baseline methods (7 well-known standalone classifiers, 5 ensemble classifiers, and 2 existing methods that merge classification and clustering) on 13 standard benchmark datasets. We show that our methods outperform other baselines for every single dataset, achieving at most 10% higher AUC. Moreover our methods are faster (1.21 times faster than the best baseline), more resilient to noise and class imbalance than the best baseline method.
[ 1, 0, 0, 1, 0, 0 ]
Title: A review of possible effects of cognitive biases on interpretation of rule-based machine learning models, Abstract: This paper investigates to what extent cognitive biases may affect human understanding of interpretable machine learning models, in particular of rules discovered from data. Twenty cognitive biases are covered, as are possible debiasing techniques that can be adopted by designers of machine learning algorithms and software. Our review transfers results obtained in cognitive psychology to the domain of machine learning, aiming to bridge the current gap between these two areas. It needs to be followed by empirical studies specifically aimed at the machine learning domain.
[ 0, 0, 0, 1, 0, 0 ]
Title: High-Tc superconductivity up to 55 K under high pressure in the heavily electron doped Lix(NH3)yFe2Se2 single crystal, Abstract: We report a high-pressure study on the heavily electron doped Lix(NH3)yFe2Se2 single crystal by using the cubic anvil cell apparatus. The superconducting transition temperature Tc = 44 K at ambient pressure is first suppressed to below 20 K upon increasing pressure to Pc = 2 GPa, above which the pressure dependence of Tc(P) reverses and Tc increases steadily to ca. 55 K at 11 GPa. These results thus evidenced a pressure-induced second high-Tc superconducting (SC-II) phase in Lix(NH3)yFe2Se2 with the highest Tcmax = 55K among the FeSe-based bulk materials. Hall data confirm that in the emergent SC-II phase the dominant electron-type carrier density undergoes a fourfold enhancement and tracks the same trend as Tc(P). Interesting, we find a nearly parallel scaling behavior between Tc and the inverse Hall coefficient for the SC-II phases of both Lix(NH3)yFe2Se2 and (Li,Fe)OHFeSe. The present work demonstrates that high pressure offers a distinctive means to further raising the maximum Tc of heavily electron doped FeSe-based materials by increasing the effective charge carrier concentration via a plausible Fermi surface reconstruction at Pc.
[ 0, 1, 0, 0, 0, 0 ]
Title: Existence theorems for a nonlinear second-order distributional differential equation, Abstract: In this work, we are concerned with existence of solutions for a nonlinear second-order distributional differential equation, which contains measure differential equations and stochastic differential equations as special cases. The proof is based on the Leray--Schauder nonlinear alternative and Kurzweil--Henstock--Stieltjes integrals. Meanwhile, examples are worked out to demonstrate that the main results are sharp.
[ 0, 0, 1, 0, 0, 0 ]
Title: Spatial coherence measurement and partially coherent diffractive imaging, Abstract: The complete characterization of spatial coherence is difficult because the mutual coherence function is a complex-valued function of four independent variables. This difficulty limits the ability of controlling and optimizing spatial coherence in a broad range of key applications. Here we propose a method for measuring the complete mutual coherence function, which does not require any prior knowledge and can be scaled to measure arbitrary coherence properties for any wavelength. Our method can also be used to retrieve objects illuminated by partially coherent beam with unknown coherence properties. This study is particularly useful for coherent diffractive imaging of nanoscale structures in the X-ray or electron regime. Our method is not limited by any assumption about the illumination and hence lays the foundation for a branch of new diffractive imaging algorithms.
[ 0, 1, 0, 0, 0, 0 ]
Title: SPECTRE: Seedless Network Alignment via Spectral Centralities, Abstract: Network alignment consists of finding a correspondence between the nodes of two networks. From aligning proteins in computational biology, to de-anonymization of social networks, to recognition tasks in computer vision, this problem has applications in many diverse areas. The current approaches to network alignment mostly focus on the case where prior information is available, either in the form of a seed set of correctly-matched nodes or attributes on the nodes and/or edges. Moreover, those approaches which assume no such prior information tend to be computationally expensive and do not scale to large-scale networks. However, many real-world networks are very large in size, and prior information can be expensive, if not impossible, to obtain. In this paper we introduce SPECTRE, a scalable, accurate algorithm able to solve the network alignment problem with no prior information. SPECTRE makes use of spectral centrality measures and percolation techniques to robustly align nodes across networks, even if those networks exhibit only moderate correlation. Through extensive numerical experiments, we show that SPECTRE is able to recover high-accuracy alignments on both synthetic and real-world networks, and outperforms other algorithms in the seedless case.
[ 1, 0, 0, 0, 0, 0 ]
Title: An efficient genetic algorithm for large-scale planning of robust industrial wireless networks, Abstract: An industrial indoor environment is harsh for wireless communications compared to an office environment, because the prevalent metal easily causes shadowing effects and affects the availability of an industrial wireless local area network (IWLAN). On the one hand, it is costly, time-consuming, and ineffective to perform trial-and-error manual deployment of wireless nodes. On the other hand, the existing wireless planning tools only focus on office environments such that it is hard to plan IWLANs due to the larger problem size and the deployed IWLANs are vulnerable to prevalent shadowing effects in harsh industrial indoor environments. To fill this gap, this paper proposes an overdimensioning model and a genetic algorithm based over-dimensioning (GAOD) algorithm for deploying large-scale robust IWLANs. As a progress beyond the state-of-the-art wireless planning, two full coverage layers are created. The second coverage layer serves as redundancy in case of shadowing. Meanwhile, the deployment cost is reduced by minimizing the number of access points (APs); the hard constraint of minimal inter-AP spatial paration avoids multiple APs covering the same area to be simultaneously shadowed by the same obstacle. The computation time and occupied memory are dedicatedly considered in the design of GAOD for large-scale optimization. A greedy heuristic based over-dimensioning (GHOD) algorithm and a random OD algorithm are taken as benchmarks. In two vehicle manufacturers with a small and large indoor environment, GAOD outperformed GHOD with up to 20% less APs, while GHOD outputted up to 25% less APs than a random OD algorithm. Furthermore, the effectiveness of this model and GAOD was experimentally validated with a real deployment system.
[ 1, 0, 0, 0, 0, 0 ]
Title: Feature selection algorithm based on Catastrophe model to improve the performance of regression analysis, Abstract: In this paper we introduce a new feature selection algorithm to remove the irrelevant or redundant features in the data sets. In this algorithm the importance of a feature is based on its fitting to the Catastrophe model. Akaike information crite- rion value is used for ranking the features in the data set. The proposed algorithm is compared with well-known RELIEF feature selection algorithm. Breast Cancer, Parkinson Telemonitoring data and Slice locality data sets are used to evaluate the model.
[ 1, 0, 0, 1, 0, 0 ]
Title: Neobility at SemEval-2017 Task 1: An Attention-based Sentence Similarity Model, Abstract: This paper describes a neural-network model which performed competitively (top 6) at the SemEval 2017 cross-lingual Semantic Textual Similarity (STS) task. Our system employs an attention-based recurrent neural network model that optimizes the sentence similarity. In this paper, we describe our participation in the multilingual STS task which measures similarity across English, Spanish, and Arabic.
[ 1, 0, 0, 0, 0, 0 ]
Title: Coherent extension of partial automorphisms, free amalgamation, and automorphism groups, Abstract: We give strengthened versions of the Herwig-Lascar and Hodkinson-Otto extension theorems for partial automorphisms of finite structures. Such strengthenings yield several combinatorial and group-theoretic consequences for homogeneous structures. For instance, we establish a coherent form of the extension property for partial automorphisms for certain Fraisse classes. We deduce from these results that the isometry group of the rational Urysohn space, the automorphism group of the Fraisse limit of any Fraisse class that is the class of all $\mathcal{F}$-free structures (in the Herwig--Lascar sense), and the automorphism group of any free homogeneous structure over a finite relational language, all contain a dense locally finite subgroup. We also show that any free homogeneous structure admits ample generics.
[ 0, 0, 1, 0, 0, 0 ]
Title: Homological indices of collections of 1-forms, Abstract: Homological index of a holomorphic 1-form on a complex analytic variety with an isolated singular point is an analogue of the usual index of a 1-form on a non-singular manifold. One can say that it corresponds to the top Chern number of a manifold. We offer a definition of homological indices for collections of 1-forms on a (purely dimensional) complex analytic variety with an isolated singular point corresponding to other Chern numbers. We also define new invariants of germs of complex analytic varieties with isolated singular points related to "vanishing Chern numbers" at them.
[ 0, 0, 1, 0, 0, 0 ]
Title: An age-structured continuum model for myxobacteria, Abstract: Myxobacteria are social bacteria, that can glide in 2D and form counter-propagating, interacting waves. Here we present a novel age-structured, continuous macroscopic model for the movement of myxobacteria. The derivation is based on microscopic interaction rules that can be formulated as a particle-based model and set within the SOH (Self-Organized Hydrodynamics) framework. The strength of this combined approach is that microscopic knowledge or data can be incorporated easily into the particle model, whilst the continuous model allows for easy numerical analysis of the different effects. However we found that the derived macroscopic model lacks a diffusion term in the density equations, which is necessary to control the number of waves, indicating that a higher order approximation during the derivation is crucial. Upon ad-hoc addition of the diffusion term, we found very good agreement between the age-structured model and the biology. In particular we analyzed the influence of a refractory (insensitivity) period following a reversal of movement. Our analysis reveals that the refractory period is not necessary for wave formation, but essential to wave synchronization, indicating separate molecular mechanisms.
[ 0, 1, 1, 0, 0, 0 ]
Title: Smoothing Properties of Bilinear Operators and Leibniz-Type Rules in Lebesgue and Mixed Lebesgue Spaces, Abstract: We prove that bilinear fractional integral operators and similar multipliers are smoothing in the sense that they improve the regularity of functions. We also treat bilinear singular multiplier operators which preserve regularity and obtain several Leibniz-type rules in the contexts of Lebesgue and mixed Lebesgue spaces.
[ 0, 0, 1, 0, 0, 0 ]
Title: Local equilibrium in the Bak-Sneppen model, Abstract: The Bak Sneppen (BS) model is a very simple model that exhibits all the richness of self-organized criticality theory. At the thermodynamic limit, the BS model converges to a situation where all particles have a fitness that is uniformly distributed between a critical value $p_c$ and 1. The $p_c$ value is unknown, as are the variables that influence and determine this value. Here, we study the Bak Sneppen model in the case in which the lowest fitness particle interacts with an arbitrary even number of $m$ nearest neighbors. We show that $p_{c,m}$ verifies a simple local equilibrium relationship. Based on this relationship, we can determine bounds for $p_{c,m}$.
[ 0, 1, 0, 0, 0, 0 ]
Title: Ribbon structures of the Drinfeld center of a finite tensor category, Abstract: We classify the ribbon structures of the Drinfeld center $\mathcal{Z}(\mathcal{C})$ of a finite tensor category $\mathcal{C}$. Our result generalizes Kauffman and Radford's classification result of the ribbon elements of the Drinfeld double of a finite-dimensional Hopf algebra. As a consequence, we see that $\mathcal{Z}(\mathcal{C})$ is a modular tensor category in the sense of Lyubashenko if $\mathcal{C}$ is a spherical finite tensor category in the sense of Douglas, Schommer-Pries and Snyder.
[ 0, 0, 1, 0, 0, 0 ]
Title: X-ray luminescence computed tomography using a focused X-ray beam, Abstract: Due to the low X-ray photon utilization efficiency and low measurement sensitivity of the electron multiplying charge coupled device (EMCCD) camera setup, the collimator based narrow beam X-ray luminescence computed tomography (XLCT) usually requires a long measurement time. In this paper, we, for the first time, report a focused X-ray beam based XLCT imaging system with measurements by a single optical fiber bundle and a photomultiplier tube (PMT). An X-ray tube with a polycapillary lens was used to generate a focused X-ray beam whose X-ray photon density is 1200 times larger than a collimated X-ray beam. An optical fiber bundle was employed to collect and deliver the emitted photons on the phantom surface to the PMT. The total measurement time was reduced to 12.5 minutes. For numerical simulations of both single and six fiber bundle cases, we were able to reconstruct six targets successfully. For the phantom experiment, two targets with an edge-to-edge distance of 0.4 mm and a center-to-center distance of 0.8 mm were successfully reconstructed by the measurement setup with a single fiber bundle and a PMT.
[ 0, 1, 0, 0, 0, 0 ]
Title: Perturbed Kitaev model: excitation spectrum and long-ranged spin correlations, Abstract: We developed general approach to the calculation of power-law infrared asymptotics of spin-spin correlation functions in the Kitaev honeycomb model with different types of perturbations. We have shown that in order to find these correlation functions, one can perform averaging of some bilinear forms composed out of free Majorana fermions, and we presented the method for explicit calculation of these fermionic densities. We demonstrated how to derive an effective Hamiltonian for the Majorana fermions, including the effects of perturbations. For specific application of the general theory, we have studied the effect of the Dzyaloshinskii-Moriya anisotropic spin-spin interaction; we demonstrated that it leads, already in the second order over its relative magnitude $D/K$, to a power-law spin correlation functions, and calculated dynamical spin structure factor of the system. We have shown that an external magnetic field $h$ in presence of the DM interaction, opens a gap in the excitation spectrum of magnitude $\Delta \propto D h$.
[ 0, 1, 0, 0, 0, 0 ]
Title: FreezeOut: Accelerate Training by Progressively Freezing Layers, Abstract: The early layers of a deep neural net have the fewest parameters, but take up the most computation. In this extended abstract, we propose to only train the hidden layers for a set portion of the training run, freezing them out one-by-one and excluding them from the backward pass. Through experiments on CIFAR, we empirically demonstrate that FreezeOut yields savings of up to 20% wall-clock time during training with 3% loss in accuracy for DenseNets, a 20% speedup without loss of accuracy for ResNets, and no improvement for VGG networks. Our code is publicly available at this https URL
[ 1, 0, 0, 1, 0, 0 ]
Title: Stochastic Subsampling for Factorizing Huge Matrices, Abstract: We present a matrix-factorization algorithm that scales to input matrices with both huge number of rows and columns. Learned factors may be sparse or dense and/or non-negative, which makes our algorithm suitable for dictionary learning, sparse component analysis, and non-negative matrix factorization. Our algorithm streams matrix columns while subsampling them to iteratively learn the matrix factors. At each iteration, the row dimension of a new sample is reduced by subsampling, resulting in lower time complexity compared to a simple streaming algorithm. Our method comes with convergence guarantees to reach a stationary point of the matrix-factorization problem. We demonstrate its efficiency on massive functional Magnetic Resonance Imaging data (2 TB), and on patches extracted from hyperspectral images (103 GB). For both problems, which involve different penalties on rows and columns, we obtain significant speed-ups compared to state-of-the-art algorithms.
[ 1, 0, 1, 1, 0, 0 ]
Title: Just ASK: Building an Architecture for Extensible Self-Service Spoken Language Understanding, Abstract: This paper presents the design of the machine learning architecture that underlies the Alexa Skills Kit (ASK) a large scale Spoken Language Understanding (SLU) Software Development Kit (SDK) that enables developers to extend the capabilities of Amazon's virtual assistant, Alexa. At Amazon, the infrastructure powers over 25,000 skills deployed through the ASK, as well as AWS's Amazon Lex SLU Service. The ASK emphasizes flexibility, predictability and a rapid iteration cycle for third party developers. It imposes inductive biases that allow it to learn robust SLU models from extremely small and sparse datasets and, in doing so, removes significant barriers to entry for software developers and dialogue systems researchers.
[ 1, 0, 0, 0, 0, 0 ]
Title: Scaling, Scattering, and Blackbody Radiation in Classical Physics, Abstract: Here we discuss blackbody radiation within the context of classical theory. We note that nonrelativistic classical mechanics and relativistic classical electrodynamics have contrasting scaling symmetries which influence the scattering of radiation. Also, nonrelativistic mechanical systems can be accurately combined with relativistic electromagnetic radiation only provided the nonrelativistic mechanical systems are the low-velocity limits of fully relativistic systems. Application of the no-interaction theorem for relativistic systems limits the scattering mechanical systems for thermal radiation to relativistic classical electrodynamic systems, which involve the Coulomb potential. Whereas the naive use of nonrelativistic scatterers or nonrelativistic classical statistical mechanics leads to the Rayleigh-Jeans spectrum, the use of fully relativistic scatterers leads to the Planck spectrum for blackbody radiation within classical physics.
[ 0, 1, 0, 0, 0, 0 ]
Title: Critical binomial ideals of Norhtcott type, Abstract: In this paper, we study a family of binomial ideals defining monomial curves in the $n-$dimensional affine space determined by $n$ hypersurfaces of the form $x_i^{c_i} - x_1^{u_{i1}} \cdots x_n^{u_{1n}} \in k[x_1, \ldots, x_n]$ with $u_{ii} = 0$, $i\in \{ 1, \ldots, n\}$. We prove that, the monomial curves in that family are set-theoretic complete intersection. Moreover, if the monomial curve is irreducible, we compute some invariants such as genus, type and Fröbenius number of the corresponding numerical semigroup. We also describe a method to produce set-theoretic complete intersection semigroup ideals of arbitrary large height.
[ 0, 0, 1, 0, 0, 0 ]
Title: Clustering of Magnetic Swimmers in a Poiseuille Flow, Abstract: We investigate the collective behavior of magnetic swimmers, which are suspended in a Poiseuille flow and placed under an external magnetic field, using analytical techniques and Brownian dynamics simulations. We find that the interplay between intrinsic activity, external alignment, and magnetic dipole-dipole interactions leads to longitudinal structure formation. Our work sheds light on a recent experimental observation of a clustering instability in this system.
[ 0, 1, 0, 0, 0, 0 ]
Title: A Time-Spectral Method for Initial-Value Problems Using a Novel Spatial Subdomain Scheme, Abstract: We analyse a new subdomain scheme for a time-spectral method for solving initial boundary value problems. Whilst spectral methods are commonplace for spatially dependent systems, finite difference schemes are typically applied for the temporal domain. The Generalized Weighted Residual Method (GWRM) is a fully spectral method in that it spectrally decomposes all specified domains, including the temporal domain, with multivariate Chebyshev polynomials. The Common Boundary-Condition method (CBC) is a spatial subdomain scheme that solves the physical equations independently from the global connection of subdomains. It is here evaluated against two finite difference methods. For the linearised Burger equation the CBC-GWRM is $\sim30\%$ faster and $\sim50\%$ more memory efficient than the semi implicit Crank-Nicolson method at a maximum error $\sim10^{-5}$. For a forced wave equation the CBC-GWRM manages to average efficiently over the small time-scale in the entire temporal domain. The CBC-GWRM is also applied to the linearised ideal magnetohydrodynamic (MHD) equations for a screw pinch equilibrium. The growth rate of the most unstable mode was efficiently computed with an error $<0.1\%$.
[ 0, 1, 0, 0, 0, 0 ]
Title: Stein Variational Adaptive Importance Sampling, Abstract: We propose a novel adaptive importance sampling algorithm which incorporates Stein variational gradient decent algorithm (SVGD) with importance sampling (IS). Our algorithm leverages the nonparametric transforms in SVGD to iteratively decrease the KL divergence between our importance proposal and the target distribution. The advantages of this algorithm are twofold: first, our algorithm turns SVGD into a standard IS algorithm, allowing us to use standard diagnostic and analytic tools of IS to evaluate and interpret the results; second, we do not restrict the choice of our importance proposal to predefined distribution families like traditional (adaptive) IS methods. Empirical experiments demonstrate that our algorithm performs well on evaluating partition functions of restricted Boltzmann machines and testing likelihood of variational auto-encoders.
[ 0, 0, 0, 1, 0, 0 ]
Title: Closure operators on dcpos, Abstract: We examine collective properties of closure operators on posets that are at least dcpos. The first theorem sets the tone of the paper: it tells how a set of preclosure maps on a dcpo determines the least closure operator above them, and pronounces the related induction principle, and its companion, the obverse induction principle. Using this theorem we prove that the poset of closure operators on a dcpo is a complete lattice, and then provide a constructive proof of the Tarski's theorem for dcpos. We go on to construct the joins in the complete lattice of Scott-continuous closure operators on a dcpo, and to prove that the complete lattice of nuclei on a preframe is a frame, giving some constructions in the special case of the frame of all nuclei on a frame. In the rather drawn-out proof if the Hofmann-Mislove-Johnstone theorem we show off the utility of the obverse induction, applying it in the proof of the crucial lemma. After that we shift a viewpoint and prove some results, analogous to results about dcpos, for posets in which certain special subsets have enough maximal elements; these results actually specialize to dcpos, but at the price of using the axiom of choice. We conclude by pointing out two convex geometries associated with closure operators on a dcpo.
[ 0, 0, 1, 0, 0, 0 ]
Title: Multi-Agent Coverage Control with Energy Depletion and Repletion, Abstract: We develop a hybrid system model to describe the behavior of multiple agents cooperatively solving an optimal coverage problem under energy depletion and repletion constraints. The model captures the controlled switching of agents between coverage (when energy is depleted) and battery charging (when energy is replenished) modes. It guarantees the feasibility of the coverage problem by defining a guard function on each agent's battery level to prevent it from dying on its way to a charging station. The charging station plays the role of a centralized scheduler to solve the contention problem of agents competing for the only charging resource in the mission space. The optimal coverage problem is transformed into a parametric optimization problem to determine an optimal recharging policy. This problem is solved through the use of Infinitesimal Perturbation Analysis (IPA), with simulation results showing that a full recharging policy is optimal.
[ 1, 0, 0, 0, 0, 0 ]
Title: Self-injective commutative rings have no nontrivial rigid ideals, Abstract: We establish a link between trace modules and rigidity in modules over Noetherian rings. Using the theory of trace ideals we make partial progress on a question of Dao, and on the Auslander-Reiten conjecture over Artinian Gorenstein rings.
[ 0, 0, 1, 0, 0, 0 ]
Title: Origin of the Drude peak and of zero sound in probe brane holography, Abstract: At zero temperature, the charge current operator appears to be conserved, within linear response, in certain holographic probe brane models of strange metals. At small but finite temperature, we analytically show that the weak non-conservation of this current leads to both a collective "zero sound" mode and a Drude peak in the electrical conductivity. This simultaneously resolves two outstanding puzzles about probe brane theories. The nonlinear dynamics of the current operator itself appears qualitatively different.
[ 0, 1, 0, 0, 0, 0 ]
Title: Multiscale dynamical network mechanisms underlying aging from birth to death, Abstract: How self-organized networks develop, mature and degenerate is a key question for sociotechnical, cyberphysical and biological systems with potential applications from tackling violent extremism through to neurological diseases. So far, it has proved impossible to measure the continuous-time evolution of any in vivo organism network from birth to death. Here we provide such a study which crosses all organizational and temporal scales, from individual components (10^1) through to the mesoscopic (10^3) and entire system scale (10^6). These continuous-time data reveal a lifespan driven by punctuated, real-time co-evolution of the structural and functional networks. Aging sees these structural and functional networks gradually diverge in terms of their small-worldness and eventually their connectivity. Dying emerges as an extended process associated with the formation of large but disjoint functional sub-networks together with an increasingly detached core. Our mathematical model quantifies the very different impacts that interventions will have on the overall lifetime, period of initial growth, peak of potency, and duration of old age, depending on when and how they are administered. In addition to their direct relevance to online extremism, our findings offer fresh insight into aging in any network system of comparable complexity for which extensive in vivo data is not yet available.
[ 0, 1, 0, 0, 0, 0 ]
Title: Theoretical calculations for precision polarimetry based on Mott scattering, Abstract: Electron polarimeters based on Mott scattering are extensively used in different fields in physics such as atomic, nuclear or particle physics. This is because spin-dependent measurements gives additional information on the physical processes under study. The main quantity that needs to be understood in very much detail, both experimentally and theoretically, is the spin-polarization function, so called analyzing power or Sherman function. A detailed theoretical analysis on all the contributions to the effective interaction potential that are relevant at the typical electron beam energies and angles commonly used in the calibration of the experimental apparatus is presented. The main contribution leading the theoretical error on the Sherman function is found to correspond to radiative corrections that have been qualitatively estimated to be below the 0.5% for the considered kinematical conditions: unpolarized electron beams of few MeV elastically scattered from a gold and silver targets at backward angles.
[ 0, 1, 0, 0, 0, 0 ]
Title: An Online Hierarchical Algorithm for Extreme Clustering, Abstract: Many modern clustering methods scale well to a large number of data items, N, but not to a large number of clusters, K. This paper introduces PERCH, a new non-greedy algorithm for online hierarchical clustering that scales to both massive N and K--a problem setting we term extreme clustering. Our algorithm efficiently routes new data points to the leaves of an incrementally-built tree. Motivated by the desire for both accuracy and speed, our approach performs tree rotations for the sake of enhancing subtree purity and encouraging balancedness. We prove that, under a natural separability assumption, our non-greedy algorithm will produce trees with perfect dendrogram purity regardless of online data arrival order. Our experiments demonstrate that PERCH constructs more accurate trees than other tree-building clustering algorithms and scales well with both N and K, achieving a higher quality clustering than the strongest flat clustering competitor in nearly half the time.
[ 1, 0, 0, 1, 0, 0 ]