text
stringlengths
57
2.88k
labels
sequencelengths
6
6
Title: Statistical Analysis on Bangla Newspaper Data to Extract Trending Topic and Visualize Its Change Over Time, Abstract: Trending topic of newspapers is an indicator to understand the situation of a country and also a way to evaluate the particular newspaper. This paper represents a model describing few techniques to select trending topics from Bangla Newspaper. Topics that are discussed more frequently than other in Bangla newspaper will be marked and how a very famous topic loses its importance with the change of time and another topic takes its place will be demonstrated. Data from two popular Bangla Newspaper with date and time were collected. Statistical analysis was performed after on these data after preprocessing. Popular and most used keywords were extracted from the stream of Bangla keyword with this analysis. This model can also cluster category wise news trend or a list of news trend in daily or weekly basis with enough data. A pattern can be found on their news trend too. Comparison among past news trend of Bangla newspapers will give a visualization of the situation of Bangladesh. This visualization will be helpful to predict future trending topics of Bangla Newspaper.
[ 1, 0, 0, 0, 0, 0 ]
Title: Effects of global gas flows on type I migration, Abstract: Magnetically-driven disk winds would alter the surface density slope of gas in the inner region of a protoplanetary disk $(r \lesssim 1 {\rm au})$. This in turn affects planet formation. Recently, the effect of disk wind torque has been considered with the suggestion that it would carve out the surface density of the disk from inside and would induce global gas flows (wind-driven accretion). We aim to investigate effects of global gas flows on type I migration and also examine planet formation. A simplified approach was taken to address this issue, and N-body simulations with isolation-mass planets were also performed. In previous studies, the effect of gas flow induced by turbulence-driven accretion has been taken into account for its desaturation effect of the corotation torque. If more rapid gas flows (e.g., wind-driven accretion) are considered, the desaturation effect can be modified. In MRI-inactive disks, in which the wind-driven accretion dominates the disk evolution, the gas flow at the midplane plays an important role. If this flow is fast, the corotation torque is efficiently desaturated. Then, the fact that the surface density slope can be positive in the inner region due to the wind torque can generate an outward migration region extended to super-Earth mass planets. In this case, we observe that no planets fall onto the central star in N-body simulations with migration forces imposed to reproduce such migration pattern. We also see that super-Earth mass planets can undergo outward migration. Relatively rapid gas flows affects type I migration and thus the formation of close-in planets.
[ 0, 1, 0, 0, 0, 0 ]
Title: Warp: a method for neural network interpretability applied to gene expression profiles, Abstract: We show a proof of principle for warping, a method to interpret the inner working of neural networks in the context of gene expression analysis. Warping is an efficient way to gain insight to the inner workings of neural nets and make them more interpretable. We demonstrate the ability of warping to recover meaningful information for a given class on a samplespecific individual basis. We found warping works well in both linearly and nonlinearly separable datasets. These encouraging results show that warping has a potential to be the answer to neural networks interpretability in computational biology.
[ 1, 0, 0, 0, 0, 0 ]
Title: Tilings of convex sets by mutually incongruent equilateral triangles contain arbitrarily small tiles, Abstract: We show that every tiling of a convex set in the Euclidean plane $\mathbb{R}^2$ by equilateral triangles of mutually different sizes contains arbitrarily small tiles. The proof is purely elementary up to the discussion of one family of tilings of the full plane $\mathbb{R}^2$, which is based on a surprising connection to a random walk on a directed graph.
[ 0, 0, 1, 0, 0, 0 ]
Title: The decomposition of 0-Hecke modules associated to quasisymmetric Schur functions, Abstract: Recently Tewari and van Willigenburg constructed modules of the 0-Hecke algebra that are mapped to the quasisymmetric Schur functions by the quasisymmetric characteristic and decomposed them into a direct sum of certain submodules. We show that these submodules are indecomposable by determining their endomorphism rings.
[ 0, 0, 1, 0, 0, 0 ]
Title: Bayesian Uncertainty Quantification and Information Fusion in CALPHAD-based Thermodynamic Modeling, Abstract: Calculation of phase diagrams is one of the fundamental tools in alloy design---more specifically under the framework of Integrated Computational Materials Engineering. Uncertainty quantification of phase diagrams is the first step required to provide confidence for decision making in property- or performance-based design. As a manner of illustration, a thorough probabilistic assessment of the CALPHAD model parameters is performed against the available data for a Hf-Si binary case study using a Markov Chain Monte Carlo sampling approach. The plausible optimum values and uncertainties of the parameters are thus obtained, which can be propagated to the resulting phase diagram. Using the parameter values obtained from deterministic optimization in a computational thermodynamic assessment tool (in this case Thermo-Calc) as the prior information for the parameter values and ranges in the sampling process is often necessary to achieve a reasonable cost for uncertainty quantification. This brings up the problem of finding an appropriate CALPHAD model with high-level of confidence which is a very hard and costly task that requires considerable expert skill. A Bayesian hypothesis testing based on Bayes' factors is proposed to fulfill the need of model selection in this case, which is applied to compare four recommended models for the Hf-Si system. However, it is demonstrated that information fusion approaches, i.e., Bayesian model averaging and an error correlation-based model fusion, can be used to combine the useful information existing in all the given models rather than just using the best selected model, which may lack some information about the system being modelled.
[ 0, 0, 0, 1, 0, 0 ]
Title: Modern Data Formats for Big Bioinformatics Data Analytics, Abstract: Next Generation Sequencing (NGS) technology has resulted in massive amounts of proteomics and genomics data. This data is of no use if it is not properly analyzed. ETL (Extraction, Transformation, Loading) is an important step in designing data analytics applications. ETL requires proper understanding of features of data. Data format plays a key role in understanding of data, representation of data, space required to store data, data I/O during processing of data, intermediate results of processing, in-memory analysis of data and overall time required to process data. Different data mining and machine learning algorithms require input data in specific types and formats. This paper explores the data formats used by different tools and algorithms and also presents modern data formats that are used on Big Data Platform. It will help researchers and developers in choosing appropriate data format to be used for a particular tool or algorithm.
[ 1, 0, 0, 0, 0, 0 ]
Title: Nearest-Neighbor Based Non-Parametric Probabilistic Forecasting with Applications in Photovoltaic Systems, Abstract: The present contribution offers a simple methodology for the obtainment of data-driven interval forecasting models by combining pairs of quantile regressions. Those regressions are created without the usage of the non-differentiable pinball-loss function, but through a k-nearest-neighbors based training set transformation and traditional regression approaches. By leaving the underlying training algorithms of the data mining techniques unchanged, the presented approach simplifies the creation of quantile regressions with more complex techniques (e.g. artificial neural networks). The quality of the presented methodology is tested on the usecase of photovoltaic power forecasting, for which quantile regressions using polynomial models as well as artificial neural networks and support vector regressions are created. From the resulting evaluation values it can be concluded that acceptable interval forecasting models are created.
[ 1, 0, 0, 0, 0, 0 ]
Title: A thermodynamic parallel of the Braess road-network paradox, Abstract: We provide here a thermodynamic analog of the Braess road-network paradox with irreversible engines working between reservoirs that are placed at vertices of the network. Paradoxes of different kinds reappear, emphasizing the specialty of the network.
[ 0, 1, 0, 0, 0, 0 ]
Title: Weighted gevrey class regularity of euler equation in the whole space, Abstract: In this paper we study the weighted Gevrey class regularity of Euler equation in the whole space R 3. We first establish the local existence of Euler equation in weighted Sobolev space, then obtain the weighted Gevrey regularity of Euler equation. We will use the weighted Sobolev-Gevrey space method to obtain the results of Gevrey regularity of Euler equation, and the use of the property of singular operator in the estimate of the pressure term is the improvement of our work.
[ 0, 0, 1, 0, 0, 0 ]
Title: Temporal connectivity in finite networks with non-uniform measures, Abstract: Soft Random Geometric Graphs (SRGGs) have been widely applied to various models including those of wireless sensor, communication, social and neural networks. SRGGs are constructed by randomly placing nodes in some space and making pairwise links probabilistically using a connection function that is system specific and usually decays with distance. In this paper we focus on the application of SRGGs to wireless communication networks where information is relayed in a multi hop fashion, although the analysis is more general and can be applied elsewhere by using different distributions of nodes and/or connection functions. We adopt a general non-uniform density which can model the stationary distribution of different mobility models, with the interesting case being when the density goes to zero along the boundaries. The global connectivity properties of these non-uniform networks are likely to be determined by highly isolated nodes, where isolation can be caused by the spatial distribution or the local geometry (boundaries). We extend the analysis to temporal-spatial networks where we fix the underlying non-uniform distribution of points and the dynamics are caused by the temporal variations in the link set, and explore the probability a node near the corner is isolated at time $T$. This work allows for insight into how non-uniformity (caused by mobility) and boundaries impact the connectivity features of temporal-spatial networks. We provide a simple method for approximating these probabilities for a range of different connection functions and verify them against simulations. Boundary nodes are numerically shown to dominate the connectivity properties of these finite networks with non-uniform measure.
[ 1, 0, 0, 0, 0, 0 ]
Title: Matching Media Contents with User Profiles by means of the Dempster-Shafer Theory, Abstract: The media industry is increasingly personalizing the offering of contents in attempt to better target the audience. This requires to analyze the relationships that goes established between users and content they enjoy, looking at one side to the content characteristics and on the other to the user profile, in order to find the best match between the two. In this paper we suggest to build that relationship using the Dempster-Shafer's Theory of Evidence, proposing a reference model and illustrating its properties by means of a toy example. Finally we suggest possible applications of the model for tasks that are common in the modern media industry.
[ 1, 0, 0, 0, 0, 0 ]
Title: Separability by Piecewise Testable Languages is PTime-Complete, Abstract: Piecewise testable languages form the first level of the Straubing-Thérien hierarchy. The membership problem for this level is decidable and testing if the language of a DFA is piecewise testable is NL-complete. The question has not yet been addressed for NFAs. We fill in this gap by showing that it is PSpace-complete. The main result is then the lower-bound complexity of separability of regular languages by piecewise testable languages. Two regular languages are separable by a piecewise testable language if the piecewise testable language includes one of them and is disjoint from the other. For languages represented by NFAs, separability by piecewise testable languages is known to be decidable in PTime. We show that it is PTime-hard and that it remains PTime-hard even for minimal DFAs.
[ 1, 0, 0, 0, 0, 0 ]
Title: Past, Present, Future: A Computational Investigation of the Typology of Tense in 1000 Languages, Abstract: We present SuperPivot, an analysis method for low-resource languages that occur in a superparallel corpus, i.e., in a corpus that contains an order of magnitude more languages than parallel corpora currently in use. We show that SuperPivot performs well for the crosslingual analysis of the linguistic phenomenon of tense. We produce analysis results for more than 1000 languages, conducting - to the best of our knowledge - the largest crosslingual computational study performed to date. We extend existing methodology for leveraging parallel corpora for typological analysis by overcoming a limiting assumption of earlier work: We only require that a linguistic feature is overtly marked in a few of thousands of languages as opposed to requiring that it be marked in all languages under investigation.
[ 1, 0, 0, 0, 0, 0 ]
Title: A Stochastic Control Approach to Managed Futures Portfolios, Abstract: We study a stochastic control approach to managed futures portfolios. Building on the Schwartz 97 stochastic convenience yield model for commodity prices, we formulate a utility maximization problem for dynamically trading a single-maturity futures or multiple futures contracts over a finite horizon. By analyzing the associated Hamilton-Jacobi-Bellman (HJB) equation, we solve the investor's utility maximization problem explicitly and derive the optimal dynamic trading strategies in closed form. We provide numerical examples and illustrate the optimal trading strategies using WTI crude oil futures data.
[ 0, 0, 0, 0, 0, 1 ]
Title: Attention-Set based Metric Learning for Video Face Recognition, Abstract: Face recognition has made great progress with the development of deep learning. However, video face recognition (VFR) is still an ongoing task due to various illumination, low-resolution, pose variations and motion blur. Most existing CNN-based VFR methods only obtain a feature vector from a single image and simply aggregate the features in a video, which less consider the correlations of face images in one video. In this paper, we propose a novel Attention-Set based Metric Learning (ASML) method to measure the statistical characteristics of image sets. It is a promising and generalized extension of Maximum Mean Discrepancy with memory attention weighting. First, we define an effective distance metric on image sets, which explicitly minimizes the intra-set distance and maximizes the inter-set distance simultaneously. Second, inspired by Neural Turing Machine, a Memory Attention Weighting is proposed to adapt set-aware global contents. Then ASML is naturally integrated into CNNs, resulting in an end-to-end learning scheme. Our method achieves state-of-the-art performance for the task of video face recognition on the three widely used benchmarks including YouTubeFace, YouTube Celebrities and Celebrity-1000.
[ 1, 0, 0, 0, 0, 0 ]
Title: Variable Annealing Length and Parallelism in Simulated Annealing, Abstract: In this paper, we propose: (a) a restart schedule for an adaptive simulated annealer, and (b) parallel simulated annealing, with an adaptive and parameter-free annealing schedule. The foundation of our approach is the Modified Lam annealing schedule, which adaptively controls the temperature parameter to track a theoretically ideal rate of acceptance of neighboring states. A sequential implementation of Modified Lam simulated annealing is almost parameter-free. However, it requires prior knowledge of the annealing length. We eliminate this parameter using restarts, with an exponentially increasing schedule of annealing lengths. We then extend this restart schedule to parallel implementation, executing several Modified Lam simulated annealers in parallel, with varying initial annealing lengths, and our proposed parallel annealing length schedule. To validate our approach, we conduct experiments on an NP-Hard scheduling problem with sequence-dependent setup constraints. We compare our approach to fixed length restarts, both sequentially and in parallel. Our results show that our approach can achieve substantial performance gains, throughout the course of the run, demonstrating our approach to be an effective anytime algorithm.
[ 1, 0, 0, 0, 0, 0 ]
Title: Low-Shapiro hydrostatic reconstruction technique for blood flow simulation in large arteries with varying geometrical and mechanical properties, Abstract: The purpose of this work is to construct a simple, efficient and accurate well-balanced numerical scheme for one-dimensional (1D) blood flow in large arteries with varying geometrical and mechanical properties. As the steady states at rest are not relevant for blood flow, we construct two well-balanced hydrostatic reconstruction techniques designed to preserve low-Shapiro number steady states that may occur in large network simulations. The Shapiro number S h = u/c is the equivalent of the Froude number for shallow water equations and the Mach number for compressible Euler equations. The first is the low-Shapiro hydrostatic reconstruction (HR-LS), which is a simple and efficient method, inspired from the hydrostatic reconstruction technique (HR). The second is the subsonic hydrostatic reconstruction (HR-S), adapted here to blood flow and designed to exactly preserve all subcritical steady states. We systematically compare HR, HR-LS and HR-S in a series of single artery and arterial network numerical tests designed to evaluate their well-balanced and wave-capturing properties. The results indicate that HR is not adapted to compute blood flow in large arteries as it is unable to capture wave reflections and transmissions when large variations of the arteries' geometrical and mechanical properties are considered. On the contrary, HR-S is exactly well-balanced and is the most accurate hydrostatic reconstruction technique. However, HR-LS is able to compute low-Shapiro number steady states as well as wave reflections and transmissions with satisfying accuracy and is simpler and computationally less expensive than HR-S. We therefore recommend using HR-LS for 1D blood flow simulations in large arterial network simulations.
[ 0, 1, 1, 0, 0, 0 ]
Title: Anisotropic two-gap superconductivity and the absence of a Pauli paramagnetic limit in single-crystalline LaO$_{0.5}$F$_{0.5}$BiS$_2$, Abstract: Ambient-pressure-grown LaO$_{0.5}$F$_{0.5}$BiS$_2$ with a superconducting transition temperature $T_{c}\sim$3K possesses a highly anisotropic normal state. By a series of electrical resistivity measurements with a magnetic field direction varying between the crystalline $c$-axis and the $ab$-plane, we present the first datasets displaying the temperature dependence of the out-of-plane upper critical field $H_{c2}^{\perp}(T)$, the in-plane upper critical field $H_{c2}^{\parallel}(T)$, as well as the angular dependence of $H_{c2}$ at fixed temperatures for ambient-pressure-grown LaO$_{0.5}$F$_{0.5}$BiS$_2$ single crystals. The anisotropy of the superconductivity, $H_{c2}^{\parallel}/H_{c2}^{\perp}$, reaches $\sim$16 on approaching 0 K, but it decreases significantly near $T_{c}$. A pronounced upward curvature of $H_{c2}^{\parallel}(T)$ is observed near $T_{c}$, which we analyze using a two-gap model. Moreover, $H_{c2}^{\parallel}(0)$ is found to exceed the Pauli paramagnetic limit, which can be understood by considering the strong spin-orbit coupling associated with Bi as well as the breaking of the local inversion symmetry at the electronically active BiS$_2$ bilayers. Hence, LaO$_{0.5}$F$_{0.5}$BiS$_2$ with a centrosymmetric lattice structure is a unique platform to explore the physics associated with local parity violation in the bulk crystal.
[ 0, 1, 0, 0, 0, 0 ]
Title: Discrete Invariants of Generically Inconsistent Systems of Laurent Polynomials, Abstract: Let $ \mathcal{A}_1, \ldots, \mathcal{A}_k $ be finite sets in $ \mathbb{Z}^n $ and let $ Y \subset (\mathbb{C}^*)^n $ be an algebraic variety defined by a system of equations \[ f_1 = \ldots = f_k = 0, \] where $ f_1, \ldots, f_k $ are Laurent polynomials with supports in $\mathcal{A}_1, \ldots, \mathcal{A}_k$. Assuming that $ f_1, \ldots, f_k $ are sufficiently generic, the Newton polyhedron theory computes discrete invariants of $ Y $ in terms of the Newton polyhedra of $ f_1, \ldots, f_k $. It may appear that the generic system with fixed supports $ \mathcal{A}_1, \ldots, \mathcal{A}_k $ is inconsistent. In this paper, we compute discrete invariants of algebraic varieties defined by system of equations which are generic in the set of consistent system with support in $\mathcal{A}_1, \ldots, \mathcal{A}_k$ by reducing the question to the Newton polyhedra theory. Unlike the classical situation, not only the Newton polyhedra of $f_1,\dots,f_k$, but also the supports $\mathcal{A}_1, \ldots, \mathcal{A}_k$ themselves appear in the answers.
[ 0, 0, 1, 0, 0, 0 ]
Title: Enabling near real-time remote search for fast transient events with lossy data compression, Abstract: We present a systematic evaluation of JPEG2000 (ISO/IEC 15444) as a transport data format to enable rapid remote searches for fast transient events as part of the Deeper Wider Faster program (DWF). DWF uses ~20 telescopes from radio to gamma-rays to perform simultaneous and rapid-response follow-up searches for fast transient events on millisecond-to-hours timescales. DWF search demands have a set of constraints that is becoming common amongst large collaborations. Here, we focus on the rapid optical data component of DWF led by the Dark Energy Camera (DECam) at CTIO. Each DECam image has 70 total CCDs saved as a ~1.2 gigabyte FITS file. Near real-time data processing and fast transient candidate identifications -- in minutes for rapid follow-up triggers on other telescopes -- requires computational power exceeding what is currently available on-site at CTIO. In this context, data files need to be transmitted rapidly to a foreign location for supercomputing post-processing, source finding, visualization and analysis. This step in the search process poses a major bottleneck, and reducing the data size helps accommodate faster data transmission. To maximise our gain in transfer time and still achieve our science goals, we opt for lossy data compression -- keeping in mind that raw data is archived and can be evaluated at a later time. We evaluate how lossy JPEG2000 compression affects the process of finding transients, and find only a negligible effect for compression ratios up to ~25:1. We also find a linear relation between compression ratio and the mean estimated data transmission speed-up factor. Adding highly customized compression and decompression steps to the science pipeline considerably reduces the transmission time -- validating its introduction to the DWF science pipeline and enabling science that was otherwise too difficult with current technology.
[ 0, 1, 0, 0, 0, 0 ]
Title: Learning with Training Wheels: Speeding up Training with a Simple Controller for Deep Reinforcement Learning, Abstract: Deep Reinforcement Learning (DRL) has been applied successfully to many robotic applications. However, the large number of trials needed for training is a key issue. Most of existing techniques developed to improve training efficiency (e.g. imitation) target on general tasks rather than being tailored for robot applications, which have their specific context to benefit from. We propose a novel framework, Assisted Reinforcement Learning, where a classical controller (e.g. a PID controller) is used as an alternative, switchable policy to speed up training of DRL for local planning and navigation problems. The core idea is that the simple control law allows the robot to rapidly learn sensible primitives, like driving in a straight line, instead of random exploration. As the actor network becomes more advanced, it can then take over to perform more complex actions, like obstacle avoidance. Eventually, the simple controller can be discarded entirely. We show that not only does this technique train faster, it also is less sensitive to the structure of the DRL network and consistently outperforms a standard Deep Deterministic Policy Gradient network. We demonstrate the results in both simulation and real-world experiments.
[ 1, 0, 0, 0, 0, 0 ]
Title: The Partition Rank of a Tensor and $k$-Right Corners in $\mathbb{F}_{q}^{n}$, Abstract: Following the breakthrough of Croot, Lev, and Pach, Tao introduced a symmetrized version of their argument, which is now known as the slice rank method. In this paper, we introduce a more general version of the slice rank of a tensor, which we call the Partition Rank. This allows us to extend the slice rank method to problems that require the variables to be distinct. Using the partition rank, we generalize a recent result of Ge and Shangguan, and prove that any set $A\subset\mathbb{F}_{q}^{n}$ of size \[|A|>(k+1)\cdot\binom{n+(k-1)q}{(k-1)(q-1)}\] contains a $k$-right-corner, that is distinct vectors $x_{1},\dots,x_{k},x_{k+1}$ where $x_{1}-x_{k+1},\dots,x_{k}-x_{k+1}$ are mutually orthogonal, for $q=p^{r}$, a prime power with $p>k$.
[ 0, 0, 1, 0, 0, 0 ]
Title: Tunable Optoelectronic Properties of Triply-Bonded Carbon Molecules with Linear and Graphyne Substructures, Abstract: In this paper we present a detailed computational study of the electronic structure and optical properties of triply-bonded hydrocarbons with linear, and graphyne substructures, with the aim of identifying their potential in opto-electronic device applications. For the purpose, we employed a correlated electron methodology based upon the Pariser-Parr-Pople model Hamiltonian, coupled with the configuration interaction (CI) approach, and studied structures containing up to 42 carbon atoms. Our calculations, based upon large-scale CI expansions, reveal that the linear structures have intense optical absorption at the HOMO-LUMO gap, while the graphyne ones have those at higher energies. Thus, the opto-electronic properties depend on the topology of the {graphyne substructures, suggesting that they can be tuned by means of structural modifications. Our results are in very good agreement with the available experimental data.
[ 0, 1, 0, 0, 0, 0 ]
Title: Higher zigzag algebras, Abstract: Given any Koszul algebra of finite global dimension one can define a new algebra, which we call a higher zigzag algebra, as a twisted trivial extension of the Koszul dual of our original algebra. If our original algebra is the path algebra of a quiver whose underlying graph is a tree, this construction recovers the zigzag algebras of Huerfano and Khovanov. We study examples of higher zigzag algebras coming from Iyama's iterative construction of type A higher representation finite algebras. We give presentations of these algebras by quivers and relations, and describe relations between spherical twists acting on their derived categories. We then make a connection to the McKay correspondence in higher dimensions: if G is a finite abelian subgroup of the special linear group acting on affine space, then the skew group algebra which controls the category of G-equivariant sheaves is Koszul dual to a higher zigzag algebra. Using this, we show that our relations between spherical twists appear naturally in examples from algebraic geometry.
[ 0, 0, 1, 0, 0, 0 ]
Title: Motivic modular forms from equivariant stable homotopy theory, Abstract: In this paper, we produce a cellular motivic spectrum of motivic modular forms over $\R$ and $\C$, answering positively to a conjecture of Dan Isaksen. This spectrum is constructed to have the appropriate cohomology, as a module over the relevant motivic Steenrod algebra. We first produce a $\G$-equivariant version of this spectrum, and then use a machinery to construct a motivic spectrum from an equivariant one. We believe that this machinery will be of independent interest.
[ 0, 0, 1, 0, 0, 0 ]
Title: Topology and experimental distinguishability, Abstract: In this work we introduce the idea that the primary application of topology in experimental sciences is to keep track of what can be distinguished through experimentation. This link provides understanding and justification as to why topological spaces and continuous functions are pervasive tools in science. We first define an experimental observation as a statement that can be verified using an experimental procedure and show that observations are closed under finite conjunction and countable disjunction. We then consider observations that identify elements within a set and show how they induce a Hausdorff and second-countable topology on that set, thus identifying an open set as one that can be associated with an experimental observation. We then show that experimental relationships are continuous functions, as they must preserve experimental distinguishability, and that they are themselves experimentally distinguishable by defining a Hausdorff and second-countable topology for this collection.
[ 0, 0, 1, 0, 0, 0 ]
Title: Edge fracture in complex fluids, Abstract: We study theoretically the edge fracture instability in sheared complex fluids, by means of linear stability analysis and direct nonlinear simulations. We derive an exact analytical expression for the onset of edge fracture in terms of the shear-rate derivative of the fluid's second normal stress difference, the shear-rate derivative of the shear stress, the jump in shear stress across the interface between the fluid and the outside medium (usually air), the surface tension of that interface, and the rheometer gap size. We provide a full mechanistic understanding of the edge fracture instability, carefully validated against our simulations. These findings, which are robust with respect to choice of rheological constitutive model, also suggest a possible route to mitigating edge fracture, potentially allowing experimentalists to achieve and accurately measure stronger flows than hitherto.
[ 0, 1, 0, 0, 0, 0 ]
Title: Testing statistical Isotropy in Cosmic Microwave Background Polarization maps, Abstract: We apply our symmetry based Power tensor technique to test conformity of PLANCK Polarization maps with statistical isotropy. On a wide range of angular scales (l=40-150), our preliminary analysis detects many statistically anisotropic multipoles in foreground cleaned full sky PLANCK polarization maps viz., COMMANDER and NILC. We also study the effect of residual foregrounds that may still be present in the galactic plane using both common UPB77 polarization mask, as well as the individual component separation method specific polarization masks. However some of the statistically anisotropic modes still persist, albeit significantly in NILC map. We further probed the data for any coherent alignments across multipoles in several bins from the chosen multipole range.
[ 0, 1, 0, 0, 0, 0 ]
Title: Generalised Lyapunov Functions and Functionally Generated Trading Strategies, Abstract: This paper investigates the dependence of functional portfolio generation, introduced by Fernholz (1999), on an extra finite variation process. The framework of Karatzas and Ruf (2017) is used to formulate conditions on trading strategies to be strong arbitrage relative to the market over sufficiently large time horizons. A mollification argument and Komlos theorem yield a general class of potential arbitrage strategies. These theoretical results are complemented by several empirical examples using data from the S&P 500 stocks.
[ 0, 0, 0, 0, 0, 1 ]
Title: Local optima of the Sherrington-Kirkpatrick Hamiltonian, Abstract: We study local optima of the Hamiltonian of the Sherrington-Kirkpatrick model. We compute the exponent of the expected number of local optima and determine the "typical" value of the Hamiltonian.
[ 1, 0, 0, 0, 0, 0 ]
Title: Joining Extractions of Regular Expressions, Abstract: Regular expressions with capture variables, also known as "regex formulas," extract relations of spans (interval positions) from text. These relations can be further manipulated via Relational Algebra as studied in the context of document spanners, Fagin et al.'s formal framework for information extraction. We investigate the complexity of querying text by Conjunctive Queries (CQs) and Unions of CQs (UCQs) on top of regex formulas. We show that the lower bounds (NP-completeness and W[1]-hardness) from the relational world also hold in our setting; in particular, hardness hits already single-character text! Yet, the upper bounds from the relational world do not carry over. Unlike the relational world, acyclic CQs, and even gamma-acyclic CQs, are hard to compute. The source of hardness is that it may be intractable to instantiate the relation defined by a regex formula, simply because it has an exponential number of tuples. Yet, we are able to establish general upper bounds. In particular, UCQs can be evaluated with polynomial delay, provided that every CQ has a bounded number of atoms (while unions and projection can be arbitrary). Furthermore, UCQ evaluation is solvable with FPT (Fixed-Parameter Tractable) delay when the parameter is the size of the UCQ.
[ 1, 0, 0, 0, 0, 0 ]
Title: Learning Generalized Reactive Policies using Deep Neural Networks, Abstract: We present a new approach to learning for planning, where knowledge acquired while solving a given set of planning problems is used to plan faster in related, but new problem instances. We show that a deep neural network can be used to learn and represent a \emph{generalized reactive policy} (GRP) that maps a problem instance and a state to an action, and that the learned GRPs efficiently solve large classes of challenging problem instances. In contrast to prior efforts in this direction, our approach significantly reduces the dependence of learning on handcrafted domain knowledge or feature selection. Instead, the GRP is trained from scratch using a set of successful execution traces. We show that our approach can also be used to automatically learn a heuristic function that can be used in directed search algorithms. We evaluate our approach using an extensive suite of experiments on two challenging planning problem domains and show that our approach facilitates learning complex decision making policies and powerful heuristic functions with minimal human input. Videos of our results are available at goo.gl/Hpy4e3.
[ 1, 0, 0, 0, 0, 0 ]
Title: Symplectic Coarse-Grained Dynamics: Chalkboard Motion in Classical and Quantum Mechanics, Abstract: In the usual approaches to mechanics (classical or quantum) the primary object of interest is the Hamiltonian, from which one tries to deduce the solutions of the equations of motion (Hamilton or Schrödinger). In the present work we reverse this paradigm and view the motions themselves as being the primary objects. This is made possible by studying arbitrary phase space motions, not of points, but of (small) ellipsoids with the requirement that the symplectic capacity of these ellipsoids is preserved. This allows us to guide and control these motions as we like. In the classical case these ellipsoids correspond to a symplectic coarse graining of phase space, and in the quantum case they correspond to the "quantum blobs" we defined in previous work, and which can be viewed as minimum uncertainty phase space cells which are in a one-to-one correspondence with Gaussian pure states.
[ 0, 0, 1, 0, 0, 0 ]
Title: Characteristic functions as bounded multipliers on anisotropic spaces, Abstract: We show that characteristic functions of domains with boundaries transversal to stable cones are bounded multipliers on a recently introduced scale $U^{t,s}_p$ of anisotropic Banach spaces, under the conditions -1+1/p<s<-t<0 and -(r-1)+t<s, with 1<p<infty. (Amended after comments from the referee and M. Jézéquel, January 10, 2018)
[ 0, 1, 0, 0, 0, 0 ]
Title: Pathwise Least Angle Regression and a Significance Test for the Elastic Net, Abstract: Least angle regression (LARS) by Efron et al. (2004) is a novel method for constructing the piece-wise linear path of Lasso solutions. For several years, it remained also as the de facto method for computing the Lasso solution before more sophisticated optimization algorithms preceded it. LARS method has recently again increased its popularity due to its ability to find the values of the penalty parameters, called knots, at which a new parameter enters the active set of non-zero coefficients. Significance test for the Lasso by Lockhart et al. (2014), for example, requires solving the knots via the LARS algorithm. Elastic net (EN), on the other hand, is a highly popular extension of Lasso that uses a linear combination of Lasso and ridge regression penalties. In this paper, we propose a new novel algorithm, called pathwise (PW-)LARS-EN, that is able to compute the EN knots over a grid of EN tuning parameter {\alpha} values. The developed PW-LARS-EN algorithm decreases the EN tuning parameter and exploits the previously found knot values and the original LARS algorithm. A covariance test statistic for the Lasso is then generalized to the EN for testing the significance of the predictors. Our simulation studies validate the fact that the test statistic has an asymptotic Exp(1) distribution.
[ 0, 0, 1, 1, 0, 0 ]
Title: 2MTF VI. Measuring the velocity power spectrum, Abstract: We present measurements of the velocity power spectrum and constraints on the growth rate of structure $f\sigma_{8}$, at redshift zero, using the peculiar motions of 2,062 galaxies in the completed 2MASS Tully-Fisher survey (2MTF). To accomplish this we introduce a model for fitting the velocity power spectrum including the effects of non-linear Redshift Space Distortions (RSD), allowing us to recover unbiased fits down to scales $k=0.2\,h\,{\rm Mpc}^{-1}$ without the need to smooth or grid the data. Our fitting methods are validated using a set of simulated 2MTF surveys. Using these simulations we also identify that the Gaussian distributed estimator for peculiar velocities of \cite{Watkins2015} is suitable for measuring the velocity power spectrum, but sub-optimal for the 2MTF data compared to using magnitude fluctuations $\delta m$, and that, whilst our fits are robust to a change in fiducial cosmology, future peculiar velocity surveys with more constraining power may have to marginalise over this. We obtain \textit{scale-dependent} constraints on the growth rate of structure in two bins, finding $f\sigma_{8} = [0.55^{+0.16}_{-0.13},0.40^{+0.16}_{-0.17}]$ in the ranges $k = [0.007-0.055, 0.55-0.150]\,h\,{\rm Mpc}^{-1}$. We also find consistent results using four bins. Assuming scale-\textit{independence} we find a value $f\sigma_{8} = 0.51^{+0.09}_{-0.08}$, a $\sim16\%$ measurement of the growth rate. Performing a consistency check of General Relativity (GR) and combining our results with CMB data only we find $\gamma = 0.45^{+0.10}_{-0.11}$, a remarkable constraint considering the small number of galaxies. All of our results are completely independent of the effects of galaxy bias, and fully consistent with the predictions of GR (scale-independent $f\sigma_{8}$ and $\gamma\approx0.55$).
[ 0, 1, 0, 0, 0, 0 ]
Title: Gradient weighted norm inequalities for very weak solutions of linear parabolic equations with BMO coefficients, Abstract: In this paper, we prove the Lorentz space $L^{q,p}$-estimates for gradients of very weak solutions to the linear parabolic equations with $\mathbf{A}_q$-weights $$u_t-\operatorname{div}(A(x,t)\nabla u)=\operatorname{div}(F),$$ in a bounded domain $\Omega\times (0,T)\subset\mathbb{R}^{N+1}$, where $A$ has a small mean oscillation, and $\Omega$ is a Lipchistz domain with a small Lipschitz constant.
[ 0, 0, 1, 0, 0, 0 ]
Title: Local incompressibility estimates for the Laughlin phase, Abstract: We prove sharp density upper bounds on optimal length-scales for the ground states of classical 2D Coulomb systems and generalizations thereof. Our method is new, based on an auxiliary Thomas-Fermi-like variational model. Moreover, we deduce density upper bounds for the related low-temperature Gibbs states. Our motivation comes from fractional quantum Hall physics, more precisely, the perturbation of the Laughlin state by external potentials or impurities. These give rise to a class of many-body wave-functions that have the form of a product of the Laughlin state and an analytic function of many variables. This class is related via Laughlin's plasma analogy to Gibbs states of the generalized classical Coulomb systems we consider. Our main result shows that the perturbation of the Laughlin state cannot increase the particle density anywhere, with implications for the response of FQHE systems to external perturbations.
[ 0, 1, 1, 0, 0, 0 ]
Title: Learning Non-local Image Diffusion for Image Denoising, Abstract: Image diffusion plays a fundamental role for the task of image denoising. Recently proposed trainable nonlinear reaction diffusion (TNRD) model defines a simple but very effective framework for image denoising. However, as the TNRD model is a local model, the diffusion behavior of which is purely controlled by information of local patches, it is prone to create artifacts in the homogenous regions and over-smooth highly textured regions, especially in the case of strong noise levels. Meanwhile, it is widely known that the non-local self-similarity (NSS) prior stands as an effective image prior for image denoising, which has been widely exploited in many non-local methods. In this work, we are highly motivated to embed the NSS prior into the TNRD model to tackle its weaknesses. In order to preserve the expected property that end-to-end training is available, we exploit the NSS prior by a set of non-local filters, and derive our proposed trainable non-local reaction diffusion (TNLRD) model for image denoising. Together with the local filters and influence functions, the non-local filters are learned by employing loss-specific training. The experimental results show that the trained TNLRD model produces visually plausible recovered images with more textures and less artifacts, compared to its local versions. Moreover, the trained TNLRD model can achieve strongly competitive performance to recent state-of-the-art image denoising methods in terms of peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM).
[ 1, 0, 0, 0, 0, 0 ]
Title: Deep vs. Diverse Architectures for Classification Problems, Abstract: This study compares various superlearner and deep learning architectures (machine-learning-based and neural-network-based) for classification problems across several simulated and industrial datasets to assess performance and computational efficiency, as both methods have nice theoretical convergence properties. Superlearner formulations outperform other methods at small to moderate sample sizes (500-2500) on nonlinear and mixed linear/nonlinear predictor relationship datasets, while deep neural networks perform well on linear predictor relationship datasets of all sizes. This suggests faster convergence of the superlearner compared to deep neural network architectures on many messy classification problems for real-world data. Superlearners also yield interpretable models, allowing users to examine important signals in the data; in addition, they offer flexible formulation, where users can retain good performance with low-computational-cost base algorithms. K-nearest-neighbor (KNN) regression demonstrates improvements using the superlearner framework, as well; KNN superlearners consistently outperform deep architectures and KNN regression, suggesting that superlearners may be better able to capture local and global geometric features through utilizing a variety of algorithms to probe the data space.
[ 1, 0, 0, 1, 0, 0 ]
Title: Means Moments and Newton's Inequalities, Abstract: It is shown that Newton's inequalities and the related Maclaurin's inequalities provide several refinements of the fundamental Arithmetic mean - Geometric mean - Harmonic mean inequality in terms of the means and variance of positive real numbers. We also obtain some inequalities involving third and fourth central moments of real numbers.
[ 0, 0, 1, 1, 0, 0 ]
Title: Automatic Exploration of Machine Learning Experiments on OpenML, Abstract: Understanding the influence of hyperparameters on the performance of a machine learning algorithm is an important scientific topic in itself and can help to improve automatic hyperparameter tuning procedures. Unfortunately, experimental meta data for this purpose is still rare. This paper presents a large, free and open dataset addressing this problem, containing results on 38 OpenML data sets, six different machine learning algorithms and many different hyperparameter configurations. Results where generated by an automated random sampling strategy, termed the OpenML Random Bot. Each algorithm was cross-validated up to 20.000 times per dataset with different hyperparameters settings, resulting in a meta dataset of around 2.5 million experiments overall.
[ 0, 0, 0, 1, 0, 0 ]
Title: Generalized Expectation Consistent Signal Recovery for Nonlinear Measurements, Abstract: In this paper, we propose a generalized expectation consistent signal recovery algorithm to estimate the signal $\mathbf{x}$ from the nonlinear measurements of a linear transform output $\mathbf{z}=\mathbf{A}\mathbf{x}$. This estimation problem has been encountered in many applications, such as communications with front-end impairments, compressed sensing, and phase retrieval. The proposed algorithm extends the prior art called generalized turbo signal recovery from a partial discrete Fourier transform matrix $\mathbf{A}$ to a class of general matrices. Numerical results show the excellent agreement of the proposed algorithm with the theoretical Bayesian-optimal estimator derived using the replica method.
[ 1, 0, 1, 0, 0, 0 ]
Title: Proactive Eavesdropping in Relaying Systems, Abstract: This paper investigates the performance of a legitimate surveillance system, where a legitimate monitor aims to eavesdrop on a dubious decode-and-forward relaying communication link. In order to maximize the effective eavesdropping rate, two strategies are proposed, where the legitimate monitor adaptively acts as an eavesdropper, a jammer or a helper. In addition, the corresponding optimal jamming beamformer and jamming power are presented. Numerical results demonstrate that the proposed strategies attain better performance compared with intuitive benchmark schemes. Moreover, it is revealed that the position of the legitimate monitor plays an important role on the eavesdropping performance for the two strategies.
[ 1, 0, 0, 0, 0, 0 ]
Title: Efficient Principal Subspace Projection of Streaming Data Through Fast Similarity Matching, Abstract: Big data problems frequently require processing datasets in a streaming fashion, either because all data are available at once but collectively are larger than available memory or because the data intrinsically arrive one data point at a time and must be processed online. Here, we introduce a computationally efficient version of similarity matching, a framework for online dimensionality reduction that incrementally estimates the top K-dimensional principal subspace of streamed data while keeping in memory only the last sample and the current iterate. To assess the performance of our approach, we construct and make public a test suite containing both a synthetic data generator and the infrastructure to test online dimensionality reduction algorithms on real datasets, as well as performant implementations of our algorithm and competing algorithms with similar aims. Among the algorithms considered we find our approach to be competitive, performing among the best on both synthetic and real data.
[ 0, 0, 0, 1, 0, 0 ]
Title: Resolving API Mentions in Informal Documents, Abstract: Developer forums contain opinions and information related to the usage of APIs. API names in forum posts are often not explicitly linked to their official resources. Automatic linking of an API mention to its official resources can be challenging for various reasons, such as, name overloading. We present a technique, ANACE, to automatically resolve API mentions in the textual contents of forum posts. Given a database of APIs, we first detect all words in a forum post that are potential references to an API. We then use a combination of heuristics and machine learning to eliminate false positives and to link true positives to the actual APIs and their resources.
[ 1, 0, 0, 0, 0, 0 ]
Title: Quantifying tidal stream disruption in a simulated Milky Way, Abstract: Simulations of tidal streams show that close encounters with dark matter subhalos induce density gaps and distortions in on-sky path along the streams. Accordingly, observing disrupted streams in the Galactic halo would substantiate the hypothesis that dark matter substructure exists there, while in contrast, observing collimated streams with smoothly varying density profiles would place strong upper limits on the number density and mass spectrum of subhalos. Here, we examine several measures of stream "disruption" and their power to distinguish between halo potentials with and without substructure and with different global shapes. We create and evolve a population of 1280 streams on a range of orbits in the Via Lactea II simulation of a Milky Way-like halo, replete with a full mass range of {\Lambda}CDM subhalos, and compare it to two control stream populations evolved in smooth spherical and smooth triaxial potentials, respectively. We find that the number of gaps observed in a stellar stream is a poor indicator of the halo potential, but that (i) the thinness of the stream on-sky, (ii) the symmetry of the leading and trailing tails, and (iii) the deviation of the tails from a low-order polynomial path on-sky ("path regularity") distinguish between the three potentials more effectively. We find that globular cluster streams on low-eccentricity orbits far from the galactic center (apocentric radius ~ 30-80 kpc) are most powerful in distinguishing between the three potentials. If they exist, such streams will shortly be discoverable and mapped in high dimensions with near-future photometric and spectroscopic surveys.
[ 0, 1, 0, 0, 0, 0 ]
Title: The dehydration of water worlds via atmospheric losses, Abstract: We present a three-species multi-fluid MHD model (H$^+$, H$_2$O$^+$ and e$^-$), endowed with the requisite atmospheric chemistry, that is capable of accurately quantifying the magnitude of water ion losses from exoplanets. We apply this model to a water world with Earth-like parameters orbiting a Sun-like star for three cases: (i) current normal solar wind conditions, (ii) ancient normal solar wind conditions, and (iii) one extreme "Carrington-type" space weather event. We demonstrate that the ion escape rate for (ii), with a value of 6.0$\times$10$^{26}$ s$^{-1}$, is about an order of magnitude higher than the corresponding value of 6.7$\times$10$^{25}$ s$^{-1}$ for (i). Studies of ion losses induced by space weather events, where the ion escape rates can reach $\sim$ 10$^{28}$ s$^{-1}$, are crucial for understanding how an active, early solar-type star (e.g., with frequent coronal mass ejections) could have accelerated the depletion of the exoplanet's atmosphere. We briefly explore the ramifications arising from the loss of water ions, especially for planets orbiting M-dwarfs where such effects are likely to be significant.
[ 0, 1, 0, 0, 0, 0 ]
Title: Replicability Analysis for Natural Language Processing: Testing Significance with Multiple Datasets, Abstract: With the ever-growing amounts of textual data from a large variety of languages, domains, and genres, it has become standard to evaluate NLP algorithms on multiple datasets in order to ensure consistent performance across heterogeneous setups. However, such multiple comparisons pose significant challenges to traditional statistical analysis methods in NLP and can lead to erroneous conclusions. In this paper, we propose a Replicability Analysis framework for a statistically sound analysis of multiple comparisons between algorithms for NLP tasks. We discuss the theoretical advantages of this framework over the current, statistically unjustified, practice in the NLP literature, and demonstrate its empirical value across four applications: multi-domain dependency parsing, multilingual POS tagging, cross-domain sentiment classification and word similarity prediction.
[ 1, 0, 0, 0, 0, 0 ]
Title: Reflections on Cyberethics Education for Millennial Software Engineers, Abstract: Software is a key component of solutions for 21st Century problems. These problems are often "wicked", complex, and unpredictable. To provide the best possible solution, millennial software engineers must be prepared to make ethical decisions, thinking critically, and acting systematically. This reality demands continuous changes in educational systems and curricula delivery, as misjudgment might have serious social impact. This study aims to investigate and reflect on Software Engineering (SE) Programs, proposing a conceptual framework for analyzing cyberethics education and a set of suggestions on how to integrate it into the SE undergraduate curriculum.
[ 1, 0, 0, 0, 0, 0 ]
Title: The effect of an offset polar cap dipolar magnetic field on the modeling of the Vela pulsar's $γ$-ray light curves, Abstract: We performed geometric pulsar light curve modeling using static, retarded vacuum, and offset polar cap (PC) dipole $B$-fields (the latter is characterized by a parameter $\epsilon$), in conjunction with standard two-pole caustic (TPC) and outer gap (OG) emission geometries. The offset-PC dipole $B$-field mimics deviations from the static dipole (which corresponds to $\epsilon=0$). In addition to constant-emissivity geometric models, we also considered a slot gap (SG) $E$-field associated with the offset-PC dipole $B$-field and found that its inclusion leads to qualitatively different light curves. Solving the particle transport equation shows that the particle energy only becomes large enough to yield significant curvature radiation at large altitudes above the stellar surface, given this relatively low $E$-field. Therefore, particles do not always attain the radiation-reaction limit. Our overall optimal light curve fit is for the retarded vacuum dipole field and OG model, at an inclination angle $\alpha=78{_{-1}^{+1}}^{\circ}$ and observer angle $\zeta=69{_{-1}^{+2}}^{\circ}$. For this $B$-field, the TPC model is statistically disfavored compared to the OG model. For the static dipole field, neither model is significantly preferred. We found that smaller values of $\epsilon$ are favored for the offset-PC dipole field when assuming constant emissivity, and larger $\epsilon$ values favored for variable emissivity, but not significantly so. When multiplying the SG $E$-field by a factor of 100, we found improved light curve fits, with $\alpha$ and $\zeta$ being closer to best fits from independent studies, as well as curvature radiation reaction at lower altitudes.
[ 0, 1, 0, 0, 0, 0 ]
Title: Survey on Models and Techniques for Root-Cause Analysis, Abstract: Automation and computer intelligence to support complex human decisions becomes essential to manage large and distributed systems in the Cloud and IoT era. Understanding the root cause of an observed symptom in a complex system has been a major problem for decades. As industry dives into the IoT world and the amount of data generated per year grows at an amazing speed, an important question is how to find appropriate mechanisms to determine root causes that can handle huge amounts of data or may provide valuable feedback in real-time. While many survey papers aim at summarizing the landscape of techniques for modelling system behavior and infering the root cause of a problem based in the resulting models, none of those focuses on analyzing how the different techniques in the literature fit growing requirements in terms of performance and scalability. In this survey, we provide a review of root-cause analysis, focusing on these particular aspects. We also provide guidance to choose the best root-cause analysis strategy depending on the requirements of a particular system and application.
[ 1, 0, 0, 0, 0, 0 ]
Title: Test Prioritization in Continuous Integration Environments, Abstract: Two heuristics namely diversity-based (DBTP) and history-based test prioritization (HBTP) have been separately proposed in the literature. Yet, their combination has not been widely studied in continuous integration (CI) environments. The objective of this study is to catch regression faults earlier, allowing developers to integrate and verify their changes more frequently and continuously. To achieve this, we investigated six open-source projects, each of which included several builds over a large time period. Findings indicate that previous failure knowledge seems to have strong predictive power in CI environments and can be used to effectively prioritize tests. HBTP does not necessarily need to have large data, and its effectiveness improves to a certain degree with larger history interval. DBTP can be used effectively during the early stages, when no historical data is available, and also combined with HBTP to improve its effectiveness. Among the investigated techniques, we found that history-based diversity using NCD Multiset is superior in terms of effectiveness but comes with relatively higher overhead in terms of method execution time. Test prioritization in CI environments can be effectively performed with negligible investment using previous failure knowledge, and its effectiveness can be further improved by considering dissimilarities among the tests.
[ 1, 0, 0, 0, 0, 0 ]
Title: Weighted integral Hankel operators with continuous spectrum, Abstract: Using the Kato-Rosenblum theorem, we describe the absolutely continuous spectrum of a class of weighted integral Hankel operators in $L^2(\mathbb R_+)$. These self-adjoint operators generalise the explicitly diagonalisable operator with the integral kernel $s^\alpha t^\alpha(s+t)^{-1-2\alpha}$, where $\alpha>-1/2$. Our analysis can be considered as an extension of J.Howland's 1992 paper which dealt with the unweighted case, corresponding to $\alpha=0$.
[ 0, 0, 1, 0, 0, 0 ]
Title: Local properties of Riesz minimal energy configurations and equilibrium measures, Abstract: We investigate separation properties of $N$-point configurations that minimize discrete Riesz $s$-energy on a compact set $A\subset \mathbb{R}^p$. When $A$ is a smooth $(p-1)$-dimensional manifold without boundary and $s\in [p-2, p-1)$, we prove that the order of separation (as $N\to \infty$) is the best possible. The same conclusions hold for the points that are a fixed positive distance from the boundary of $A$ whenever $A$ is any $p$-dimensional set. These estimates extend a result of Dahlberg for certain smooth $(p-1)$-dimensional surfaces when $s=p-2$ (the harmonic case). Furthermore, we obtain the same separation results for `greedy' $s$-energy points. We deduce our results from an upper regularity property of the $s$-equilibrium measure (i.e., the measure that solves the continuous minimal Riesz $s$-energy problem), and we show that this property holds under a local smoothness assumption on the set $A$.
[ 0, 0, 1, 0, 0, 0 ]
Title: Cascaded Segmentation-Detection Networks for Word-Level Text Spotting, Abstract: We introduce an algorithm for word-level text spotting that is able to accurately and reliably determine the bounding regions of individual words of text "in the wild". Our system is formed by the cascade of two convolutional neural networks. The first network is fully convolutional and is in charge of detecting areas containing text. This results in a very reliable but possibly inaccurate segmentation of the input image. The second network (inspired by the popular YOLO architecture) analyzes each segment produced in the first stage, and predicts oriented rectangular regions containing individual words. No post-processing (e.g. text line grouping) is necessary. With execution time of 450 ms for a 1000-by-560 image on a Titan X GPU, our system achieves the highest score to date among published algorithms on the ICDAR 2015 Incidental Scene Text dataset benchmark.
[ 1, 0, 0, 0, 0, 0 ]
Title: Hilbert $C^*$-modules over $Σ^*$-algebras II: $Σ^*$-Morita equivalence, Abstract: In previous work, we defined and studied $\Sigma^*$-modules, a class of Hilbert $C^*$-modules over $\Sigma^*$-algebras (the latter are $C^*$-algebras that are sequentially closed in the weak operator topology). The present work continues this study by developing the appropriate $\Sigma^*$-algebraic analogue of the notion of strong Morita equivalence for $C^*$-algebras. We define strong $\Sigma^*$-Morita equivalence, prove a few characterizations, look at the relationship with equivalence of categories of a certain type of Hilbert space representation, study $\Sigma^*$-versions of the interior and exterior tensor products, and prove a $\Sigma^*$-version of the Brown-Green-Rieffel stable isomorphism theorem.
[ 0, 0, 1, 0, 0, 0 ]
Title: Topological degeneracy and pairing in a one-dimensional gas of spinless Fermions, Abstract: We revisit the low energy physics of one dimensional spinless fermion liquids, showing that with sufficiently strong interactions the conventional Luttinger liquid can give way to a strong pairing phase. While the density fluctuations in both phases are described by a gapless Luttinger liquid, single fermion excitations are gapped only in the strong pairing phase. Smooth spatial Interfaces between the two phases lead to topological degeneracies in the ground state and low energy phonon spectrum. Using a concrete microscopic model, with both single particle and pair hopping, we show that the strong pairing state is established through emergence of a new low energy fermionic mode. We characterize the two phases with numerical calculations using the density matrix renormalization group. In particular we find enhancement of the central charge from $c=1$ in the two Luttinger liquid phases to $c=3/2$ at the critical point, which gives direct evidence for an emergent critical Majorana mode. Finally, we confirm the existence of topological degeneracies in the low energy phonon spectrum, associated with spatial interfaces between the two phases.
[ 0, 1, 0, 0, 0, 0 ]
Title: Rejecting inadmissible rules in reduced normal forms in S4, Abstract: Several methods for checking admissibility of rules in the modal logic $S4$ are presented in [1], [15]. These methods determine admissibility of rules in $S4$, but they don't determine or give substitutions rejecting inadmissible rules. In this paper, we investigate some relations between one of the above methods, based on the reduced normal form rules, and sets of substitutions which reject them. We also generalize the method in [1], [15] for one rule to admissibility of a set of rules.
[ 1, 0, 1, 0, 0, 0 ]
Title: An Executable Sequential Specification for Spark Aggregation, Abstract: Spark is a new promising platform for scalable data-parallel computation. It provides several high-level application programming interfaces (APIs) to perform parallel data aggregation. Since execution of parallel aggregation in Spark is inherently non-deterministic, a natural requirement for Spark programs is to give the same result for any execution on the same data set. We present PureSpark, an executable formal Haskell specification for Spark aggregate combinators. Our specification allows us to deduce the precise condition for deterministic outcomes from Spark aggregation. We report case studies analyzing deterministic outcomes and correctness of Spark programs.
[ 1, 0, 0, 0, 0, 0 ]
Title: Obstructions to a small hyperbolicity in Helly graphs, Abstract: It is known that for every graph $G$ there exists the smallest Helly graph $\cal H(G)$ into which $G$ isometrically embeds ($\cal H(G)$ is called the injective hull of $G$) such that the hyperbolicity of $\cal H(G)$ is equal to the hyperbolicity of $G$. Motivated by this, we investigate structural properties of Helly graphs that govern their hyperbolicity and identify three isometric subgraphs of the King-grid as structural obstructions to a small hyperbolicity in Helly graphs.
[ 1, 0, 0, 0, 0, 0 ]
Title: Calculation of time resolution of the J-PET tomograph using the Kernel Density Estimation, Abstract: In this paper we estimate the time resolution of the J-PET scanner built from plastic scintillators. We incorporate the method of signal processing using the Tikhonov regularization framework and the Kernel Density Estimation method. We obtain simple, closed-form analytical formulas for time resolutions. The proposed method is validated using signals registered by means of the single detection unit of the J-PET tomograph built out from 30 cm long plastic scintillator strip. It is shown that the experimental and theoretical results, obtained for the J-PET scanner equipped with vacuum tube photomultipliers, are consistent.
[ 0, 1, 0, 0, 0, 0 ]
Title: Writer Independent Offline Signature Recognition Using Ensemble Learning, Abstract: The area of Handwritten Signature Verification has been broadly researched in the last decades, but remains an open research problem. In offline (static) signature verification, the dynamic information of the signature writing process is lost, and it is difficult to design good feature extractors that can distinguish genuine signatures and skilled forgeries. This verification task is even harder in writer independent scenarios which is undeniably fiscal for realistic cases. In this paper, we have proposed an Ensemble model for offline writer, independent signature verification task with Deep learning. We have used two CNNs for feature extraction, after that RGBT for classification & Stacking to generate final prediction vector. We have done extensive experiments on various datasets from various sources to maintain a variance in the dataset. We have achieved the state of the art performance on various datasets.
[ 1, 0, 0, 1, 0, 0 ]
Title: Automated Directed Fairness Testing, Abstract: Fairness is a critical trait in decision making. As machine-learning models are increasingly being used in sensitive application domains (e.g. education and employment) for decision making, it is crucial that the decisions computed by such models are free of unintended bias. But how can we automatically validate the fairness of arbitrary machine-learning models? For a given machine-learning model and a set of sensitive input parameters, our AEQUITAS approach automatically discovers discriminatory inputs that highlight fairness violation. At the core of AEQUITAS are three novel strategies to employ probabilistic search over the input space with the objective of uncovering fairness violation. Our AEQUITAS approach leverages inherent robustness property in common machine-learning models to design and implement scalable test generation methodologies. An appealing feature of our generated test inputs is that they can be systematically added to the training set of the underlying model and improve its fairness. To this end, we design a fully automated module that guarantees to improve the fairness of the underlying model. We implemented AEQUITAS and we have evaluated it on six state-of-the-art classifiers, including a classifier that was designed with fairness constraints. We show that AEQUITAS effectively generates inputs to uncover fairness violation in all the subject classifiers and systematically improves the fairness of the respective models using the generated test inputs. In our evaluation, AEQUITAS generates up to 70% discriminatory inputs (w.r.t. the total number of inputs generated) and leverages these inputs to improve the fairness up to 94%.
[ 1, 0, 0, 1, 0, 0 ]
Title: Intel MPX Explained: An Empirical Study of Intel MPX and Software-based Bounds Checking Approaches, Abstract: Memory-safety violations are a prevalent cause of both reliability and security vulnerabilities in systems software written in unsafe languages like C/C++. Unfortunately, all the existing software-based solutions to this problem exhibit high performance overheads preventing them from wide adoption in production runs. To address this issue, Intel recently released a new ISA extension - Memory Protection Extensions (Intel MPX), a hardware-assisted full-stack solution to protect against memory safety violations. In this work, we perform an exhaustive study of the Intel MPX architecture to understand its advantages and caveats. We base our study along three dimensions: (a) performance overheads, (b) security guarantees, and (c) usability issues. To put our results in perspective, we compare Intel MPX with three prominent software-based approaches: (1) trip-wire - AddressSanitizer, (2) object-based - SAFECode, and (3) pointer-based - SoftBound. Our main conclusion is that Intel MPX is a promising technique that is not yet practical for widespread adoption. Intel MPX's performance overheads are still high (roughly 50% on average), and the supporting infrastructure has bugs which may cause compilation or runtime errors. Moreover, we showcase the design limitations of Intel MPX: it cannot detect temporal errors, may have false positives and false negatives in multithreaded code, and its restrictions on memory layout require substantial code changes for some programs.
[ 1, 0, 0, 0, 0, 0 ]
Title: Generating Long-term Trajectories Using Deep Hierarchical Networks, Abstract: We study the problem of modeling spatiotemporal trajectories over long time horizons using expert demonstrations. For instance, in sports, agents often choose action sequences with long-term goals in mind, such as achieving a certain strategic position. Conventional policy learning approaches, such as those based on Markov decision processes, generally fail at learning cohesive long-term behavior in such high-dimensional state spaces, and are only effective when myopic modeling lead to the desired behavior. The key difficulty is that conventional approaches are "shallow" models that only learn a single state-action policy. We instead propose a hierarchical policy class that automatically reasons about both long-term and short-term goals, which we instantiate as a hierarchical neural network. We showcase our approach in a case study on learning to imitate demonstrated basketball trajectories, and show that it generates significantly more realistic trajectories compared to non-hierarchical baselines as judged by professional sports analysts.
[ 1, 0, 0, 0, 0, 0 ]
Title: The usefulness of Poynting's theorem in magnetic turbulence, Abstract: We rewrite Poynting's theorem, already used in a previous publication (Treumann & Baumjohann 2017) to derive relations between the turbulent magnetic and electric power spectral densities, to make explicit where the mechanical contributions enter. We then make explicit use of the relativistic transformation of the turbulent electric fluctuations to obtain expressions which depend only on the magnetic and velocity fluctuations. Any electric fluctuations play just an intermediate role. Equations are constructed for the turbulent conductivity spectrum in Alfvénic and non-Alfvénic turbulence in extension of the results in the above citation. An observation-based discussion of their use in application to solar wind turbulence is given. The inertial range solar wind turbulence exhibits signs of chaos and self-organisation.
[ 0, 1, 0, 0, 0, 0 ]
Title: The {\it victory} project v1.0: an efficient parquet equations solver, Abstract: {\it Victory}, i.e. \underline{vi}enna \underline{c}omputational \underline{to}ol deposito\underline{ry}, is a collection of numerical tools for solving the parquet equations for the Hubbard model and similar many body problems. The parquet formalism is a self-consistent theory at both the single- and two-particle levels, and can thus describe individual fermions as well as their collective behavior on equal footing. This is essential for the understanding of various emergent phases and their transitions in many-body systems, in particular for cases in which a single-particle description fails. Our implementation of {\it victory} is in modern Fortran and it fully respects the structure of various vertex functions in both momentum and Matsubara frequency space. We found the latter to be crucial for the convergence of the parquet equations, as well as for the correct determination of various physical observables. In this release, we thoroughly explain the program structure and the controlled approximations to efficiently solve the parquet equations, i.e. the two-level kernel approximation and the high-frequency regulation.
[ 0, 1, 0, 0, 0, 0 ]
Title: An In Vitro Vascularized Tumor Platform for Modeling Breast Tumor Stromal Interactions and Characterizing the Subsequent Response, Abstract: Tumor stromal interactions have been shown to be the driving force behind the poor prognosis associated with aggressive breast tumors. These interactions, specifically between tumor and the surrounding ECM, and tumor and vascular endothelium, promote tumor formation, angiogenesis, and metastasis. In this study, we develop an in vitro vascularized tumor platform that allows for investigation of tumor-stromal interactions in three breast tumor derived cell lines of varying aggressiveness: MDA-IBC3, SUM149, and MDA-MB-231. The platform recreates key features of breast tumors, including increased vascular permeability, vessel sprouting, and ECM remodeling. Morphological and quantitative analysis reveals differential effects from each tumor cell type on endothelial coverage, permeability, expression of VEGF, and collagen remodeling. Triple negative tumors, SUM149 and MDA-MB-321, resulted in a significantly (p<0.05) higher endothelial permeability and decreased endothelial coverage compared to the control TIME only platform. SUM149/TIME platforms were 1.3 fold lower (p<0.05), and MDA-MB-231/TIME platforms were 1.5 fold lower (p<0.01) in endothelial coverage compared to the control TIME only platform. HER2+ MDA-IBC3 tumor cells expressed high levels of VEGF (p<0.01) and induced vessel sprouting. Vessels sprouting was tracked for 3 weeks and with increasing time exhibited formation of multiple vessel sprouts that invaded into the ECM and surrounded clusters of MDA-IBC3 cells. Both IBC cell lines, SUM149 and MDA-IBC3, resulted in a collagen ECM with significantly greater porosity with 1.6 and 1.1 fold higher compared to control, p<0.01. The breast cancer in vitro vascularized platforms introduced in this paper are an adaptable, high throughout tool for unearthing tumor-stromal mechanisms and dynamics behind tumor progression and may prove essential in developing effective targeted therapeutics.
[ 0, 0, 0, 0, 1, 0 ]
Title: Testing for Principal Component Directions under Weak Identifiability, Abstract: We consider the problem of testing, on the basis of a $p$-variate Gaussian random sample, the null hypothesis ${\cal H}_0: {\pmb \theta}_1= {\pmb \theta}_1^0$ against the alternative ${\cal H}_1: {\pmb \theta}_1 \neq {\pmb \theta}_1^0$, where ${\pmb \theta}_1$ is the "first" eigenvector of the underlying covariance matrix and ${\pmb \theta}_1^0$ is a fixed unit $p$-vector. In the classical setup where eigenvalues $\lambda_1>\lambda_2\geq \ldots\geq \lambda_p$ are fixed, the Anderson (1963) likelihood ratio test (LRT) and the Hallin, Paindaveine and Verdebout (2010) Le Cam optimal test for this problem are asymptotically equivalent under the null hypothesis, hence also under sequences of contiguous alternatives. We show that this equivalence does not survive asymptotic scenarios where $\lambda_{n1}/\lambda_{n2}=1+O(r_n)$ with $r_n=O(1/\sqrt{n})$. For such scenarios, the Le Cam optimal test still asymptotically meets the nominal level constraint, whereas the LRT severely overrejects the null hypothesis. Consequently, the former test should be favored over the latter one whenever the two largest sample eigenvalues are close to each other. By relying on the Le Cam's asymptotic theory of statistical experiments, we study the non-null and optimality properties of the Le Cam optimal test in the aforementioned asymptotic scenarios and show that the null robustness of this test is not obtained at the expense of power. Our asymptotic investigation is extensive in the sense that it allows $r_n$ to converge to zero at an arbitrary rate. While we restrict to single-spiked spectra of the form $\lambda_{n1}>\lambda_{n2}=\ldots=\lambda_{np}$ to make our results as striking as possible, we extend our results to the more general elliptical case. Finally, we present an illustrative real data example.
[ 0, 0, 1, 1, 0, 0 ]
Title: Locally-adaptive Bayesian nonparametric inference for phylodynamics, Abstract: Phylodynamics is an area of population genetics that uses genetic sequence data to estimate past population dynamics. Modern state-of-the-art Bayesian nonparametric methods for phylodynamics use either change-point models or Gaussian process priors to recover population size trajectories of unknown form. Change-point models suffer from computational issues when the number of change-points is unknown and needs to be estimated. Gaussian process-based methods lack local adaptivity and cannot accurately recover trajectories that exhibit features such as abrupt changes in trend or varying levels of smoothness. We propose a novel, locally-adaptive approach to Bayesian nonparametric phylodynamic inference that has the flexibility to accommodate a large class of functional behaviors. Local adaptivity results from modeling the log-transformed effective population size a priori as a horseshoe Markov random field, a recently proposed statistical model that blends together the best properties of the change-point and Gaussian process modeling paradigms. We use simulated data to assess model performance, and find that our proposed method results in reduced bias and increased precision when compared to contemporary methods. We also use our models to reconstruct past changes in genetic diversity of human hepatitis C virus in Egypt and to estimate population size changes of ancient and modern steppe bison. These analyses show that our new method captures features of the population size trajectories that were missed by the state-of-the-art phylodynamic methods.
[ 0, 0, 0, 0, 1, 0 ]
Title: Out of sight out of mind: Perceived physical distance between the observer and someone in pain shapes observer's neural empathic reactions, Abstract: Social and affective relations may shape empathy to others' affective states. Previous studies also revealed that people tend to form very different mental representations of stimuli on the basis of their physical distance. In this regard, embodied cognition proposes that different physical distances between individuals activate different interpersonal processing modes, such that close physical distance tends to activate the interpersonal processing mode typical of socially and affectively close relationships. In Experiment 1, two groups of participants were administered a pain decision task involving upright and inverted face stimuli painfully or neutrally stimulated, and we monitored their neural empathic reactions by means of event-related potentials (ERPs) technique. Crucially, participants were presented with face stimuli of one of two possible sizes in order to manipulate retinal size and perceived physical distance, roughly corresponding to the close and far portions of social distance. ERPs modulations compatible with an empathic reaction were observed only for the group exposed to face stimuli appearing to be at a close social distance from the participants. This reaction was absent in the group exposed to smaller stimuli corresponding to face stimuli observed from a far social distance. In Experiment 2, one different group of participants was engaged in a match-to-sample task involving the two-size upright face stimuli of Experiment 1 to test whether the modulation of neural empathic reaction observed in Experiment 1 could be ascribable to differences in the ability to identify faces of the two different sizes. Results suggested that face stimuli of the two sizes could be equally identifiable. In line with the Construal Level and Embodied Simulation theoretical frameworks, we conclude that perceived physical distance may shape empathy as well as social and affective distance.
[ 0, 0, 0, 0, 1, 0 ]
Title: Dynamic Difficulty Adjustment on MOBA Games, Abstract: This paper addresses the dynamic difficulty adjustment on MOBA games as a way to improve the player's entertainment. Although MOBA is currently one of the most played genres around the world, it is known as a game that offer less autonomy, more challenges and consequently more frustration. Due to these characteristics, the use of a mechanism that performs the difficulty balance dynamically seems to be an interesting alternative to minimize and/or avoid that players experience such frustrations. In this sense, this paper presents a dynamic difficulty adjustment mechanism for MOBA games. The main idea is to create a computer controlled opponent that adapts dynamically to the player performance, trying to offer to the player a better game experience. This is done by evaluating the performance of the player using a metric based on some game features and switching the difficulty of the opponent's artificial intelligence behavior accordingly. Quantitative and qualitative experiments were performed and the results showed that the system is capable of adapting dynamically to the opponent's skills. In spite of that, the qualitative experiments with users showed that the player's expertise has a greater influence on the perception of the difficulty level and dynamic adaptation.
[ 1, 0, 0, 0, 0, 0 ]
Title: On iteration of Cox rings, Abstract: We characterize all varieties with a torus action of complexity one that admit iteration of Cox rings.
[ 0, 0, 1, 0, 0, 0 ]
Title: Stability Conditions and Lagrangian Cobordisms, Abstract: In this paper we study the interplay between Lagrangian cobordisms and stability conditions. We show that any stability condition on the derived Fukaya category $D\mathcal{F}uk(M)$ of a symplectic manifold $(M,\omega)$ induces a stability condition on the derived Fukaya category of Lagrangian cobordisms $D\mathcal{F}uk(\mathbb{C} \times M)$. In addition, using stability conditions, we provide general conditions under which the homomorphism $\Theta: \Omega_{Lag}(M)\to K_0(D\mathcal{F}uk(M))$, introduced by Biran and Cornea, is an isomorphism. This yields a better understanding of how stability conditions affect $\Theta$ and it allows us to elucidate Haug's result, that the Lagrangian cobordism group of $T^2$ is isomorphic to $K_0(D\mathcal{F}uk(T^2))$.
[ 0, 0, 1, 0, 0, 0 ]
Title: Sentiment Analysis of Citations Using Word2vec, Abstract: Citation sentiment analysis is an important task in scientific paper analysis. Existing machine learning techniques for citation sentiment analysis are focusing on labor-intensive feature engineering, which requires large annotated corpus. As an automatic feature extraction tool, word2vec has been successfully applied to sentiment analysis of short texts. In this work, I conducted empirical research with the question: how well does word2vec work on the sentiment analysis of citations? The proposed method constructed sentence vectors (sent2vec) by averaging the word embeddings, which were learned from Anthology Collections (ACL-Embeddings). I also investigated polarity-specific word embeddings (PS-Embeddings) for classifying positive and negative citations. The sentence vectors formed a feature space, to which the examined citation sentence was mapped to. Those features were input into classifiers (support vector machines) for supervised classification. Using 10-cross-validation scheme, evaluation was conducted on a set of annotated citations. The results showed that word embeddings are effective on classifying positive and negative citations. However, hand-crafted features performed better for the overall classification.
[ 1, 0, 0, 0, 0, 0 ]
Title: Hybrid Clustering based on Content and Connection Structure using Joint Nonnegative Matrix Factorization, Abstract: We present a hybrid method for latent information discovery on the data sets containing both text content and connection structure based on constrained low rank approximation. The new method jointly optimizes the Nonnegative Matrix Factorization (NMF) objective function for text clustering and the Symmetric NMF (SymNMF) objective function for graph clustering. We propose an effective algorithm for the joint NMF objective function, based on a block coordinate descent (BCD) framework. The proposed hybrid method discovers content associations via latent connections found using SymNMF. The method can also be applied with a natural conversion of the problem when a hypergraph formulation is used or the content is associated with hypergraph edges. Experimental results show that by simultaneously utilizing both content and connection structure, our hybrid method produces higher quality clustering results compared to the other NMF clustering methods that uses content alone (standard NMF) or connection structure alone (SymNMF). We also present some interesting applications to several types of real world data such as citation recommendations of papers. The hybrid method proposed in this paper can also be applied to general data expressed with both feature space vectors and pairwise similarities and can be extended to the case with multiple feature spaces or multiple similarity measures.
[ 1, 0, 0, 1, 0, 0 ]
Title: Embedding dimension and codimension of tensor products of algebras over a field, Abstract: Let k be a field. This paper investigates the embedding dimension and codimension of Noetherian local rings arising as localizations of tensor products of k-algebras. We use results and techniques from prime spectra and dimension theory to establish an analogue of the "special chain theorem" for the embedding dimension of tensor products, with effective consequence on the transfer or defect of regularity as exhibited by the (embedding) codimension.
[ 0, 0, 1, 0, 0, 0 ]
Title: The Cosmic Axion Spin Precession Experiment (CASPEr): a dark-matter search with nuclear magnetic resonance, Abstract: The Cosmic Axion Spin Precession Experiment (CASPEr) is a nuclear magnetic resonance experiment (NMR) seeking to detect axion and axion-like particles which could make up the dark matter present in the universe. We review the predicted couplings of axions and axion-like particles with baryonic matter that enable their detection via NMR. We then describe two measurement schemes being implemented in CASPEr. The first method, presented in the original CASPEr proposal, consists of a resonant search via continuous-wave NMR spectroscopy. This method offers the highest sensitivity for frequencies ranging from a few Hz to hundreds of MHz, corresponding to masses $ m_{\rm a} \sim 10^{-14}$--$10^{-6}$ eV. Sub-Hz frequencies are typically difficult to probe with NMR due to the diminishing sensitivity of magnetometers in this region. To circumvent this limitation, we suggest new detection and data processing modalities. We describe a non-resonant frequency-modulation detection scheme, enabling searches from mHz to Hz frequencies ($m_{\rm a} \sim 10^{-17}$--$10^{-14} $ eV), extending the detection bandwidth by three decades.
[ 0, 1, 0, 0, 0, 0 ]
Title: On eccentricity version of Laplacian energy of a graph, Abstract: The energy of a graph G is equal to the sum of absolute values of the eigenvalues of the adjacency matrix of G, whereas the Laplacian energy of a graph G is equal to the sum of the absolute value of the difference between the eigenvalues of the Laplacian matrix of G and average degree of the vertices of G. Motivated by the work from Sharafdini et al. [R. Sharafdini, H. Panahbar, Vertex weighted Laplacian graph energy and other topological indices. J. Math. Nanosci. 2016, 6, 49-57.], in this paper we investigate the eccentricity version of Laplacian energy of a graph G.
[ 1, 0, 1, 0, 0, 0 ]
Title: LARNN: Linear Attention Recurrent Neural Network, Abstract: The Linear Attention Recurrent Neural Network (LARNN) is a recurrent attention module derived from the Long Short-Term Memory (LSTM) cell and ideas from the consciousness Recurrent Neural Network (RNN). Yes, it LARNNs. The LARNN uses attention on its past cell state values for a limited window size $k$. The formulas are also derived from the Batch Normalized LSTM (BN-LSTM) cell and the Transformer Network for its Multi-Head Attention Mechanism. The Multi-Head Attention Mechanism is used inside the cell such that it can query its own $k$ past values with the attention window. This has the effect of augmenting the rank of the tensor with the attention mechanism, such that the cell can perform complex queries to question its previous inner memories, which should augment the long short-term effect of the memory. With a clever trick, the LARNN cell with attention can be easily used inside a loop on the cell state, just like how any other Recurrent Neural Network (RNN) cell can be looped linearly through time series. This is due to the fact that its state, which is looped upon throughout time steps within time series, stores the inner states in a "first in, first out" queue which contains the $k$ most recent states and on which it is easily possible to add static positional encoding when the queue is represented as a tensor. This neural architecture yields better results than the vanilla LSTM cells. It can obtain results of 91.92% for the test accuracy, compared to the previously attained 91.65% using vanilla LSTM cells. Note that this is not to compare to other research, where up to 93.35% is obtained, but costly using 18 LSTM cells rather than with 2 to 3 cells as analyzed here. Finally, an interesting discovery is made, such that adding activation within the multi-head attention mechanism's linear layers can yield better results in the context researched hereto.
[ 0, 0, 0, 1, 0, 0 ]
Title: Improved nonparametric estimation of the drift in diffusion processes, Abstract: In this paper, we consider the robust adaptive non parametric estimation problem for the drift coefficient in diffusion processes. An adaptive model selection procedure, based on the improved weighted least square estimates, is proposed. Sharp oracle inequalities for the robust risk have been obtained.
[ 0, 0, 1, 1, 0, 0 ]
Title: Energy-efficient Hybrid CMOS-NEMS LIF Neuron Circuit in 28 nm CMOS Process, Abstract: Designing analog sub-threshold neuromorphic circuits in deep sub-micron technologies e.g. 28 nm can be a daunting task due to the problem of excessive leakage current. We propose novel energy-efficient hybrid CMOS-nano electro-mechanical switches (NEMS) Leaky Integrate and Fire (LIF) neuron and synapse circuits and investigate the impact of NEM switches on the leakage power and overall energy consumption. We analyze the performance of biologically-inspired neuron circuit in terms of leakage power consumption and present new energy-efficient neural circuits that operate with biologically plausible firing rates. Our results show the proposed CMOS-NEMS neuron circuit is, on average, 35% more energy-efficient than its CMOS counterpart with same complexity in 28 nm process. Moreover, we discuss how NEM switches can be utilized to further improve the scalability of mixed-signal neuromorphic circuits.
[ 1, 0, 0, 0, 0, 0 ]
Title: Quasar Rain: the Broad Emission Line Region as Condensations in the Warm Accretion Disk Wind, Abstract: The origin of the broad emission line region (BELR) in quasars and active galactic nuclei is still unclear. I propose that condensations form in the warm, radiation pressure driven, accretion disk wind of quasars creating the BEL clouds and uniting them with the other two manifestations of cool, 10,000 K, gas in quasars, the low ionization phase of the warm absorbers (WAs) and the clouds causing X-ray eclipses. The cool clouds will condense quickly (days to years), before the WA outflows reach escape velocity (which takes months to centuries). Cool clouds form in equilibrium with the warm phase of the wind because the rapidly varying X-ray quasar continuum changes the force multiplier, causing pressure waves to move gas into stable locations in pressure-temperature space. The narrow range of 2-phase equilibrium densities may explain the scaling of the BELR size with the square root of luminosity, while the scaling of cloud formation timescales could produce the Baldwin effect. These dense clouds have force multipliers of order unity and so cannot be accelerated to escape velocity. They fall back on a dynamical timescale (months to centuries), producing an inflow that rains down toward the central black hole. As they soon move at Mach ~40 with respect to the WA outflow, these 'raindrops' will be rapidly destroyed within months. This rain of clouds may produce the elliptical BELR orbits implied by velocity resolved reverberation mapping in some objects, and can explain the opening angle and destruction timescale of the narrow 'cometary' tails of the clouds seen in X-ray eclipse observations. Some consequences and challenges of this 'quasar rain' model are presented along with several avenues for theoretical investigation.
[ 0, 1, 0, 0, 0, 0 ]
Title: Consistent polynomial-time unseeded graph matching for Lipschitz graphons, Abstract: We propose a consistent polynomial-time method for the unseeded node matching problem for networks with smooth underlying structures. Despite widely conjectured by the research community that the structured graph matching problem to be significantly easier than its worst case counterpart, well-known to be NP-hard, the statistical version of the problem has stood a challenge that resisted any solution both provable and polynomial-time. The closest existing work requires quasi-polynomial time. Our method is based on the latest advances in graphon estimation techniques and analysis on the concentration of empirical Wasserstein distances. Its core is a simple yet unconventional sampling-and-matching scheme that reduces the problem from unseeded to seeded. Our method allows flexible efficiencies, is convenient to analyze and potentially can be extended to more general settings. Our work enables a rich variety of subsequent estimations and inferences.
[ 0, 0, 0, 1, 0, 0 ]
Title: Photoinduced filling of near nodal gap in Bi$_2$Sr$_2$CaCu$_2$O$_{8+δ}$, Abstract: We report time and angle resolved spectroscopic measurements in optimally doped Bi$_2$Sr$_2$CaCu$_2$O$_{8+\delta}$. The spectral function is monitored as a function of temperature, photoexcitation density and delay time from the pump pulse. According to our data, the superconducting gap becomes slightly stiffer when moving off the nodal direction. The nodal quasiparticles develop a faster dynamics when pumping the superconductor with a fluence that is large enough to induce the total collapse of the gap. We discuss the observed relaxation in terms of a dynamical reformation of Cooper pairs.
[ 0, 1, 0, 0, 0, 0 ]
Title: Globally convergent Jacobi-type algorithms for simultaneous orthogonal symmetric tensor diagonalization, Abstract: In this paper, we consider a family of Jacobi-type algorithms for simultaneous orthogonal diagonalization problem of symmetric tensors. For the Jacobi-based algorithm of [SIAM J. Matrix Anal. Appl., 2(34):651--672, 2013], we prove its global convergence for simultaneous orthogonal diagonalization of symmetric matrices and 3rd-order tensors. We also propose a new Jacobi-based algorithm in the general setting and prove its global convergence for sufficiently smooth functions.
[ 1, 0, 1, 0, 0, 0 ]
Title: Representing the Deligne-Hinich-Getzler $\infty$-groupoid, Abstract: The goal of the present paper is to introduce a smaller, but equivalent version of the Deligne-Hinich-Getzler $\infty$-groupoid associated to a homotopy Lie algebra. In the case of differential graded Lie algebras, we represent it by a universal cosimplicial object.
[ 0, 0, 1, 0, 0, 0 ]
Title: Sentence-level quality estimation by predicting HTER as a multi-component metric, Abstract: This submission investigates alternative machine learning models for predicting the HTER score on the sentence level. Instead of directly predicting the HTER score, we suggest a model that jointly predicts the amount of the 4 distinct post-editing operations, which are then used to calculate the HTER score. This also gives the possibility to correct invalid (e.g. negative) predicted values prior to the calculation of the HTER score. Without any feature exploration, a multi-layer perceptron with 4 outputs yields small but significant improvements over the baseline.
[ 1, 0, 0, 0, 0, 0 ]
Title: Machine Assisted Analysis of Vowel Length Contrasts in Wolof, Abstract: Growing digital archives and improving algorithms for automatic analysis of text and speech create new research opportunities for fundamental research in phonetics. Such empirical approaches allow statistical evaluation of a much larger set of hypothesis about phonetic variation and its conditioning factors (among them geographical / dialectal variants). This paper illustrates this vision and proposes to challenge automatic methods for the analysis of a not easily observable phenomenon: vowel length contrast. We focus on Wolof, an under-resourced language from Sub-Saharan Africa. In particular, we propose multiple features to make a fine evaluation of the degree of length contrast under different factors such as: read vs semi spontaneous speech ; standard vs dialectal Wolof. Our measures made fully automatically on more than 20k vowel tokens show that our proposed features can highlight different degrees of contrast for each vowel considered. We notably show that contrast is weaker in semi-spontaneous speech and in a non standard semi-spontaneous dialect.
[ 1, 0, 0, 0, 0, 0 ]
Title: Precision of Evaluation Methods in White Light Interferometry: Correlogram Correlation Method, Abstract: In this paper we promote a method for the evaluation of a surface topography which we call the correlogram correlation method. Employing a theoretical analysis as well as numerical simulations the method is proven to be the most accurate among available evaluation algorithms in the common case of uncorrelated noise. Examples illustrate the superiority of the correlogram correlation method over the common envelope and phase methods.
[ 0, 1, 0, 0, 0, 0 ]
Title: Non-wetting drops at liquid interfaces: From liquid marbles to Leidenfrost drops, Abstract: We consider the flotation of deformable, non-wetting drops on a liquid interface. We consider the deflection of both the liquid interface and the droplet itself in response to the buoyancy forces, density difference and the various surface tensions within the system. Our results suggest new insight into a range of phenomena in which such drops occur, including Leidenfrost droplets and floating liquid marbles. In particular, we show that the floating state of liquid marbles is very sensitive to the tension of the particle-covered interface and suggest that this sensitivity may make such experiments a useful assay of the properties of these complex interfaces.
[ 0, 1, 0, 0, 0, 0 ]
Title: Kinetics of the Crystalline Nuclei Growth in Glassy Systems, Abstract: In this work, we study the crystalline nuclei growth in glassy systems focusing primarily on the early stages of the process, at which the size of a growing nucleus is still comparable with the critical size. On the basis of molecular dynamics simulation results for two crystallizing glassy systems, we evaluate the growth laws of the crystalline nuclei and the parameters of the growth kinetics at the temperatures corresponding to deep supercoolings; herein, the statistical treatment of the simulation results is done within the mean-first-passage-time method. It is found for the considered systems at different temperatures that the crystal growth laws rescaled onto the waiting times of the critically-sized nucleus follow the unified dependence, that can simplify significantly theoretical description of the post-nucleation growth of crystalline nuclei. The evaluated size-dependent growth rates are characterized by transition to the steady-state growth regime, which depends on the temperature and occurs in the glassy systems when the size of a growing nucleus becomes two-three times larger than a critical size. It is suggested to consider the temperature dependencies of the crystal growth rate characteristics by using the reduced temperature scale $\widetilde{T}$. Thus, it is revealed that the scaled values of the crystal growth rate characteristics (namely, the steady-state growth rate and the attachment rate for the critically-sized nucleus) as functions of the reduced temperature $\widetilde{T}$ for glassy systems follow the unified power-law dependencies. This finding is supported by available simulation results; the correspondence with the experimental data for the crystal growth rate in glassy systems at the temperatures near the glass transition is also discussed.
[ 0, 1, 0, 0, 0, 0 ]
Title: Electron Cloud Trapping In Recycler Combined Function Dipole Magnets, Abstract: Electron cloud can lead to a fast instability in intense proton and positron beams in circular accelerators. In the Fermilab Recycler the electron cloud is confined within its combined function magnets. We show that the field of combined function magnets traps the electron cloud, present the results of analytical estimates of trapping, and compare them to numerical simulations of electron cloud formation. The electron cloud is located at the beam center and up to 1% of the particles can be trapped by the magnetic field. Since the process of electron cloud build-up is exponential, once trapped this amount of electrons significantly increases the density of the cloud on the next revolution. In a Recycler combined function dipole this multi-turn accumulation allows the electron cloud reaching final intensities orders of magnitude greater than in a pure dipole. The multi-turn build-up can be stopped by injection of a clearing bunch of $10^{10}$ p at any position in the ring.
[ 0, 1, 0, 0, 0, 0 ]
Title: Causal Inference Under Network Interference: A Framework for Experiments on Social Networks, Abstract: No man is an island, as individuals interact and influence one another daily in our society. When social influence takes place in experiments on a population of interconnected individuals, the treatment on a unit may affect the outcomes of other units, a phenomenon known as interference. This thesis develops a causal framework and inference methodology for experiments where interference takes place on a network of influence (i.e. network interference). In this framework, the network potential outcomes serve as the key quantity and flexible building blocks for causal estimands that represent a variety of primary, peer, and total treatment effects. These causal estimands are estimated via principled Bayesian imputation of missing outcomes. The theory on the unconfoundedness assumptions leading to simplified imputation highlights the importance of including relevant network covariates in the potential outcome model. Additionally, experimental designs that result in balanced covariates and sizes across treatment exposure groups further improve the causal estimate, especially by mitigating potential outcome model mis-specification. The true potential outcome model is not typically known in real-world experiments, so the best practice is to account for interference and confounding network covariates through both balanced designs and model-based imputation. A full factorial simulated experiment is formulated to demonstrate this principle by comparing performance across different randomization schemes during the design phase and estimators during the analysis phase, under varying network topology and true potential outcome models. Overall, this thesis asserts that interference is not just a nuisance for analysis but rather an opportunity for quantifying and leveraging peer effects in real-world experiments.
[ 0, 0, 1, 1, 0, 0 ]
Title: Polynomial configurations in sets of positive upper density over local fields, Abstract: Let $F(x)=(f_1(x), \dots, f_m(x))$ be such that $1, f_1, \dots, f_m$ are linearly independent polynomials with real coefficients. Based on ideas of Bachoc, DeCorte, Oliveira and Vallentin in combination with estimating certain oscillatory integrals with polynomial phase we will show that the independence ratio of the Cayley graph of $\mathbb{R}^m$ with respect to the portion of the graph of $F$ defined by $a\leq \log |s| \leq T$ is at most $O(1/(T-a))$. We conclude that if $I \subseteq \mathbb{R}^m$ has positive upper density, then the difference set $I-I$ contains vectors of the form $F(s)$ for an unbounded set of values $s \in \mathbb{R}$. It follows that the Borel chromatic number of the Cayley graph of $\mathbb{R}^m$ with respect to the set $\{ \pm F(s): s \in \mathbb{R} \}$ is infinite. Analogous results are also proven when $\mathbb{R}$ is replaced by the field of $p$-adic numbers $\mathbb{Q}_p$. At the end, we will also the existence of real analytic functions $f_1, \dots, f_m$, for which the analogous statements no longer hold.
[ 0, 0, 1, 0, 0, 0 ]
Title: Deterministic Genericity for Polynomial Ideals, Abstract: We consider several notions of genericity appearing in algebraic geometry and commutative algebra. Special emphasis is put on various stability notions which are defined in a combinatorial manner and for which a number of equivalent algebraic characterisations are provided. It is shown that in characteristic zero the corresponding generic positions can be obtained with a simple deterministic algorithm. In positive characteristic, only adapted stable positions are reachable except for quasi-stability which is obtainable in any characteristic.
[ 1, 0, 1, 0, 0, 0 ]
Title: Precise Pointing of Cubesat Telescopes: Comparison Between Heat and Light Induced Attitude Control Methods, Abstract: CubeSats are emerging as low-cost tools to perform astronomy, exoplanet searches and earth observation. These satellites can target an object for science observation for weeks on end. This is typically not possible on larger missions where usage time is shared. The problem of designing an attitude control system for CubeSat telescopes is very challenging because current choice of actuators such as reaction-wheels and magnetorquers can induce jitter on the spacecraft due to moving mechanical parts and due to external disturbances. These telescopes may contain cryo-pumps and servos that introduce additional vibrations. A better solution is required. In our paper, we analyze the feasibility of utilizing solar radiation pressure (SRP) and radiometric force to achieve precise attitude control. Our studies show radiometric actuators to be a viable method to achieve precise pointing. The device uses 8 thin vanes of different temperatures placed in a near-vacuum chamber. These chambers contain trace quantities of lightweight, inert gasses like argon. The temperature gradient across the vanes causes the gas molecules to strike the vanes differently and thus inducing a force. By controlling these forces, it's possible to produce a torque to precisely point or spin a spacecraft. We present a conceptual design of a CubeSat that is equipped with these actuators. We then analyze the potential slew maneuver and slew rates possible with these actuators by simulating their performance. Our analytical and simulation results point towards a promising pathway for laboratory testing of this technology and demonstration of this technology in space.
[ 0, 1, 0, 0, 0, 0 ]
Title: Cyclicity in weighted $\ell^p$ spaces, Abstract: We study the cyclicity in weighted $\ell^p(\mathbb{Z})$ spaces. For $p \geq 1$ and $\beta \geq 0$, let $\ell^p\_\beta(\mathbb{Z})$ be the space of sequences $u=(u\_n)\_{n\in \mathbb{Z}}$ such that $(u\_n |n|^{\beta})\in \ell^p(\mathbb{Z}) $. We obtain both necessary conditions and sufficient conditions for $u$ to be cyclic in $\ell^p\_\beta(\mathbb{Z})$, in other words, for $ \{(u\_{n+k})\_{n \in \mathbb{Z}},~ k \in \mathbb{Z} \}$ to span a dense subspace of $\ell^p\_\beta(\mathbb{Z})$. The conditions are given in terms of the Hausdorff dimension and the capacity of the zero set of the Fourier transform of $u$.
[ 0, 0, 1, 0, 0, 0 ]