text
stringlengths
138
2.38k
labels
sequencelengths
6
6
Predictions
sequencelengths
1
3
Title: Derivative Principal Component Analysis for Representing the Time Dynamics of Longitudinal and Functional Data, Abstract: We propose a nonparametric method to explicitly model and represent the derivatives of smooth underlying trajectories for longitudinal data. This representation is based on a direct Karhunen--Loève expansion of the unobserved derivatives and leads to the notion of derivative principal component analysis, which complements functional principal component analysis, one of the most popular tools of functional data analysis. The proposed derivative principal component scores can be obtained for irregularly spaced and sparsely observed longitudinal data, as typically encountered in biomedical studies, as well as for functional data which are densely measured. Novel consistency results and asymptotic convergence rates for the proposed estimates of the derivative principal component scores and other components of the model are derived under a unified scheme for sparse or dense observations and mild conditions. We compare the proposed representations for derivatives with alternative approaches in simulation settings and also in a wallaby growth curve application. It emerges that representations using the proposed derivative principal component analysis recover the underlying derivatives more accurately compared to principal component analysis-based approaches especially in settings where the functional data are represented with only a very small number of components or are densely sampled. In a second wheat spectra classification example, derivative principal component scores were found to be more predictive for the protein content of wheat than the conventional functional principal component scores.
[ 0, 0, 1, 1, 0, 0 ]
[ "Statistics", "Mathematics" ]
Title: Stokes phenomenon and confluence in non-autonomous Hamiltonian systems, Abstract: This article studies a confluence of a pair of regular singular points to an irregular one in a generic family of time-dependent Hamiltonian systems in dimension 2. This is a general setting for the understanding of the degeneration of the sixth Painleve equation to the fifth one. The main result is a theorem of sectoral normalization of the family to an integrable formal normal form, through which is explained the relation between the local monodromy operators at the two regular singularities and the non-linear Stokes phenomenon at the irregular singularity of the limit system. The problem of analytic classification is also addressed. Key words: Non-autonomous Hamiltonian systems; irregular singularity; non-linear Stokes phenomenon; wild monodromy; confluence; local analytic classification; Painleve equations.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics", "Physics" ]
Title: The independence number of the Birkhoff polytope graph, and applications to maximally recoverable codes, Abstract: Maximally recoverable codes are codes designed for distributed storage which combine quick recovery from single node failure and optimal recovery from catastrophic failure. Gopalan et al [SODA 2017] studied the alphabet size needed for such codes in grid topologies and gave a combinatorial characterization for it. Consider a labeling of the edges of the complete bipartite graph $K_{n,n}$ with labels coming from $F_2^d$ , that satisfies the following condition: for any simple cycle, the sum of the labels over its edges is nonzero. The minimal d where this is possible controls the alphabet size needed for maximally recoverable codes in n x n grid topologies. Prior to the current work, it was known that d is between $(\log n)^2$ and $n\log n$. We improve both bounds and show that d is linear in n. The upper bound is a recursive construction which beats the random construction. The lower bound follows by first relating the problem to the independence number of the Birkhoff polytope graph, and then providing tight bounds for it using the representation theory of the symmetric group.
[ 1, 0, 1, 0, 0, 0 ]
[ "Computer Science", "Mathematics" ]
Title: Spin Angular Momentum of Proton Spin Puzzle in Complex Octonion Spaces, Abstract: The paper focuses on considering some special precessional motions as the spin motions, separating the octonion angular momentum of a proton into six components, elucidating the proton angular momentum in the proton spin puzzle, especially the proton spin, decomposition, quarks and gluons, and polarization and so forth. J. C. Maxwell was the first to use the quaternions to study the electromagnetic fields. Subsequently the complex octonions are utilized to depict the electromagnetic field, gravitational field, and quantum mechanics and so forth. In the complex octonion space, the precessional equilibrium equation infers the angular velocity of precession. The external electromagnetic strength may induce a new precessional motion, generating a new term of angular momentum, even if the orbital angular momentum is zero. This new term of angular momentum can be regarded as the spin angular momentum, and its angular velocity of precession is different from the angular velocity of revolution. The study reveals that the angular momentum of the proton must be separated into more components than ever before. In the proton spin puzzle, the orbital angular momentum and magnetic dipole moment are independent of each other, and they should be measured and calculated respectively.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics", "Mathematics" ]
Title: Physics-Informed Regularization of Deep Neural Networks, Abstract: This paper presents a novel physics-informed regularization method for training of deep neural networks (DNNs). In particular, we focus on the DNN representation for the response of a physical or biological system, for which a set of governing laws are known. These laws often appear in the form of differential equations, derived from first principles, empirically-validated laws, and/or domain expertise. We propose a DNN training approach that utilizes these known differential equations in addition to the measurement data, by introducing a penalty term to the training loss function to penalize divergence form the governing laws. Through three numerical examples, we will show that the proposed regularization produces surrogates that are physically interpretable with smaller generalization errors, when compared to other common regularization methods.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Physics" ]
Title: Gradient-based Filter Design for the Dual-tree Wavelet Transform, Abstract: The wavelet transform has seen success when incorporated into neural network architectures, such as in wavelet scattering networks. More recently, it has been shown that the dual-tree complex wavelet transform can provide better representations than the standard transform. With this in mind, we extend our previous method for learning filters for the 1D and 2D wavelet transforms into the dual-tree domain. We show that with few modifications to our original model, we can learn directional filters that leverage the properties of the dual-tree wavelet transform.
[ 0, 0, 0, 1, 0, 0 ]
[ "Computer Science", "Mathematics" ]
Title: Steganographic Generative Adversarial Networks, Abstract: Steganography is collection of methods to hide secret information ("payload") within non-secret information ("container"). Its counterpart, Steganalysis, is the practice of determining if a message contains a hidden payload, and recovering it if possible. Presence of hidden payloads is typically detected by a binary classifier. In the present study, we propose a new model for generating image-like containers based on Deep Convolutional Generative Adversarial Networks (DCGAN). This approach allows to generate more setganalysis-secure message embedding using standard steganography algorithms. Experiment results demonstrate that the new model successfully deceives the steganography analyzer, and for this reason, can be used in steganographic applications.
[ 1, 0, 0, 1, 0, 0 ]
[ "Computer Science" ]
Title: Markov Chain Lifting and Distributed ADMM, Abstract: The time to converge to the steady state of a finite Markov chain can be greatly reduced by a lifting operation, which creates a new Markov chain on an expanded state space. For a class of quadratic objectives, we show an analogous behavior where a distributed ADMM algorithm can be seen as a lifting of Gradient Descent algorithm. This provides a deep insight for its faster convergence rate under optimal parameter tuning. We conjecture that this gain is always present, as opposed to the lifting of a Markov chain which sometimes only provides a marginal speedup.
[ 1, 0, 1, 1, 0, 0 ]
[ "Computer Science", "Mathematics" ]
Title: Fast failover of multicast sessions in software-defined networks, Abstract: With the rapid growth of services that stream to groups of users comes an increased importance of and demand for reliable multicast. In this paper, we turn to software-defined networking and develop a novel general-purpose multi-failure protection algorithm to provide quick failure recovery, via Fast Failover (FF) groups, for dynamic multicast groups. This extends previous research, which either could not realize fast failover, worked only for single link failures, or was only applicable to static multicast groups. However, while FF is know to be fast, it requires pre-installing back-up rules. These additional memory requirements, which in a multicast setting are even more pronounced than for unicast, are often mentioned as a big disadvantage of using FF. We develop an OpenFlow application for resilient multicast, with which we study FF resource usage, in an attempt to better understand the trade-off between recovery time and resource usage. Our tests on a realistic network suggest that using FF groups can reduce the recovery time of the network significantly compared to other methods, especially when the latency between the controller and the switches is relatively large.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science" ]
Title: Stacked Structure Learning for Lifted Relational Neural Networks, Abstract: Lifted Relational Neural Networks (LRNNs) describe relational domains using weighted first-order rules which act as templates for constructing feed-forward neural networks. While previous work has shown that using LRNNs can lead to state-of-the-art results in various ILP tasks, these results depended on hand-crafted rules. In this paper, we extend the framework of LRNNs with structure learning, thus enabling a fully automated learning process. Similarly to many ILP methods, our structure learning algorithm proceeds in an iterative fashion by top-down searching through the hypothesis space of all possible Horn clauses, considering the predicates that occur in the training examples as well as invented soft concepts entailed by the best weighted rules found so far. In the experiments, we demonstrate the ability to automatically induce useful hierarchical soft concepts leading to deep LRNNs with a competitive predictive power.
[ 1, 0, 0, 1, 0, 0 ]
[ "Computer Science", "Statistics" ]
Title: Spec-QP: Speculative Query Planning for Joins over Knowledge Graphs, Abstract: Organisations store huge amounts of data from multiple heterogeneous sources in the form of Knowledge Graphs (KGs). One of the ways to query these KGs is to use SPARQL queries over a database engine. Since SPARQL follows exact match semantics, the queries may return too few or no results. Recent works have proposed query relaxation where the query engine judiciously replaces a query predicate with similar predicates using weighted relaxation rules mined from the KG. The space of possible relaxations is potentially too large to fully explore and users are typically interested in only top-k results, so such query engines use top-k algorithms for query processing. However, they may still process all the relaxations, many of whose answers do not contribute towards top-k answers. This leads to computation overheads and delayed response times. We propose Spec-QP, a query planning framework that speculatively determines which relaxations will have their results in the top-k answers. Only these relaxations are processed using the top-k operators. We, therefore, reduce the computation overheads and achieve faster response times without adversely affecting the quality of results. We tested Spec-QP over two datasets - XKG and Twitter, to demonstrate the efficiency of our planning framework at reducing runtimes with reasonable accuracy for query engines supporting relaxations.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science" ]
Title: Single Index Latent Variable Models for Network Topology Inference, Abstract: A semi-parametric, non-linear regression model in the presence of latent variables is applied towards learning network graph structure. These latent variables can correspond to unmodeled phenomena or unmeasured agents in a complex system of interacting entities. This formulation jointly estimates non-linearities in the underlying data generation, the direct interactions between measured entities, and the indirect effects of unmeasured processes on the observed data. The learning is posed as regularized empirical risk minimization. Details of the algorithm for learning the model are outlined. Experiments demonstrate the performance of the learned model on real data.
[ 0, 0, 0, 1, 0, 0 ]
[ "Statistics", "Computer Science" ]
Title: Quadratically-Regularized Optimal Transport on Graphs, Abstract: Optimal transportation provides a means of lifting distances between points on a geometric domain to distances between signals over the domain, expressed as probability distributions. On a graph, transportation problems can be used to express challenging tasks involving matching supply to demand with minimal shipment expense; in discrete language, these become minimum-cost network flow problems. Regularization typically is needed to ensure uniqueness for the linear ground distance case and to improve optimization convergence; state-of-the-art techniques employ entropic regularization on the transportation matrix. In this paper, we explore a quadratic alternative to entropic regularization for transport over a graph. We theoretically analyze the behavior of quadratically-regularized graph transport, characterizing how regularization affects the structure of flows in the regime of small but nonzero regularization. We further exploit elegant second-order structure in the dual of this problem to derive an easily-implemented Newton-type optimization algorithm.
[ 1, 0, 1, 0, 0, 0 ]
[ "Computer Science", "Mathematics" ]
Title: Discussion on "Random-projection ensemble classification" by T. Cannings and R. Samworth, Abstract: Discussion on "Random-projection ensemble classification" by T. Cannings and R. Samworth. We believe that the proposed approach can find many applications in economics such as credit scoring (e.g. Altman (1968)) and can be extended to more general type of classifiers. In this discussion we would like to draw authors attention to the copula-based discriminant analysis (Han et al. (2013) and He et al. (2016)).
[ 0, 0, 1, 1, 0, 0 ]
[ "Statistics", "Quantitative Finance" ]
Title: Sub-nanometre resolution of atomic motion during electronic excitation in phase-change materials, Abstract: Phase-change materials based on Ge-Sb-Te alloys are widely used in industrial applications such as nonvolatile memories, but reaction pathways for crystalline-to-amorphous phase-change on picosecond timescales remain unknown. Femtosecond laser excitation and an ultrashort x-ray probe is used to show the temporal separation of electronic and thermal effects in a long-lived ($>$100 ps) transient metastable state of Ge$_{2}$Sb$_{2}$Te$_{5}$ with muted interatomic interaction induced by a weakening of resonant bonding. Due to a specific electronic state, the lattice undergoes a reversible nondestructive modification over a nanoscale region, remaining cold for 4 ps. An independent time-resolved x-ray absorption fine structure experiment confirms the existence of an intermediate state with disordered bonds. This newly unveiled effect allows the utilization of non-thermal ultra-fast pathways enabling artificial manipulation of the switching process, ultimately leading to a redefined speed limit, and improved energy efficiency and reliability of phase-change memory technologies.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: A Density Result for Real Hyperelliptic Curves, Abstract: Let $\{\infty^+, \infty^-\}$ be the two points above $\infty$ on the real hyperelliptic curve $H: y^2 = (x^2 - 1) \prod_{i=1}^{2g} (x - a_i)$. We show that the divisor $([\infty^+] - [\infty^-])$ is torsion in $\operatorname{Jac} J$ for a dense set of $(a_1, a_2, \ldots, a_{2g}) \in (-1, 1)^{2g}$. In fact, we prove by degeneration to a nodal $\mathbb{P}^1$ that an associated period map has derivative generically of full rank.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics" ]
Title: On a class of integrable systems of Monge-Ampère type, Abstract: We investigate a class of multi-dimensional two-component systems of Monge-Ampère type that can be viewed as generalisations of heavenly-type equations appearing in self-dual Ricci-flat geometry. Based on the Jordan-Kronecker theory of skew-symmetric matrix pencils, a classification of normal forms of such systems is obtained. All two-component systems of Monge-Ampère type turn out to be integrable, and can be represented as the commutativity conditions of parameter-dependent vector fields. Geometrically, systems of Monge-Ampère type are associated with linear sections of the Grassmannians. This leads to an invariant differential-geometric characterisation of the Monge-Ampère property.
[ 0, 1, 1, 0, 0, 0 ]
[ "Mathematics", "Physics" ]
Title: Unsupervised Motion Artifact Detection in Wrist-Measured Electrodermal Activity Data, Abstract: One of the main benefits of a wrist-worn computer is its ability to collect a variety of physiological data in a minimally intrusive manner. Among these data, electrodermal activity (EDA) is readily collected and provides a window into a person's emotional and sympathetic responses. EDA data collected using a wearable wristband are easily influenced by motion artifacts (MAs) that may significantly distort the data and degrade the quality of analyses performed on the data if not identified and removed. Prior work has demonstrated that MAs can be successfully detected using supervised machine learning algorithms on a small data set collected in a lab setting. In this paper, we demonstrate that unsupervised learning algorithms perform competitively with supervised algorithms for detecting MAs on EDA data collected in both a lab-based setting and a real-world setting comprising about 23 hours of data. We also find, somewhat surprisingly, that incorporating accelerometer data as well as EDA improves detection accuracy only slightly for supervised algorithms and significantly degrades the accuracy of unsupervised algorithms.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Statistics", "Quantitative Biology" ]
Title: Early Experiences with Crowdsourcing Airway Annotations in Chest CT, Abstract: Measuring airways in chest computed tomography (CT) images is important for characterizing diseases such as cystic fibrosis, yet very time-consuming to perform manually. Machine learning algorithms offer an alternative, but need large sets of annotated data to perform well. We investigate whether crowdsourcing can be used to gather airway annotations which can serve directly for measuring the airways, or as training data for the algorithms. We generate image slices at known locations of airways and request untrained crowd workers to outline the airway lumen and airway wall. Our results show that the workers are able to interpret the images, but that the instructions are too complex, leading to many unusable annotations. After excluding unusable annotations, quantitative results show medium to high correlations with expert measurements of the airways. Based on this positive experience, we describe a number of further research directions and provide insight into the challenges of crowdsourcing in medical images from the perspective of first-time users.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Quantitative Biology" ]
Title: Convergent Iteration in Sobolev Space for Time Dependent Closed Quantum Systems, Abstract: Time dependent quantum systems have become indispensable in science and its applications, particularly at the atomic and molecular levels. Here, we discuss the approximation of closed time dependent quantum systems on bounded domains, via iterative methods in Sobolev space based upon evolution operators. Recently, existence and uniqueness of weak solutions were demonstrated by a contractive fixed point mapping defined by the evolution operators. Convergent successive approximation is then guaranteed. This article uses the same mapping to define quadratically convergent Newton and approximate Newton methods. Estimates for the constants used in the convergence estimates are provided. The evolution operators are ideally suited to serve as the framework for this operator approximation theory, since the Hamiltonian is time dependent. In addition, the hypotheses required to guarantee quadratic convergence of the Newton iteration build naturally upon the hypotheses used for the existence/uniqueness theory.
[ 0, 0, 1, 0, 0, 0 ]
[ "Physics", "Mathematics" ]
Title: Image-based immersed boundary model of the aortic root, Abstract: Each year, approximately 300,000 heart valve repair or replacement procedures are performed worldwide, including approximately 70,000 aortic valve replacement surgeries in the United States alone. This paper describes progress in constructing anatomically and physiologically realistic immersed boundary (IB) models of the dynamics of the aortic root and ascending aorta. This work builds on earlier IB models of fluid-structure interaction (FSI) in the aortic root, which previously achieved realistic hemodynamics over multiple cardiac cycles, but which also were limited to simplified aortic geometries and idealized descriptions of the biomechanics of the aortic valve cusps. By contrast, the model described herein uses an anatomical geometry reconstructed from patient-specific computed tomography angiography (CTA) data, and employs a description of the elasticity of the aortic valve leaflets based on a fiber-reinforced constitutive model fit to experimental tensile test data. Numerical tests show that the model is able to resolve the leaflet biomechanics in diastole and early systole at practical grid spacings. The model is also used to examine differences in the mechanics and fluid dynamics yielded by fresh valve leaflets and glutaraldehyde-fixed leaflets similar to those used in bioprosthetic heart valves. Although there are large differences in the leaflet deformations during diastole, the differences in the open configurations of the valve models are relatively small, and nearly identical hemodynamics are obtained in all cases considered.
[ 1, 1, 0, 0, 0, 0 ]
[ "Quantitative Biology", "Physics" ]
Title: A Data-Driven Analysis of the Influence of Care Coordination on Trauma Outcome, Abstract: OBJECTIVE: To test the hypothesis that variation in care coordination is related to LOS. DESIGN We applied a spectral co-clustering methodology to simultaneously infer groups of patients and care coordination patterns, in the form of interaction networks of health care professionals, from electronic medical record (EMR) utilization data. The care coordination pattern for each patient group was represented by standard social network characteristics and its relationship with hospital LOS was assessed via a negative binomial regression with a 95% confidence interval. SETTING AND PATIENTS This study focuses on 5,588 adult patients hospitalized for trauma at the Vanderbilt University Medical Center. The EMRs were accessed by healthcare professionals from 179 operational areas during 158,467 operational actions. MAIN OUTCOME MEASURES: Hospital LOS for trauma inpatients, as an indicator of care coordination efficiency. RESULTS: Three general types of care coordination patterns were discovered, each of which was affiliated with a specific patient group. The first patient group exhibited the shortest hospital LOS and was managed by a care coordination pattern that involved the smallest number of operational areas (102 areas, as opposed to 125 and 138 for the other patient groups), but exhibited the largest number of collaborations between operational areas (e.g., an average of 27.1 connections per operational area compared to 22.5 and 23.3 for the other two groups). The hospital LOS for the second and third patient groups was 14 hours (P = 0.024) and 10 hours (P = 0.042) longer than the first patient group, respectively.
[ 1, 0, 0, 0, 0, 0 ]
[ "Statistics", "Quantitative Biology" ]
Title: Construction of Non-asymptotic Confidence Sets in 2-Wasserstein Space, Abstract: In this paper, we consider a probabilistic setting where the probability measures are considered to be random objects. We propose a procedure of construction non-asymptotic confidence sets for empirical barycenters in 2-Wasserstein space and develop the idea further to construction of a non-parametric two-sample test that is then applied to the detection of structural breaks in data with complex geometry. Both procedures mainly rely on the idea of multiplier bootstrap (Spokoiny and Zhilova (2015), Chernozhukov et al. (2014)). The main focus lies on probability measures that have commuting covariance matrices and belong to the same scatter-location family: we proof the validity of a bootstrap procedure that allows to compute confidence sets and critical values for a Wasserstein-based two-sample test.
[ 0, 0, 1, 1, 0, 0 ]
[ "Mathematics", "Statistics" ]
Title: Limits on statistical anisotropy from BOSS DR12 galaxies using bipolar spherical harmonics, Abstract: We measure statistically anisotropic signatures imprinted in three-dimensional galaxy clustering using bipolar spherical harmonics (BipoSHs) in both Fourier space and configuration space. We then constrain a well-known quadrupolar anisotropy parameter $g_{2M}$ in the primordial power spectrum, parametrized by $P(\vec{k}) = \bar{P}(k) [ 1 + \sum_{M} g_{2M} Y_{2M}(\hat{k}) ]$, with $M$ determining the direction of the anisotropy. Such an anisotropic signal is easily contaminated by artificial asymmetries due to specific survey geometry. We precisely estimate the contaminated signal and finally subtract it from the data. Using the galaxy samples obtained by the Baryon Oscillation Spectroscopic Survey Data Release 12, we find no evidence for violation of statistical isotropy, $g_{2M}$ for all $M$ to be of zero within the $2\sigma$ level. The $g_{2M}$-type anisotropy can originate from the primordial curvature power spectrum involving a directional-dependent modulation $g_* (\hat{k} \cdot \hat{p})^2$. The bound on $g_{2M}$ is translated into $g_*$ as $-0.09 < g_* < 0.08$ with a $95\%$ confidence level when $\hat{p}$ is marginalized over.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics", "Statistics" ]
Title: Confidence Intervals for Stochastic Arithmetic, Abstract: Quantifying errors and losses due to the use of Floating-Point (FP) calculations in industrial scientific computing codes is an important part of the Verification, Validation and Uncertainty Quantification (VVUQ) process. Stochastic Arithmetic is one way to model and estimate FP losses of accuracy, which scales well to large, industrial codes. It exists in different flavors, such as CESTAC or MCA, implemented in various tools such as CADNA, Verificarlo or Verrou. These methodologies and tools are based on the idea that FP losses of accuracy can be modeled via randomness. Therefore, they share the same need to perform a statistical analysis of programs results in order to estimate the significance of the results. In this paper, we propose a framework to perform a solid statistical analysis of Stochastic Arithmetic. This framework unifies all existing definitions of the number of significant digits (CESTAC and MCA), and also proposes a new quantity of interest: the number of digits contributing to the accuracy of the results. Sound confidence intervals are provided for all estimators, both in the case of normally distributed results, and in the general case. The use of this framework is demonstrated by two case studies of large, industrial codes: Europlexus and code\_aster.
[ 1, 0, 0, 1, 0, 0 ]
[ "Computer Science", "Statistics" ]
Title: Co-segmentation for Space-Time Co-located Collections, Abstract: We present a co-segmentation technique for space-time co-located image collections. These prevalent collections capture various dynamic events, usually by multiple photographers, and may contain multiple co-occurring objects which are not necessarily part of the intended foreground object, resulting in ambiguities for traditional co-segmentation techniques. Thus, to disambiguate what the common foreground object is, we introduce a weakly-supervised technique, where we assume only a small seed, given in the form of a single segmented image. We take a distributed approach, where local belief models are propagated and reinforced with similar images. Our technique progressively expands the foreground and background belief models across the entire collection. The technique exploits the power of the entire set of image without building a global model, and thus successfully overcomes large variability in appearance of the common foreground object. We demonstrate that our method outperforms previous co-segmentation techniques on challenging space-time co-located collections, including dense benchmark datasets which were adapted for our novel problem setting.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science" ]
Title: Connecting pairwise spheres by depth: DCOPS, Abstract: We extend the classical notion of the spherical depth in \mathbb{R}^k, to the important setup of data on a Riemannian manifold. We show that this notion of depth satisfies a set of desirable properties. For the empirical version of this depth function both uniform consistency and the asymptotic distribution are studied. Consistency is also shown for functional data. The behaviour of the depth is illustrated through several examples for Riemannian manifold data.
[ 0, 0, 1, 1, 0, 0 ]
[ "Mathematics", "Statistics" ]
Title: TRAMP: Tracking by a Real-time AMbisonic-based Particle filter, Abstract: This article presents a multiple sound source localization and tracking system, fed by the Eigenmike array. The First Order Ambisonics (FOA) format is used to build a pseudointensity-based spherical histogram, from which the source position estimates are deduced. These instantaneous estimates are processed by a wellknown tracking system relying on a set of particle filters. While the novelty within localization and tracking is incremental, the fully-functional, complete and real-time running system based on these algorithms is proposed for the first time. As such, it could serve as an additional baseline method of the LOCATA challenge.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science" ]
Title: General Bounds for Incremental Maximization, Abstract: We propose a theoretical framework to capture incremental solutions to cardinality constrained maximization problems. The defining characteristic of our framework is that the cardinality/support of the solution is bounded by a value $k\in\mathbb{N}$ that grows over time, and we allow the solution to be extended one element at a time. We investigate the best-possible competitive ratio of such an incremental solution, i.e., the worst ratio over all $k$ between the incremental solution after $k$ steps and an optimum solution of cardinality $k$. We define a large class of problems that contains many important cardinality constrained maximization problems like maximum matching, knapsack, and packing/covering problems. We provide a general $2.618$-competitive incremental algorithm for this class of problems, and show that no algorithm can have competitive ratio below $2.18$ in general. In the second part of the paper, we focus on the inherently incremental greedy algorithm that increases the objective value as much as possible in each step. This algorithm is known to be $1.58$-competitive for submodular objective functions, but it has unbounded competitive ratio for the class of incremental problems mentioned above. We define a relaxed submodularity condition for the objective function, capturing problems like maximum (weighted) ($b$-)matching and a variant of the maximum flow problem. We show that the greedy algorithm has competitive ratio (exactly) $2.313$ for the class of problems that satisfy this relaxed submodularity condition. Note that our upper bounds on the competitive ratios translate to approximation ratios for the underlying cardinality constrained problems.
[ 1, 0, 1, 0, 0, 0 ]
[ "Computer Science", "Mathematics" ]
Title: Scalable methods for Bayesian selective inference, Abstract: Modeled along the truncated approach in Panigrahi (2016), selection-adjusted inference in a Bayesian regime is based on a selective posterior. Such a posterior is determined together by a generative model imposed on data and the selection event that enforces a truncation on the assumed law. The effective difference between the selective posterior and the usual Bayesian framework is reflected in the use of a truncated likelihood. The normalizer of the truncated law in the adjusted framework is the probability of the selection event; this is typically intractable and it leads to the computational bottleneck in sampling from such a posterior. The current work lays out a primal-dual approach of solving an approximating optimization problem to provide valid post-selective Bayesian inference. The selection procedures are posed as data-queries that solve a randomized version of a convex learning program which have the advantage of preserving more left-over information for inference. We propose a randomization scheme under which the optimization has separable constraints that result in a partially separable objective in lower dimensions for many commonly used selective queries to approximate the otherwise intractable selective posterior. We show that the approximating optimization under a Gaussian randomization gives a valid exponential rate of decay for the selection probability on a large deviation scale. We offer a primal-dual method to solve the optimization problem leading to an approximate posterior; this allows us to exploit the usual merits of a Bayesian machinery in both low and high dimensional regimes where the underlying signal is effectively sparse. We show that the adjusted estimates empirically demonstrate better frequentist properties in comparison to the unadjusted estimates based on the usual posterior, when applied to a wide range of constrained, convex data queries.
[ 0, 0, 0, 1, 0, 0 ]
[ "Statistics", "Mathematics", "Computer Science" ]
Title: Modified Recursive Cholesky (Rchol) Algorithm: An Explicit Estimation and Pseudo-inverse of Correlation Matrices, Abstract: The Cholesky decomposition plays an important role in finding the inverse of the correlation matrices. As it is a fast and numerically stable for linear system solving, inversion, and factorization compared to singular valued decomposition (SVD), QR factorization and LU decomposition. As different methods exist to find the Cholesky decomposition of a given matrix. This paper presents the comparative study of a proposed RChol algorithm with the conventional methods. The RChol algorithm is an explicit way to estimate the modified Cholesky factors of a dynamic correlation matrix.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics", "Statistics", "Computer Science" ]
Title: A Utility-Driven Multi-Queue Admission Control Solution for Network Slicing, Abstract: The combination of recent emerging technologies such as network function virtualization (NFV) and network programmability (SDN) gave birth to the Network Slicing revolution. 5G networks consist of multi-tenant infrastructures capable of offering leased network "slices" to new customers (e.g., vertical industries) enabling a new telecom business model: Slice-as-aService (SlaaS). In this paper, we aim i ) to study the slicing admission control problem by means of a multi-queuing system for heterogeneous tenant requests, ii ) to derive its statistical behavior model, and iii ) to provide a utility-based admission control optimization. Our results analyze the capability of the proposed SlaaS system to be approximately Markovian and evaluate its performance as compared to legacy solutions.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Statistics" ]
Title: Statistical inference for network samples using subgraph counts, Abstract: We consider that a network is an observation, and a collection of observed networks forms a sample. In this setting, we provide methods to test whether all observations in a network sample are drawn from a specified model. We achieve this by deriving, under the null of the graphon model, the joint asymptotic properties of average subgraph counts as the number of observed networks increases but the number of nodes in each network remains finite. In doing so, we do not require that each observed network contains the same number of nodes, or is drawn from the same distribution. Our results yield joint confidence regions for subgraph counts, and therefore methods for testing whether the observations in a network sample are drawn from: a specified distribution, a specified model, or from the same model as another network sample. We present simulation experiments and an illustrative example on a sample of brain networks where we find that highly creative individuals' brains present significantly more short cycles.
[ 1, 0, 0, 1, 0, 0 ]
[ "Statistics", "Mathematics", "Quantitative Biology" ]
Title: Data Race Detection on Compressed Traces, Abstract: We consider the problem of detecting data races in program traces that have been compressed using straight line programs (SLP), which are special context-free grammars that generate exactly one string, namely the trace that they represent. We consider two classical approaches to race detection --- using the happens-before relation and the lockset discipline. We present algorithms for both these methods that run in time that is linear in the size of the compressed, SLP representation. Typical program executions almost always exhibit patterns that lead to significant compression. Thus, our algorithms are expected to result in large speedups when compared with analyzing the uncompressed trace. Our experimental evaluation of these new algorithms on standard benchmarks confirms this observation.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science" ]
Title: Two-Party Function Computation on the Reconciled Data, Abstract: In this paper, we initiate a study of a new problem termed function computation on the reconciled data, which generalizes a set reconciliation problem in the literature. Assume a distributed data storage system with two users $A$ and $B$. The users possess a collection of binary vectors $S_{A}$ and $S_{B}$, respectively. They are interested in computing a function $\phi$ of the reconciled data $S_{A} \cup S_{B}$. It is shown that any deterministic protocol, which computes a sum and a product of reconciled sets of binary vectors represented as nonnegative integers, has to communicate at least $2^n + n - 1$ and $2^n + n - 2$ bits in the worst-case scenario, respectively, where $n$ is the length of the binary vectors. Connections to other problems in computer science, such as set disjointness and finding the intersection, are established, yielding a variety of additional upper and lower bounds on the communication complexity. A protocol for computation of a sum function, which is based on use of a family of hash functions, is presented, and its characteristics are analyzed.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Mathematics" ]
Title: Halo assembly bias and the tidal anisotropy of the local halo environment, Abstract: We study the role of the local tidal environment in determining the assembly bias of dark matter haloes. Previous results suggest that the anisotropy of a halo's environment (i.e, whether it lies in a filament or in a more isotropic region) can play a significant role in determining the eventual mass and age of the halo. We statistically isolate this effect using correlations between the large-scale and small-scale environments of simulated haloes at $z=0$ with masses between $10^{11.6}\lesssim (m/h^{-1}M_{\odot})\lesssim10^{14.9}$. We probe the large-scale environment using a novel halo-by-halo estimator of linear bias. For the small-scale environment, we identify a variable $\alpha_R$ that captures the $\textit{tidal anisotropy}$ in a region of radius $R=4R_{\textrm{200b}}$ around the halo and correlates strongly with halo bias at fixed mass. Segregating haloes by $\alpha_R$ reveals two distinct populations. Haloes in highly isotropic local environments ($\alpha_R\lesssim0.2$) behave as expected from the simplest, spherically averaged analytical models of structure formation, showing a $\textit{negative}$ correlation between their concentration and large-scale bias at $\textit{all}$ masses. In contrast, haloes in anisotropic, filament-like environments ($\alpha_R\gtrsim0.5$) tend to show a $\textit{positive}$ correlation between bias and concentration at any mass. Our multi-scale analysis cleanly demonstrates how the overall assembly bias trend across halo mass emerges as an average over these different halo populations, and provides valuable insights towards building analytical models that correctly incorporate assembly bias. We also discuss potential implications for the nature and detectability of galaxy assembly bias.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: Towards a general theory for non-linear locally stationary processes, Abstract: In this paper some general theory is presented for locally stationary processes based on the stationary approximation and the stationary derivative. Laws of large numbers, central limit theorems as well as deterministic and stochastic bias expansions are proved for processes obeying an expansion in terms of the stationary approximation and derivative. In addition it is shown that this applies to some general nonlinear non-stationary Markov-models. In addition the results are applied to derive the asymptotic properties of maximum likelihood estimates of parameter curves in such models.
[ 0, 0, 1, 1, 0, 0 ]
[ "Statistics", "Mathematics" ]
Title: Single-cell diffraction tomography with optofluidic rotation about a tilted axis, Abstract: Optical diffraction tomography (ODT) is a tomographic technique that can be used to measure the three-dimensional (3D) refractive index distribution within living cells without the requirement of any marker. In principle, ODT can be regarded as a generalization of optical projection tomography which is equivalent to computerized tomography (CT). Both optical tomographic techniques require projection-phase images of cells measured at multiple angles. However, the reconstruction of the 3D refractive index distribution post-measurement differs for the two techniques. It is known that ODT yields better results than projection tomography, because it takes into account diffraction of the imaging light due to the refractive index structure of the sample. Here, we apply ODT to biological cells in a microfluidic chip which combines optical trapping and microfluidic flow to achieve an optofluidic single-cell rotation. In particular, we address the problem that arises when the trapped cell is not rotating about an axis perpendicular to the imaging plane, but instead about an arbitrarily tilted axis. In this paper we show that the 3D reconstruction can be improved by taking into account such a tilted rotational axis in the reconstruction process.
[ 0, 0, 0, 0, 1, 0 ]
[ "Physics", "Quantitative Biology" ]
Title: Maps on statistical manifolds exactly reduced from the Perron-Frobenius equations for solvable chaotic maps, Abstract: Maps on a parameter space for expressing distribution functions are exactly derived from the Perron-Frobenius equations for a generalized Boole transform family. Here the generalized Boole transform family is a one-parameter family of maps where it is defined on a subset of the real line and its probability distribution function is the Cauchy distribution with some parameters. With this reduction, some relations between the statistical picture and the orbital one are shown. From the viewpoint of information geometry, the parameter space can be identified with a statistical manifold, and then it is shown that the derived maps can be characterized. Also, with an induced symplectic structure from a statistical structure, symplectic and information geometric aspects of the derived maps are discussed.
[ 0, 1, 0, 0, 0, 0 ]
[ "Mathematics", "Statistics", "Physics" ]
Title: Sparse Matrix Multiplication On An Associative Processor, Abstract: Sparse matrix multiplication is an important component of linear algebra computations. Implementing sparse matrix multiplication on an associative processor (AP) enables high level of parallelism, where a row of one matrix is multiplied in parallel with the entire second matrix, and where the execution time of vector dot product does not depend on the vector size. Four sparse matrix multiplication algorithms are explored in this paper, combining AP and baseline CPU processing to various levels. They are evaluated by simulation on a large set of sparse matrices. The computational complexity of sparse matrix multiplication on AP is shown to be an O(nnz) where nnz is the number of nonzero elements. The AP is found to be especially efficient in binary sparse matrix multiplication. AP outperforms conventional solutions in power efficiency.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Mathematics" ]
Title: On the intersection of tame subgroups in groups acting on trees, Abstract: Let $G$ be a group acting on a tree $T$ with finite edge stabilizers of bounded order. We provide, in some very interesting cases, upper bounds for the complexity of the intersection $H\cap K$ of two tame subgroups $H$ and $K$ of $G$ in terms of the complexities of $H$ and $K$. In particular, we obtain bounds for the Kurosh rank $Kr(H\cap K)$ of the intersection in terms of Kurosh ranks $Kr(H)$ and $Kr(K)$, in the case where $H$ and $K$ act freely on the edges of $T$.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics" ]
Title: Digital Advertising Traffic Operation: Flow Management Analysis, Abstract: In a Web Advertising Traffic Operation the Trafficking Routing Problem (TRP) consists in scheduling the management of Web Advertising (Adv) campaign between Trafficking campaigns in the most efficient way to oversee and manage relationship with partners and internal teams, managing expectations through integration and post-launch in order to ensure success for every stakeholders involved. For our own interest we did that independent research projects also through specific innovative tasks validate towards average working time declared on "specification required" by the main worldwide industry leading Advertising Agency. We present a Mixed Integer Linear Programming (MILP) formulation for end-to-end management of campaign workflow along a predetermined path and generalize it to include alternative path to oversee and manage detail-oriented relationship with partners and internal teams to achieve the goals above mentioned. To meet clients' KPIs, we consider an objective function that includes the punctuality indicators (the average waiting time and completion times) but also the main punctuality indicators (the average delay and the on time performance). Then we investigate their analytical relationships in the advertising domain through experiments based on real data from a Traffic Office. We show that the classic punctuality indicators are in contradiction with the task of reducing waiting times. We propose new indicators used for a synthesize analysis on projects or process changes for the wider team that are more sustainable, but also more relevant for stakeholders. We also show that the flow of a campaign (adv-ways) is the main bottleneck of a Traffic Office and that alternate paths cannot improve the performance indicators.
[ 1, 0, 1, 0, 0, 0 ]
[ "Computer Science", "Statistics", "Quantitative Finance" ]
Title: Addressing Class Imbalance in Classification Problems of Noisy Signals by using Fourier Transform Surrogates, Abstract: Randomizing the Fourier-transform (FT) phases of temporal-spatial data generates surrogates that approximate examples from the data-generating distribution. We propose such FT surrogates as a novel tool to augment and analyze training of neural networks and explore the approach in the example of sleep-stage classification. By computing FT surrogates of raw EEG, EOG, and EMG signals of under-represented sleep stages, we balanced the CAPSLPDB sleep database. We then trained and tested a convolutional neural network for sleep stage classification, and found that our surrogate-based augmentation improved the mean F1-score by 7%. As another application of FT surrogates, we formulated an approach to compute saliency maps for individual sleep epochs. The visualization is based on the response of inferred class probabilities under replacement of short data segments by partial surrogates. To quantify how well the distributions of the surrogates and the original data match, we evaluated a trained classifier on surrogates of correctly classified examples, and summarized these conditional predictions in a confusion matrix. We show how such conditional confusion matrices can qualitatively explain the performance of surrogates in class balancing. The FT-surrogate augmentation approach may improve classification on noisy signals if carefully adapted to the data distribution under analysis.
[ 0, 0, 0, 1, 1, 0 ]
[ "Computer Science", "Statistics" ]
Title: Simultaneous smoothness and simultaneous stability of a $C^\infty$ strictly convex integrand and its dual, Abstract: In this paper, we investigate simultaneous properties of a convex integrand $\gamma$ and its dual $\delta$. The main results are the following three. (1) For a $C^\infty$ convex integrand $\gamma: S^n\to \mathbb{R}_+$, its dual convex integrand $\delta: S^n\to \mathbb{R}_+$ is of class $C^\infty$ if and only if $\gamma$ is a strictly convex integrand. (2) Let $\gamma: S^n\to \mathbb{R}_+$ be a $C^\infty$ strictly convex integrand. Then, $\gamma$ is stable if and only if its dual convex integrand $\delta: S^n\to \mathbb{R}_+$ is stable. (3) Let $\gamma: S^n\to \mathbb{R}_+$ be a $C^\infty$ strictly convex integrand. Suppose that $\gamma$ is stable. Then, for any $i$ $(0\le i\le n)$, a point $\theta_0\in S^n$ is a non-degenerate critical point of $\gamma$ with Morse index $i$ if and only if its antipodal point $-\theta_0\in S^n$ is a non-degenerate critical point of the dual convex integrand $\delta$ with Morse index $(n-i)$.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics" ]
Title: Anomalous transport effects on switching currents of graphene-based Josephson junctions, Abstract: We explore the effect of noise on the ballistic graphene-based small Josephson junctions in the framework of the resistively and capacitively shunted model. We use the non-sinusoidal current-phase relation specific for graphene layers partially covered by superconducting electrodes. The noise induced escapes from the metastable states, when the external bias current is ramped, give the switching current distribution, i.e. the probability distribution of the passages to finite voltage from the superconducting state as a function of the bias current, that is the information more promptly available in the experiments. We consider a noise source that is a mixture of two different types of processes: a Gaussian contribution to simulate an uncorrelated ordinary thermal bath, and non-Gaussian, $\alpha$-stable (or Lévy) term, generally associated to non-equilibrium transport phenomena. We find that the analysis of the switching current distribution makes it possible to efficiently detect a non-Gaussian noise component in a Gaussian background.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: PbTe(111) Sub-Thermionic Photocathode: A Route to High-Quality Electron Pulses, Abstract: The emission properties of PbTe(111) single crystal have been extensively investigated to demonstrate that PbTe(111) is a promising low root mean square transverse momentum ({\Delta}p$_T$) and high brightness photocathode. The density functional theory (DFT) based photoemission analysis successfully elucidates that the 'hole-like' {\Lambda}$^+_6$ energy band in the $L$ valley with low effective mass $m^*$ results in low {\Delta}p$_T$. Especially, as a 300K solid planar photocathode, Te-terminated PbTe(111) single crystal is expected to be a potential 50K electron source.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: Envy-free Matchings with Lower Quotas, Abstract: While every instance of the Hospitals/Residents problem admits a stable matching, the problem with lower quotas (HR-LQ) has instances with no stable matching. For such an instance, we expect the existence of an envy-free matching, which is a relaxation of a stable matching preserving a kind of fairness property. In this paper, we investigate the existence of an envy-free matching in several settings, in which hospitals have lower quotas and not all doctor-hospital pairs are acceptable. We first show that, for an HR-LQ instance, we can efficiently decide the existence of an envy-free matching. Then, we consider envy-freeness in the Classified Stable Matching model due to Huang (2010), i.e., each hospital has lower and upper quotas on subsets of doctors. We show that, for this model, deciding the existence of an envy-free matching is NP-hard in general, but solvable in polynomial time if quotas are paramodular.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Mathematics" ]
Title: Interacting Multi-particle Classical Szilard Engine, Abstract: Szilard engine(SZE) is one of the best example of how information can be used to extract work from a system. Initially, the working substance of SZE was considered to be a single particle. Later on, researchers has extended the studies of SZE to multi-particle systems and even to quantum regime. Here we present a detailed study of classical SZE consisting of $N$ particles with inter-particle interactions, i.e., the working substance is a low density non-ideal gas and compare the work extraction with respect to SZE with non-interacting multi particle system as working substance. We have considered two cases of interactions namely: (i) hard core interactions and (ii) square well interaction. Our study reveals that work extraction is less when more particles are interacting through hard core interactions. More work is extracted when the particles are interacting via square well interaction. Another important result for the second case is that as we increase the particle number the work extraction becomes independent of the initial position of the partition, as opposed to the first case. Work extraction depends crucially on the initial position of the partition. More work can be extracted with larger number of particles when partition is inserted at positions near the boundary walls.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: Graph of Virtual Actors (GOVA): a Big Data Analytics Architecture for IoT, Abstract: With the emergence of cloud computing and sensor technologies, Big Data analytics for the Internet of Things (IoT) has become the main force behind many innovative solutions for our society's problems. This paper provides practical explanations for the question "why is the number of Big Data applications that succeed and have an effect on our daily life so limited, compared with all of the solutions proposed and tested in the literature?", with examples taken from Smart Grids. We argue that "noninvariants" are the most challenging issues in IoT applications, which can be easily revealed if we use the term "invariant" to replace the more common terms such as "information", "knowledge", or "insight" in any Big Data for IoT research. From our experience with developing Smart Grid applications, we produced a list of "noninvariants", which we believe to be the main causes of the gaps between Big Data in a laboratory and in practice in IoT applications. This paper also proposes Graph of Virtual Actors (GOVA) as a Big Data analytics architecture for IoT applications, which not only can solve the noninvariants issues, but can also quickly scale horizontally in terms of computation, data storage, caching requirements, and programmability of the system.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science" ]
Title: A Variation of the $q$-Painlevé System with Affine Weyl Group Symmetry of Type $E_7^{(1)}$, Abstract: Recently a certain $q$-Painlevé type system has been obtained from a reduction of the $q$-Garnier system. In this paper it is shown that the $q$-Painlevé type system is associated with another realization of the affine Weyl group symmetry of type $E_7^{(1)}$ and is different from the well-known $q$-Painlevé system of type $E_7^{(1)}$ from the point of view of evolution directions. We also study a connection between the $q$-Painlevé type system and the $q$-Painlevé system of type $E_7^{(1)}$. Furthermore determinant formulas of particular solutions for the $q$-Painlevé type system are constructed in terms of the terminating $q$-hypergeometric function.
[ 0, 1, 1, 0, 0, 0 ]
[ "Mathematics" ]
Title: Riemannian geometry in infinite dimensional spaces, Abstract: We lay foundations of the subject in the title, on which we build in another paper devoted to isometries in spaces of Kähler metrics.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics" ]
Title: Full-angle Negative Reflection with An Ultrathin Acoustic Gradient Metasurface: Floquet-Bloch Modes Perspective and Experimental Verification, Abstract: Metasurface with gradient phase response offers new alternative for steering the propagation of waves. Conventional Snell's law has been revised by taking the contribution of local phase gradient into account. However, the requirement of momentum matching along the metasurface sets its nontrivial beam manipulation functionality within a limited-angle incidence. In this work, we theoretically and experimentally demonstrate that the acoustic gradient metasurface supports the negative reflection for full-angle incidence. The mode expansion theory is developed to help understand how the gradient metasurface tailors the incident beams, and the full-angle negative reflection occurs when the first negative order Floquet-Bloch mode dominates. The coiling-up space structures are utilized to build desired acoustic gradient metasurface and the full-angle negative reflections have been perfectly verified by experimental measurements. Our work offers the Floquet-Bloch modes perspective for qualitatively understanding the reflection behaviors of the acoustic gradient metasurface and enables a new degree of the acoustic wave manipulating.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: Semigroup C*-algebras and toric varieties, Abstract: Let S be a finitely generated subsemigroup of Z^2. We derive a general formula for the K-theory of the left regular C*-algebra for S.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics" ]
Title: Error analysis for small-sample, high-variance data: Cautions for bootstrapping and Bayesian bootstrapping, Abstract: Recent advances in molecular simulations allow the direct evaluation of kinetic parameters such as rate constants for protein folding or unfolding. However, these calculations are usually computationally expensive and even significant computing resources may result in a small number of independent rate estimates spread over many orders of magnitude. Such small, high-variance samples are not readily amenable to analysis using the standard uncertainty ("standard error of the mean") because unphysical negative limits of confidence intervals result. Bootstrapping, a natural alternative guaranteed to yield a confidence interval within the minimum and maximum values, also exhibits a striking systematic bias of the lower confidence limit. As we show, bootstrapping artifactually assigns high probability to improbably low mean values. A second alternative, the Bayesian bootstrap strategy, does not suffer from the same deficit and is more logically consistent with the type of confidence interval desired, but must be used with caution nevertheless. Neither standard nor Bayesian bootstrapping can overcome the intrinsic challenge of under-estimating the mean from small, high-variance samples. Our report is based on extensive re-analysis of multiple estimates for rate constants obtained from independent atomistic simulations. Although we only analyze rate constants, similar considerations may apply to other types of high-variance calculations, such as may occur in highly non-linear averages like the Jarzynski relation.
[ 0, 0, 0, 1, 0, 0 ]
[ "Statistics", "Quantitative Biology" ]
Title: Computing Influence of a Product through Uncertain Reverse Skyline, Abstract: Understanding the influence of a product is crucially important for making informed business decisions. This paper introduces a new type of skyline queries, called uncertain reverse skyline, for measuring the influence of a probabilistic product in uncertain data settings. More specifically, given a dataset of probabilistic products P and a set of customers C, an uncertain reverse skyline of a probabilistic product q retrieves all customers c in C which include q as one of their preferred products. We present efficient pruning ideas and techniques for processing the uncertain reverse skyline query of a probabilistic product using R-Tree data index. We also present an efficient parallel approach to compute the uncertain reverse skyline and influence score of a probabilistic product. Our approach significantly outperforms the baseline approach derived from the existing literature. The efficiency of our approach is demonstrated by conducting extensive experiments with both real and synthetic datasets.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Statistics" ]
Title: A branch-and-price approach with MILP formulation to modularity density maximization on graphs, Abstract: For clustering of an undirected graph, this paper presents an exact algorithm for the maximization of modularity density, a more complicated criterion to overcome drawbacks of the well-known modularity. The problem can be interpreted as the set-partitioning problem, which reminds us of its integer linear programming (ILP) formulation. We provide a branch-and-price framework for solving this ILP, or column generation combined with branch-and-bound. Above all, we formulate the column generation subproblem to be solved repeatedly as a simpler mixed integer linear programming (MILP) problem. Acceleration techniques called the set-packing relaxation and the multiple-cutting-planes-at-a-time combined with the MILP formulation enable us to optimize the modularity density for famous test instances including ones with over 100 vertices in around four minutes by a PC. Our solution method is deterministic and the computation time is not affected by any stochastic behavior. For one of them, column generation at the root node of the branch-and-bound tree provides a fractional upper bound solution and our algorithm finds an integral optimal solution after branching.
[ 1, 0, 1, 0, 0, 0 ]
[ "Computer Science", "Mathematics" ]
Title: Survey of Gravitationally-lensed Objects in HSC Imaging (SuGOHI). I. Automatic search for galaxy-scale strong lenses, Abstract: The Hyper Suprime-Cam Subaru Strategic Program (HSC SSP) is an excellent survey for the search for strong lenses, thanks to its area, image quality and depth. We use three different methods to look for lenses among 43,000 luminous red galaxies from the Baryon Oscillation Spectroscopic Survey (BOSS) sample with photometry from the S16A internal data release of the HSC SSP. The first method is a newly developed algorithm, named YATTALENS, which looks for arc-like features around massive galaxies and then estimates the likelihood of an object being a lens by performing a lens model fit. The second method, CHITAH, is a modeling-based algorithm originally developed to look for lensed quasars. The third method makes use of spectroscopic data to look for emission lines from objects at a different redshift from that of the main galaxy. We find 15 definite lenses, 36 highly probable lenses and 282 possible lenses. Among the three methods, YATTALENS, which was developed specifically for this problem, performs best in terms of both completeness and purity. Nevertheless five highly probable lenses were missed by YATTALENS but found by the other two methods, indicating that the three methods are highly complementary. Based on these numbers we expect to find $\sim$300 definite or probable lenses by the end of the HSC SSP.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics", "Astrophysics" ]
Title: Learning Sparse Neural Networks through $L_0$ Regularization, Abstract: We propose a practical method for $L_0$ norm regularization for neural networks: pruning the network during training by encouraging weights to become exactly zero. Such regularization is interesting since (1) it can greatly speed up training and inference, and (2) it can improve generalization. AIC and BIC, well-known model selection criteria, are special cases of $L_0$ regularization. However, since the $L_0$ norm of weights is non-differentiable, we cannot incorporate it directly as a regularization term in the objective function. We propose a solution through the inclusion of a collection of non-negative stochastic gates, which collectively determine which weights to set to zero. We show that, somewhat surprisingly, for certain distributions over the gates, the expected $L_0$ norm of the resulting gated weights is differentiable with respect to the distribution parameters. We further propose the \emph{hard concrete} distribution for the gates, which is obtained by "stretching" a binary concrete distribution and then transforming its samples with a hard-sigmoid. The parameters of the distribution over the gates can then be jointly optimized with the original network parameters. As a result our method allows for straightforward and efficient learning of model structures with stochastic gradient descent and allows for conditional computation in a principled way. We perform various experiments to demonstrate the effectiveness of the resulting approach and regularizer.
[ 1, 0, 0, 1, 0, 0 ]
[ "Computer Science", "Statistics" ]
Title: Sequential Randomized Matrix Factorization for Gaussian Processes: Efficient Predictions and Hyper-parameter Optimization, Abstract: This paper presents a sequential randomized lowrank matrix factorization approach for incrementally predicting values of an unknown function at test points using the Gaussian Processes framework. It is well-known that in the Gaussian processes framework, the computational bottlenecks are the inversion of the (regularized) kernel matrix and the computation of the hyper-parameters defining the kernel. The main contributions of this paper are two-fold. First, we formalize an approach to compute the inverse of the kernel matrix using randomized matrix factorization algorithms in a streaming scenario, i.e., data is generated incrementally over time. The metrics of accuracy and computational efficiency of the proposed method are compared against a batch approach based on use of randomized matrix factorization and an existing streaming approach based on approximating the Gaussian process by a finite set of basis vectors. Second, we extend the sequential factorization approach to a class of kernel functions for which the hyperparameters can be efficiently optimized. All results are demonstrated on two publicly available datasets.
[ 1, 0, 0, 1, 0, 0 ]
[ "Computer Science", "Statistics" ]
Title: Large-scale dynamos in rapidly rotating plane layer convection, Abstract: Context: Convectively-driven flows play a crucial role in the dynamo processes that are responsible for producing magnetic activity in stars and planets. It is still not fully understood why many astrophysical magnetic fields have a significant large-scale component. Aims: Our aim is to investigate the dynamo properties of compressible convection in a rapidly rotating Cartesian domain, focusing upon a parameter regime in which the underlying hydrodynamic flow is known to be unstable to a large-scale vortex instability. Methods: The governing equations of three-dimensional nonlinear magnetohydrodynamics (MHD) are solved numerically. Different numerical schemes are compared and we propose a possible benchmark case for other similar codes. Results: In keeping with previous related studies, we find that convection in this parameter regime can drive a large-scale dynamo. The components of the mean horizontal magnetic field oscillate, leading to a continuous overall rotation of the mean field. Whilst the large-scale vortex instability dominates the early evolution of the system, it is suppressed by the magnetic field and makes a negligible contribution to the mean electromotive force that is responsible for driving the large-scale dynamo. The cycle period of the dynamo is comparable to the ohmic decay time, with longer cycles for dynamos in convective systems that are closer to onset. In these particular simulations, large-scale dynamo action is found only when vertical magnetic field boundary conditions are adopted at the upper and lower boundaries. Strongly modulated large-scale dynamos are found at higher Rayleigh numbers, with periods of reduced activity ("grand minima"-like events) occurring during transient phases in which the large-scale vortex temporarily re-establishes itself, before being suppressed again by the magnetic field.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: Statistical inference for misspecified ergodic Lévy driven stochastic differential equation models, Abstract: This paper deals with the estimation problem of misspecified ergodic Lévy driven stochastic differential equation models based on high-frequency samples. We utilize the widely applicable and tractable Gaussian quasi-likelihood approach which focuses on (conditional) mean and variance structure. It is shown that the corresponding Gaussian quasi-likelihood estimators of drift and scale parameters satisfy tail probability estimates and asymptotic normality at the same rate as correctly specified case. In this process, extended Poisson equation for time-homogeneous Feller Markov processes plays an important role to handle misspecification effect. Our result confirms the practical usefulness of the Gaussian quasi-likelihood approach for SDE models, more firmly.
[ 0, 0, 1, 1, 0, 0 ]
[ "Statistics", "Mathematics" ]
Title: Automated and Robust Quantification of Colocalization in Dual-Color Fluorescence Microscopy: A Nonparametric Statistical Approach, Abstract: Colocalization is a powerful tool to study the interactions between fluorescently labeled molecules in biological fluorescence microscopy. However, existing techniques for colocalization analysis have not undergone continued development especially in regards to robust statistical support. In this paper, we examine two of the most popular quantification techniques for colocalization and argue that they could be improved upon using ideas from nonparametric statistics and scan statistics. In particular, we propose a new colocalization metric that is robust, easily implementable, and optimal in a rigorous statistical testing framework. Application to several benchmark datasets, as well as biological examples, further demonstrates the usefulness of the proposed technique.
[ 0, 0, 0, 1, 0, 0 ]
[ "Quantitative Biology", "Statistics" ]
Title: Subadditivity and additivity of the Yang-Mills action functional in Noncommutative Geometry, Abstract: We formulate notions of subadditivity and additivity of the Yang-Mills action functional in noncommutative geometry. We identify a suitable hypothesis on spectral triples which proves that the Yang-Mills functional is always subadditive, as per expectation. The additivity property is much stronger in the sense that it implies the subadditivity property. Under this hypothesis we obtain a necessary and sufficient condition for the additivity of the Yang-Mills functional. An instance of additivity is shown for the case of noncommutative $n$-tori. We also investigate the behaviour of critical points of the Yang-Mills functional under additivity. At the end we discuss few examples involving compact spin manifolds, matrix algebras, noncommutative $n$-torus and the quantum Heisenberg manifolds which validate our hypothesis.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics", "Physics" ]
Title: Specifying a positive threshold function via extremal points, Abstract: An extremal point of a positive threshold Boolean function $f$ is either a maximal zero or a minimal one. It is known that if $f$ depends on all its variables, then the set of its extremal points completely specifies $f$ within the universe of threshold functions. However, in some cases, $f$ can be specified by a smaller set. The minimum number of points in such a set is the specification number of $f$. It was shown in [S.-T. Hu. Threshold Logic, 1965] that the specification number of a threshold function of $n$ variables is at least $n+1$. In [M. Anthony, G. Brightwell, and J. Shawe-Taylor. On specifying Boolean functions by labelled examples. Discrete Applied Mathematics, 1995] it was proved that this bound is attained for nested functions and conjectured that for all other threshold functions the specification number is strictly greater than $n+1$. In the present paper, we resolve this conjecture negatively by exhibiting threshold Boolean functions of $n$ variables, which are non-nested and for which the specification number is $n+1$. On the other hand, we show that the set of extremal points satisfies the statement of the conjecture, i.e., a positive threshold Boolean function depending on all its $n$ variables has $n+1$ extremal points if and only if it is nested. To prove this, we reveal an underlying structure of the set of extremal points.
[ 1, 0, 0, 0, 0, 0 ]
[ "Mathematics", "Computer Science" ]
Title: Using Stock Prices as Ground Truth in Sentiment Analysis to Generate Profitable Trading Signals, Abstract: The increasing availability of "big" (large volume) social media data has motivated a great deal of research in applying sentiment analysis to predict the movement of prices within financial markets. Previous work in this field investigates how the true sentiment of text (i.e. positive or negative opinions) can be used for financial predictions, based on the assumption that sentiments expressed online are representative of the true market sentiment. Here we consider the converse idea, that using the stock price as the ground-truth in the system may be a better indication of sentiment. Tweets are labelled as Buy or Sell dependent on whether the stock price discussed rose or fell over the following hour, and from this, stock-specific dictionaries are built for individual companies. A Bayesian classifier is used to generate stock predictions, which are input to an automated trading algorithm. Placing 468 trades over a 1 month period yields a return rate of 5.18%, which annualises to approximately 83% per annum. This approach performs significantly better than random chance and outperforms two baseline sentiment analysis methods tested.
[ 0, 0, 0, 0, 0, 1 ]
[ "Computer Science", "Quantitative Finance", "Statistics" ]
Title: When does every definable nonempty set have a definable element?, Abstract: The assertion that every definable set has a definable element is equivalent over ZF to the principle $V=\text{HOD}$, and indeed, we prove, so is the assertion merely that every $\Pi_2$-definable set has an ordinal-definable element. Meanwhile, every model of ZFC has a forcing extension satisfying $V\neq\text{HOD}$ in which every $\Sigma_2$-definable set has an ordinal-definable element. Similar results hold for $\text{HOD}(\mathbb{R})$ and $\text{HOD}(\text{Ord}^\omega)$ and other natural instances of $\text{HOD}(X)$.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics" ]
Title: Measurement and Analysis of Quality of Service of Mobile Networks in Afghanistan End User Perspective, Abstract: Enhanced Quality of Service (QoS) and satisfaction of mobile phone user are major concerns of a service provider. In order to manage network efficiently and to provide enhanced end to end Quality of Experience (QoE), operator is expected to measure and analyze QoS from various perspectives and at different relevant points of network. The scope of this paper is measurement and statistically analysis of QoS of mobile networks from end user perspective in Afghanistan. The study is based on primary data collected on random basis from 1,515 mobile phone users of five cellular operators. The paper furthermore proposes adequate technical solutions to mobile operators in order to address existing challenges in the area of QoS and to remain competitive in the market. Based on the result of processed data, considering geographical locations, population and telecom regulations of the government, authors recommend deployment of small cells (SCs), increasing number of regular performance tests, optimal placement of base stations, increasing number of carriers, and high order sectorization as proposed technical solutions.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Statistics" ]
Title: On the uncertainty of temperature estimation in a rapid compression machine, Abstract: Rapid compression machines (RCMs) have been widely used in the combustion literature to study the low-to-intermediate temperature ignition of many fuels. In a typical RCM, the pressure during and after the compression stroke is measured. However, measurement of the temperature history in the RCM reaction chamber is challenging. Thus, the temperature is generally calculated by the isentropic relations between pressure and temperature, assuming that the adiabatic core hypothesis holds. To estimate the uncertainty in the calculated temperature, an uncertainty propagation analysis must be carried out. Our previous analyses assumed that the uncertainties of the parameters in the equation to calculate the temperature were normally distributed and independent, but these assumptions do not hold for typical RCM operating procedures. In this work, a Monte Carlo method is developed to estimate the uncertainty in the calculated temperature, while taking into account the correlation between parameters and the possibility of non-normal probability distributions. In addition, the Monte Carlo method is compared to an analysis that assumes normally distributed, independent parameters. Both analysis methods show that the magnitude of the initial pressure and the uncertainty of the initial temperature have strong influences on the magnitude of the uncertainty. Finally, the uncertainty estimation methods studied here provide a reference value for the uncertainty of the reference temperature in an RCM and can be generalized to other similar facilities.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics", "Statistics" ]
Title: Improved Training of Wasserstein GANs, Abstract: Generative Adversarial Networks (GANs) are powerful generative models, but suffer from training instability. The recently proposed Wasserstein GAN (WGAN) makes progress toward stable training of GANs, but sometimes can still generate only low-quality samples or fail to converge. We find that these problems are often due to the use of weight clipping in WGAN to enforce a Lipschitz constraint on the critic, which can lead to undesired behavior. We propose an alternative to clipping weights: penalize the norm of gradient of the critic with respect to its input. Our proposed method performs better than standard WGAN and enables stable training of a wide variety of GAN architectures with almost no hyperparameter tuning, including 101-layer ResNets and language models over discrete data. We also achieve high quality generations on CIFAR-10 and LSUN bedrooms.
[ 1, 0, 0, 1, 0, 0 ]
[ "Computer Science", "Statistics" ]
Title: A cancellation theorem for Milnor-Witt correspondences, Abstract: We show that finite Milnor-Witt correspondences satisfy a cancellation theorem with respect to the pointed multiplicative group scheme. This has several notable applications in the theory of Milnor-Witt motives and Milnor-Witt motivic cohomology.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics" ]
Title: Weak Keys and Cryptanalysis of a Cold War Block Cipher, Abstract: T-310 is a cipher that was used for encryption of governmental communications in East Germany during the final years of the Cold War. Due to its complexity and the encryption process,there was no published attack for a period of more than 40 years until 2018 by Nicolas T. Courtois et al. in [10]. In this thesis we study the so called 'long term keys' that were used in the cipher, in order to expose weaknesses which will assist the design of various attacks on T-310.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science" ]
Title: Spectroscopy of Ultra-diffuse Galaxies in the Coma Cluster, Abstract: We present spectra of 5 ultra-diffuse galaxies (UDGs) in the vicinity of the Coma Cluster obtained with the Multi-Object Double Spectrograph on the Large Binocular Telescope. We confirm 4 of these as members of the cluster, quintupling the number of spectroscopically confirmed systems. Like the previously confirmed large (projected half light radius $>$ 4.6 kpc) UDG, DF44, the systems we targeted all have projected half light radii $> 2.9$ kpc. As such, we spectroscopically confirm a population of physically large UDGs in the Coma cluster. The remaining UDG is located in the field, about $45$ Mpc behind the cluster. We observe Balmer and Ca II H \& K absorption lines in all of our UDG spectra. By comparing the stacked UDG spectrum against stellar population synthesis models, we conclude that, on average, these UDGs are composed of metal-poor stars ([Fe/H] $\lesssim -1.5$). We also discover the first UDG with [OII] and [OIII] emission lines within a clustered environment, demonstrating that not all cluster UDGs are devoid of gas and sources of ionizing radiation.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: On Robust Tie-line Scheduling in Multi-Area Power Systems, Abstract: The tie-line scheduling problem in a multi-area power system seeks to optimize tie-line power flows across areas that are independently operated by different system operators (SOs). In this paper, we leverage the theory of multi-parametric linear programming to propose algorithms for optimal tie-line scheduling within a deterministic and a robust optimization framework. Through a coordinator, the proposed algorithms are proved to converge to the optimal schedule within a finite number of iterations. A key feature of the proposed algorithms, besides their finite step convergence, is the privacy of the information exchanges; the SO in an area does not need to reveal its dispatch cost structure, network constraints, or the nature of the uncertainty set to the coordinator. The performance of the algorithms is evaluated using several power system examples.
[ 0, 0, 1, 0, 0, 0 ]
[ "Computer Science", "Mathematics" ]
Title: Stochastic Input Models in Online Computing, Abstract: In this paper, we study twelve stochastic input models for online problems and reveal the relationships among the competitive ratios for the models. The competitive ratio is defined as the worst ratio between the expected optimal value and the expected profit of the solution obtained by the online algorithm where the input distribution is restricted according to the model. To handle a broad class of online problems, we use a framework called request-answer games that is introduced by Ben-David et al. The stochastic input models consist of two types: known distribution and unknown distribution. For each type, we consider six classes of distributions: dependent distributions, deterministic input, independent distributions, identical independent distribution, random order of a deterministic input, and random order of independent distributions. As an application of the models, we consider two basic online problems, which are variants of the secretary problem and the prophet inequality problem, under the twelve stochastic input models. We see the difference of the competitive ratios through these problems.
[ 1, 0, 1, 1, 0, 0 ]
[ "Computer Science", "Mathematics", "Statistics" ]
Title: Generating Visual Representations for Zero-Shot Classification, Abstract: This paper addresses the task of learning an image clas-sifier when some categories are defined by semantic descriptions only (e.g. visual attributes) while the others are defined by exemplar images as well. This task is often referred to as the Zero-Shot classification task (ZSC). Most of the previous methods rely on learning a common embedding space allowing to compare visual features of unknown categories with semantic descriptions. This paper argues that these approaches are limited as i) efficient discrimi-native classifiers can't be used ii) classification tasks with seen and unseen categories (Generalized Zero-Shot Classification or GZSC) can't be addressed efficiently. In contrast , this paper suggests to address ZSC and GZSC by i) learning a conditional generator using seen classes ii) generate artificial training examples for the categories without exemplars. ZSC is then turned into a standard supervised learning problem. Experiments with 4 generative models and 5 datasets experimentally validate the approach, giving state-of-the-art results on both ZSC and GZSC.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Statistics" ]
Title: Mitigating radiation damage of single photon detectors for space applications, Abstract: Single-photon detectors in space must retain useful performance characteristics despite being bombarded with sub-atomic particles. Mitigating the effects of this space radiation is vital to enabling new space applications which require high-fidelity single-photon detection. To this end, we conducted proton radiation tests of various models of avalanche photodiodes (APDs) and one model of photomultiplier tube potentially suitable for satellite-based quantum communications. The samples were irradiated with 106 MeV protons at doses approximately equivalent to lifetimes of 0.6 , 6, 12 and 24 months in a low-Earth polar orbit. Although most detection properties were preserved, including efficiency, timing jitter and afterpulsing probability, all APD samples demonstrated significant increases in dark count rate (DCR) due to radiation-induced damage, many orders of magnitude higher than the 200 counts per second (cps) required for ground-to-satellite quantum communications. We then successfully demonstrated the mitigation of this DCR degradation through the use of deep cooling, to as low as -86 degrees C. This achieved DCR below the required 200 cps over the 24 months orbit duration. DCR was further reduced by thermal annealing at temperatures of +50 to +100 degrees C.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: An exact solution to a Stefan problem with variable thermal conductivity and a Robin boundary condition, Abstract: In this article it is proved the existence of similarity solutions for a one-phase Stefan problem with temperature-dependent thermal conductivity and a Robin condition at the fixed face. The temperature distribution is obtained through a generalized modified error function which is defined as the solution to a nonlinear ordinary differential problem of second order. It is proved that the latter has a unique non-negative bounded analytic solution when the parameter on which it depends assumes small positive values. Moreover, it is shown that the generalized modified error function is concave and increasing, and explicit approximations are proposed for it. Relation between the Stefan problem considered in this article with those with either constant thermal conductivity or a temperature boundary condition is also analysed.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics", "Physics" ]
Title: State-of-the-art Speech Recognition With Sequence-to-Sequence Models, Abstract: Attention-based encoder-decoder architectures such as Listen, Attend, and Spell (LAS), subsume the acoustic, pronunciation and language model components of a traditional automatic speech recognition (ASR) system into a single neural network. In previous work, we have shown that such architectures are comparable to state-of-theart ASR systems on dictation tasks, but it was not clear if such architectures would be practical for more challenging tasks such as voice search. In this work, we explore a variety of structural and optimization improvements to our LAS model which significantly improve performance. On the structural side, we show that word piece models can be used instead of graphemes. We also introduce a multi-head attention architecture, which offers improvements over the commonly-used single-head attention. On the optimization side, we explore synchronous training, scheduled sampling, label smoothing, and minimum word error rate optimization, which are all shown to improve accuracy. We present results with a unidirectional LSTM encoder for streaming recognition. On a 12, 500 hour voice search task, we find that the proposed changes improve the WER from 9.2% to 5.6%, while the best conventional system achieves 6.7%; on a dictation task our model achieves a WER of 4.1% compared to 5% for the conventional system.
[ 1, 0, 0, 1, 0, 0 ]
[ "Computer Science" ]
Title: Bayesian inference for Stable Levy driven Stochastic Differential Equations with high-frequency data, Abstract: In this article we consider parametric Bayesian inference for stochastic differential equations (SDE) driven by a pure-jump stable Levy process, which is observed at high frequency. In most cases of practical interest, the likelihood function is not available, so we use a quasi-likelihood and place an associated prior on the unknown parameters. It is shown under regularity conditions that there is a Bernstein-von Mises theorem associated to the posterior. We then develop a Markov chain Monte Carlo (MCMC) algorithm for Bayesian inference and assisted by our theoretical results, we show how to scale Metropolis-Hastings proposals when the frequency of the data grows, in order to prevent the acceptance ratio going to zero in the large data limit. Our algorithm is presented on numerical examples that help to verify our theoretical findings.
[ 0, 0, 1, 1, 0, 0 ]
[ "Statistics", "Mathematics" ]
Title: The Evolution of Reputation-Based Cooperation in Regular Networks, Abstract: Despite recent advances in reputation technologies, it is not clear how reputation systems can affect human cooperation in social networks. Although it is known that two of the major mechanisms in the evolution of cooperation are spatial selection and reputation-based reciprocity, theoretical study of the interplay between both mechanisms remains almost uncharted. Here, we present a new individual-based model for the evolution of reciprocal cooperation between reputation and networks. We comparatively analyze four of the leading moral assessment rules---shunning, image scoring, stern judging, and simple standing---and base the model on the giving game in regular networks for Cooperators, Defectors, and Discriminators. Discriminators rely on a proper moral assessment rule. By using individual-based models, we show that the four assessment rules are differently characterized in terms of how cooperation evolves, depending on the benefit-to-cost ratio, the network-node degree, and the observation and error conditions. Our findings show that the most tolerant rule---simple standing---is the most robust among the four assessment rules in promoting cooperation in regular networks.
[ 1, 1, 0, 0, 0, 0 ]
[ "Quantitative Biology", "Mathematics" ]
Title: The ABCD of topological recursion, Abstract: Kontsevich and Soibelman reformulated and slightly generalised the topological recursion of math-ph/0702045, seeing it as a quantization of certain quadratic Lagrangians in $T^*V$ for some vector space $V$. KS topological recursion is a procedure which takes as initial data a quantum Airy structure -- a family of at most quadratic differential operators on $V$ satisfying some axioms -- and gives as outcome a formal series of functions in $V$ (the partition function) simultaneously annihilated by these operators. Finding and classifying quantum Airy structures modulo gauge group action, is by itself an interesting problem which we study here. We provide some elementary, Lie-algebraic tools to address this problem, and give some elements of classification for ${\rm dim}\,V = 2$. We also describe four more interesting classes of quantum Airy structures, coming from respectively Frobenius algebras (here we retrieve the 2d TQFT partition function as a special case), non-commutative Frobenius algebras, loop spaces of Frobenius algebras and a $\mathbb{Z}_{2}$-invariant version of the latter. This $\mathbb{Z}_{2}$-invariant version in the case of a semi-simple Frobenius algebra corresponds to the topological recursion of math-ph/0702045.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics", "Physics" ]
Title: Evaluating Compositionality in Sentence Embeddings, Abstract: An important challenge for human-like AI is compositional semantics. Recent research has attempted to address this by using deep neural networks to learn vector space embeddings of sentences, which then serve as input to other tasks. We present a new dataset for one such task, `natural language inference' (NLI), that cannot be solved using only word-level knowledge and requires some compositionality. We find that the performance of state of the art sentence embeddings (InferSent; Conneau et al., 2017) on our new dataset is poor. We analyze the decision rules learned by InferSent and find that they are consistent with simple heuristics that are ecologically valid in its training dataset. Further, we find that augmenting training with our dataset improves test performance on our dataset without loss of performance on the original training dataset. This highlights the importance of structured datasets in better understanding and improving AI systems.
[ 0, 0, 0, 1, 0, 0 ]
[ "Computer Science", "Statistics" ]
Title: Half-range lattice Boltzmann models for the simulation of Couette flow using the Shakhov collision term, Abstract: The three-dimensional Couette flow between parallel plates is addressed using mixed lattice Boltzmann models which implement the half-range and the full-range Gauss-Hermite quadratures on the Cartesian axes perpendicular and parallel to the walls, respectively. The ability of our models to simulate rarefied flows are validated through comparison against previously reported results obtained using the linearized Boltzmann-BGK equation for values of the Knudsen number (Kn) up to $100$. We find that recovering the non-linear part of the velocity profile (i.e., its deviation from a linear function) at ${\rm Kn} \gtrsim 1$ requires high quadrature orders. We then employ the Shakhov model for the collision term to obtain macroscopic profiles for Maxwell molecules using the standard $\mu \sim T^\omega$ law, as well as for monatomic Helium and Argon gases, modeled through ab-initio potentials, where the viscosity is recovered using the Sutherland model. We validate our implementation by comparison with DSMC results and find excellent match for all macroscopic quantities for ${\rm Kn} \lesssim 0.1$. At ${\rm Kn} \gtrsim 0.1$, small deviations can be seen in the profiles of the diagonal components of the pressure tensor, the heat flux parallel to the plates, and the velocity profile, as well as in the values of the velocity gradient at the channel center. We attribute these deviations to the limited applicability of the Shakhov collision model for highly out of equilibrium flows.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics", "Mathematics" ]
Title: Analyzing and improving maximal attainable accuracy in the communication hiding pipelined BiCGStab method, Abstract: Pipelined Krylov subspace methods avoid communication latency by reducing the number of global synchronization bottlenecks and by hiding global communication behind useful computational work. In exact arithmetic pipelined Krylov subspace algorithms are equivalent to classic Krylov subspace methods and generate identical series of iterates. However, as a consequence of the reformulation of the algorithm to improve parallelism, pipelined methods may suffer from severely reduced attainable accuracy in a practical finite precision setting. This work presents a numerical stability analysis that describes and quantifies the impact of local rounding error propagation on the maximal attainable accuracy of the multi-term recurrences in the preconditioned pipelined BiCGStab method. Theoretical expressions for the gaps between the true and computed residual as well as other auxiliary variables used in the algorithm are derived, and the elementary dependencies between the gaps on the various recursively computed vector variables are analyzed. The norms of the corresponding propagation matrices and vectors provide insights in the possible amplification of local rounding errors throughout the algorithm. Stability of the pipelined BiCGStab method is compared numerically to that of pipelined CG on a symmetric benchmark problem. Furthermore, numerical evidence supporting the effectiveness of employing a residual replacement type strategy to improve the maximal attainable accuracy for the pipelined BiCGStab method is provided.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Mathematics" ]
Title: Trigonometric integrators for quasilinear wave equations, Abstract: Trigonometric time integrators are introduced as a class of explicit numerical methods for quasilinear wave equations. Second-order convergence for the semi-discretization in time with these integrators is shown for a sufficiently regular exact solution. The time integrators are also combined with a Fourier spectral method into a fully discrete scheme, for which error bounds are provided without requiring any CFL-type coupling of the discretization parameters. The proofs of the error bounds are based on energy techniques and on the semiclassical G\aa rding inequality.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics", "Physics", "Computer Science" ]
Title: A Highly Efficient Polarization-Independent Metamaterial-Based RF Energy-Harvesting Rectenna for Low-Power Applications, Abstract: A highly-efficient multi-resonant RF energy-harvesting rectenna based on a metamaterial perfect absorber featuring closely-spaced polarization-independent absorption modes is presented. Its effective area is larger than its physical area, and so efficiencies of 230% and 130% are measured at power densities of 10 uW/cm2 and 1 uW/cm2 respectively, for a linear absorption mode at 0.75 GHz. The rectenna exhibits a broad polarization-independent region between 1.4 GHz and 1.7 GHz with maximum efficiencies of 167% and 36% for those same power densities. Additionally, by adjustment of the distance between the rectenna and a reflecting ground plane, the absorption frequency can be adjusted to a limited extent within the polarization-independent region. Lastly, the rectenna should be capable of delivering 100 uW of power to a device located within 50 m of a cell-phone tower under ideal conditions.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: Mixed Graphical Models for Causal Analysis of Multi-modal Variables, Abstract: Graphical causal models are an important tool for knowledge discovery because they can represent both the causal relations between variables and the multivariate probability distributions over the data. Once learned, causal graphs can be used for classification, feature selection and hypothesis generation, while revealing the underlying causal network structure and thus allowing for arbitrary likelihood queries over the data. However, current algorithms for learning sparse directed graphs are generally designed to handle only one type of data (continuous-only or discrete-only), which limits their applicability to a large class of multi-modal biological datasets that include mixed type variables. To address this issue, we developed new methods that modify and combine existing methods for finding undirected graphs with methods for finding directed graphs. These hybrid methods are not only faster, but also perform better than the directed graph estimation methods alone for a variety of parameter settings and data set sizes. Here, we describe a new conditional independence test for learning directed graphs over mixed data types and we compare performances of different graph learning strategies on synthetic data.
[ 1, 0, 0, 1, 0, 0 ]
[ "Computer Science", "Statistics", "Quantitative Biology" ]
Title: Deep Robust Framework for Protein Function Prediction using Variable-Length Protein Sequences, Abstract: Amino acid sequence portrays most intrinsic form of a protein and expresses primary structure of protein. The order of amino acids in a sequence enables a protein to acquire a particular stable conformation that is responsible for the functions of the protein. This relationship between a sequence and its function motivates the need to analyse the sequences for predicting protein functions. Early generation computational methods using BLAST, FASTA, etc. perform function transfer based on sequence similarity with existing databases and are computationally slow. Although machine learning based approaches are fast, they fail to perform well for long protein sequences (i.e., protein sequences with more than 300 amino acid residues). In this paper, we introduce a novel method for construction of two separate feature sets for protein sequences based on analysis of 1) single fixed-sized segments and 2) multi-sized segments, using bi-directional long short-term memory network. Further, model based on proposed feature set is combined with the state of the art Multi-lable Linear Discriminant Analysis (MLDA) features based model to improve the accuracy. Extensive evaluations using separate datasets for biological processes and molecular functions demonstrate promising results for both single-sized and multi-sized segments based feature sets. While former showed an improvement of +3.37% and +5.48%, the latter produces an improvement of +5.38% and +8.00% respectively for two datasets over the state of the art MLDA based classifier. After combining two models, there is a significant improvement of +7.41% and +9.21% respectively for two datasets compared to MLDA based classifier. Specifically, the proposed approach performed well for the long protein sequences and superior overall performance.
[ 0, 0, 0, 0, 1, 0 ]
[ "Computer Science", "Quantitative Biology" ]
Title: Helicity of convective flows from localized heat source in a rotating layer, Abstract: Experimental and numerical study of the steady-state cyclonic vortex from isolated heat source in a rotating fluid layer is described. The structure of laboratory cyclonic vortex is similar to the typical structure of tropical cyclones from observational data and numerical modelling including secondary flows in the boundary layer. Differential characteristics of the flow were studied by numerical simulation using CFD software FlowVision. Helicity distribution in rotating fluid layer with localized heat source was analysed. Two mechanisms which play role in helicity generation are found. The first one is the strong correlation of cyclonic vortex and intensive upward motion in the central part of the vessel. The second one is due to large gradients of velocity on the periphery. The integral helicity in the considered case is substantial and its relative level is high.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: Fine-Gray competing risks model with high-dimensional covariates: estimation and Inference, Abstract: The purpose of this paper is to construct confidence intervals for the regression coefficients in the Fine-Gray model for competing risks data with random censoring, where the number of covariates can be larger than the sample size. Despite strong motivation from biostatistics applications, high-dimensional Fine-Gray model has attracted relatively little attention among the methodological or theoretical literatures. We fill in this blank by proposing first a consistent regularized estimator and then the confidence intervals based on the one-step bias-correcting estimator. We are able to generalize the partial likelihood approach for the Fine-Gray model under random censoring despite many technical difficulties. We lay down a methodological and theoretical framework for the one-step bias-correcting estimator with the partial likelihood, which does not have independent and identically distributed entries. We also handle for our theory the approximation error from the inverse probability weighting (IPW), proposing novel concentration results for time dependent processes. In addition to the theoretical results and algorithms, we present extensive numerical experiments and an application to a study of non-cancer mortality among prostate cancer patients using the linked Medicare-SEER data.
[ 0, 0, 1, 1, 0, 0 ]
[ "Statistics", "Mathematics" ]
Title: Cartan's Conjecture for Moving Hypersurfaces, Abstract: Let $f$ be a holomorphic curve in $\mathbb{P}^n({\mathbb{C}})$ and let $\mathcal{D}=\{D_1,\ldots,D_q\}$ be a family of moving hypersurfaces defined by a set of homogeneous polynomials $\mathcal{Q}=\{Q_1,\ldots,Q_q\}$. For $j=1,\ldots,q$, denote by $Q_j=\sum\limits_{i_0+\cdots+i_n=d_j}a_{j,I}(z)x_0^{i_0}\cdots x_n^{i_n}$, where $I=(i_0,\ldots,i_n)\in\mathbb{Z}_{\ge 0}^{n+1}$ and $a_{j,I}(z)$ are entire functions on ${\mathbb{C}}$ without common zeros. Let $\mathcal{K}_{\mathcal{Q}}$ be the smallest subfield of meromorphic function field $\mathcal{M}$ which contains ${\mathbb{C}}$ and all $\frac{a_{j,I'}(z)}{a_{j,I''}(z)}$ with $a_{j,I''}(z)\not\equiv 0$, $1\le j\le q$. In previous known second main theorems for $f$ and $\mathcal{D}$, $f$ is usually assumed to be algebraically nondegenerate over $\mathcal{K}_{\mathcal{Q}}$. In this paper, we prove a second main theorem in which $f$ is only assumed to be nonconstant. This result can be regarded as a generalization of Cartan's conjecture for moving hypersurfaces.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics" ]
Title: A Fluid-Flow Interpretation of SCED Scheduling, Abstract: We show that a fluid-flow interpretation of Service Curve Earliest Deadline First (SCED) scheduling simplifies deadline derivations for this scheduler. By exploiting the recently reported isomorphism between min-plus and max-plus network calculus, and expressing deadlines in a max-plus algebra, deadline computations no longer require pseudo-inverse computations. SCED deadlines are provided for general convex or concave piecewise linear service curves.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Mathematics" ]
Title: Emergence of Topological Nodal Lines and Type II Weyl Nodes in Strong Spin--Orbit Coupling System InNbX2(X=S,Se), Abstract: Using first--principles density functional calculations, we systematically investigate electronic structures and topological properties of InNbX2 (X=S, Se). In the absence of spin--orbit coupling (SOC), both compounds show nodal lines protected by mirror symmetry. Including SOC, the Dirac rings in InNbS2 split into two Weyl rings. This unique property is distinguished from other dicovered nodal line materials which normally requires the absence of SOC. On the other hand, SOC breaks the nodal lines in InNbSe2 and the compound becomes a type II Weyl semimetal with 12 Weyl points in the Brillouin Zone. Using a supercell slab calculation we study the dispersion of Fermi arcs surface states in InNbSe2, we also utilize a coherent potential approximation to probe their tolernace to the surface disorder effects. The quasi two--dimensionality and the absence of toxic elements makes these two compounds an ideal experimental platform for investigating novel properties of topological semimetals.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: Scheduling with regular performance measures and optional job rejection on a single machine, Abstract: We address single machine problems with optional jobs - rejection, studied recently in Zhang et al. [21] and Cao et al. [2]. In these papers, the authors focus on minimizing regular performance measures, i.e., functions that are non-decreasing in the jobs completion time, subject to the constraint that the total rejection cost cannot exceed a predefined upper bound. They also prove that the considered problems are ordinary NP-hard and provide pseudo-polynomial-time Dynamic Programming (DP) solutions. In this paper, we focus on three of these problems: makespan with release-dates; total completion times; and total weighted completion, and present enhanced DP solutions demonstrating both theoretical and practical improvements. Moreover, we provide extensive numerical studies verifying their efficiency.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Mathematics" ]
Title: Data-Driven Stochastic Robust Optimization: A General Computational Framework and Algorithm for Optimization under Uncertainty in the Big Data Era, Abstract: A novel data-driven stochastic robust optimization (DDSRO) framework is proposed for optimization under uncertainty leveraging labeled multi-class uncertainty data. Uncertainty data in large datasets are often collected from various conditions, which are encoded by class labels. Machine learning methods including Dirichlet process mixture model and maximum likelihood estimation are employed for uncertainty modeling. A DDSRO framework is further proposed based on the data-driven uncertainty model through a bi-level optimization structure. The outer optimization problem follows a two-stage stochastic programming approach to optimize the expected objective across different data classes; adaptive robust optimization is nested as the inner problem to ensure the robustness of the solution while maintaining computational tractability. A decomposition-based algorithm is further developed to solve the resulting multi-level optimization problem efficiently. Case studies on process network design and planning are presented to demonstrate the applicability of the proposed framework and algorithm.
[ 1, 0, 1, 0, 0, 0 ]
[ "Computer Science", "Mathematics", "Statistics" ]
Title: Algebraic multiscale method for flow in heterogeneous porous media with embedded discrete fractures (F-AMS), Abstract: This paper introduces an Algebraic MultiScale method for simulation of flow in heterogeneous porous media with embedded discrete Fractures (F-AMS). First, multiscale coarse grids are independently constructed for both porous matrix and fracture networks. Then, a map between coarse- and fine-scale is obtained by algebraically computing basis functions with local support. In order to extend the localization assumption to the fractured media, four types of basis functions are investigated: (1) Decoupled-AMS, in which the two media are completely decoupled, (2) Frac-AMS and (3) Rock-AMS, which take into account only one-way transmissibilities, and (4) Coupled-AMS, in which the matrix and fracture interpolators are fully coupled. In order to ensure scalability, the F-AMS framework permits full flexibility in terms of the resolution of the fracture coarse grids. Numerical results are presented for two- and three-dimensional heterogeneous test cases. During these experiments, the performance of F-AMS, paired with ILU(0) as second-stage smoother in a convergent iterative procedure, is studied by monitoring CPU times and convergence rates. Finally, in order to investigate the scalability of the method, an extensive benchmark study is conducted, where a commercial algebraic multigrid solver is used as reference. The results show that, given an appropriate coarsening strategy, F-AMS is insensitive to severe fracture and matrix conductivity contrasts, as well as the length of the fracture networks. Its unique feature is that a fine-scale mass conservative flux field can be reconstructed after any iteration, providing efficient approximate solutions in time-dependent simulations.
[ 1, 1, 0, 0, 0, 0 ]
[ "Mathematics", "Physics" ]
Title: A Bayesian Mixture Model for Clustering on the Stiefel Manifold, Abstract: Analysis of a Bayesian mixture model for the Matrix Langevin distribution on the Stiefel manifold is presented. The model exploits a particular parametrization of the Matrix Langevin distribution, various aspects of which are elaborated on. A general, and novel, family of conjugate priors, and an efficient Markov chain Monte Carlo (MCMC) sampling scheme for the corresponding posteriors is then developed for the mixture model. Theoretical properties of the prior and posterior distributions, including posterior consistency, are explored in detail. Extensive simulation experiments are presented to validate the efficacy of the framework. Real-world examples, including a large scale neuroimaging dataset, are analyzed to demonstrate the computational tractability of the approach.
[ 0, 0, 0, 1, 0, 0 ]
[ "Statistics", "Mathematics" ]
Title: Multi-color image compression-encryption algorithm based on chaotic system and fuzzy transform, Abstract: In this paper an algorithm for multi-color image compression-encryption is introduced. For compression step fuzzy transform based on exponential b-spline function is used. In encryption step, a novel combination chaotic system based on Sine and Tent systems is proposed. Also in the encryption algorithm, 3D shift based on chaotic system is introduced. The simulation results and security analysis show that the proposed algorithm is secure and efficient.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science" ]
Title: The Dynamical History of Chariklo and its Rings, Abstract: Chariklo is the only small Solar system body confirmed to have rings. Given the instability of its orbit, the presence of rings is surprising, and their origin remains poorly understood. In this work, we study the dynamical history of the Chariklo system by integrating almost 36,000 Chariklo clones backwards in time for one Gyr under the influence of the Sun and the four giant planets. By recording all close encounters between the clones and planets, we investigate the likelihood that Chariklo's rings could have survived since its capture to the Centaur population. Our results reveal that Chariklo's orbit occupies a region of stable chaos, resulting in its orbit being marginally more stable than those of the other Centaurs. Despite this, we find that it was most likely captured to the Centaur population within the last 20 Myr, and that its orbital evolution has been continually punctuated by regular close encounters with the giant planets. The great majority (> 99%) of those encounters within one Hill radius of the planet have only a small effect on the rings. We conclude that close encounters with giant planets have not had a significant effect on the ring structure. Encounters within the Roche limit of the giant planets are rare, making ring creation through tidal disruption unlikely.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: Smooth positon solutions of the focusing modified Korteweg-de Vries equation, Abstract: The $n$-fold Darboux transformation $T_{n}$ of the focusing real mo\-di\-fied Kor\-te\-weg-de Vries (mKdV) equation is expressed in terms of the determinant representation. Using this representation, the $n$-soliton solutions of the mKdV equation are also expressed by determinants whose elements consist of the eigenvalues $\lambda_{j}$ and the corresponding eigenfunctions of the associated Lax equation. The nonsingular $n$-positon solutions of the focusing mKdV equation are obtained in the special limit $\lambda_{j}\rightarrow\lambda_{1}$, from the corresponding $n$-soliton solutions and by using the associated higher-order Taylor expansion. Furthermore, the decomposition method of the $n$-positon solution into $n$ single-soliton solutions, the trajectories, and the corresponding "phase shifts" of the multi-positons are also investigated.
[ 0, 1, 0, 0, 0, 0 ]
[ "Mathematics", "Physics" ]