text
stringlengths 138
2.38k
| labels
sequencelengths 6
6
| Predictions
sequencelengths 1
3
|
---|---|---|
Title: Learning to Optimize Neural Nets,
Abstract: Learning to Optimize is a recently proposed framework for learning
optimization algorithms using reinforcement learning. In this paper, we explore
learning an optimization algorithm for training shallow neural nets. Such
high-dimensional stochastic optimization problems present interesting
challenges for existing reinforcement learning algorithms. We develop an
extension that is suited to learning optimization algorithms in this setting
and demonstrate that the learned optimization algorithm consistently
outperforms other known optimization algorithms even on unseen tasks and is
robust to changes in stochasticity of gradients and the neural net
architecture. More specifically, we show that an optimization algorithm trained
with the proposed method on the problem of training a neural net on MNIST
generalizes to the problems of training neural nets on the Toronto Faces
Dataset, CIFAR-10 and CIFAR-100. | [
1,
0,
1,
1,
0,
0
] | [
"Computer Science",
"Statistics"
] |
Title: Notes on complexity of packing coloring,
Abstract: A packing $k$-coloring for some integer $k$ of a graph $G=(V,E)$ is a mapping
$\varphi:V\to\{1,\ldots,k\}$ such that any two vertices $u, v$ of color
$\varphi(u)=\varphi(v)$ are in distance at least $\varphi(u)+1$. This concept
is motivated by frequency assignment problems. The \emph{packing chromatic
number} of $G$ is the smallest $k$ such that there exists a packing
$k$-coloring of $G$.
Fiala and Golovach showed that determining the packing chromatic number for
chordal graphs is \NP-complete for diameter exactly 5. While the problem is
easy to solve for diameter 2, we show \NP-completeness for any diameter at
least 3. Our reduction also shows that the packing chromatic number is hard to
approximate within $n^{{1/2}-\varepsilon}$ for any $\varepsilon > 0$.
In addition, we design an \FPT algorithm for interval graphs of bounded
diameter. This leads us to exploring the problem of finding a partial coloring
that maximizes the number of colored vertices. | [
1,
0,
0,
0,
0,
0
] | [
"Computer Science",
"Mathematics"
] |
Title: Improved Absolute Frequency Measurement of the 171Yb Optical Lattice Clock at KRISS Relative to the SI Second,
Abstract: We measured the absolute frequency of the $^1S_0$ - $^3P_0$ transition of
$^{171}$Yb atoms confined in a one-dimensional optical lattice relative to the
SI second. The determined frequency was 518 295 836 590 863.38(57) Hz. The
uncertainty was reduced by a factor of 14 compared with our previously reported
value in 2013 due to the significant improvements in decreasing the systematic
uncertainties. This result is expected to contribute to the determination of a
new recommended value for the secondary representations of the second. | [
0,
1,
0,
0,
0,
0
] | [
"Physics"
] |
Title: A Hilbert Space of Stationary Ergodic Processes,
Abstract: Identifying meaningful signal buried in noise is a problem of interest
arising in diverse scenarios of data-driven modeling. We present here a
theoretical framework for exploiting intrinsic geometry in data that resists
noise corruption, and might be identifiable under severe obfuscation. Our
approach is based on uncovering a valid complete inner product on the space of
ergodic stationary finite valued processes, providing the latter with the
structure of a Hilbert space on the real field. This rigorous construction,
based on non-standard generalizations of the notions of sum and scalar
multiplication of finite dimensional probability vectors, allows us to
meaningfully talk about "angles" between data streams and data sources, and,
make precise the notion of orthogonal stochastic processes. In particular, the
relative angles appear to be preserved, and identifiable, under severe noise,
and will be developed in future as the underlying principle for robust
classification, clustering and unsupervised featurization algorithms. | [
0,
0,
0,
1,
0,
1
] | [
"Mathematics",
"Statistics"
] |
Title: Total variation regularized non-negative matrix factorization for smooth hyperspectral unmixing,
Abstract: Hyperspectral analysis has gained popularity over recent years as a way to
infer what materials are displayed on a picture whose pixels consist of a
mixture of spectral signatures. Computing both signatures and mixture
coefficients is known as unsupervised unmixing, a set of techniques usually
based on non-negative matrix factorization. Unmixing is a difficult non-convex
problem, and algorithms may converge to one out of many local minima, which may
be far removed from the true global minimum. Computing this true minimum is
NP-hard and seems therefore out of reach. Aiming for interesting local minima,
we investigate the addition of total variation regularization terms. Advantages
of these regularizers are two-fold. Their computation is typically rather
light, and they are deemed to preserve sharp edges in pictures. This paper
describes an algorithm for regularized hyperspectral unmixing based on the
Alternating Direction Method of Multipliers. | [
0,
0,
1,
0,
0,
0
] | [
"Computer Science",
"Mathematics"
] |
Title: Efficient Antihydrogen Detection in Antimatter Physics by Deep Learning,
Abstract: Antihydrogen is at the forefront of antimatter research at the CERN
Antiproton Decelerator. Experiments aiming to test the fundamental CPT symmetry
and antigravity effects require the efficient detection of antihydrogen
annihilation events, which is performed using highly granular tracking
detectors installed around an antimatter trap. Improving the efficiency of the
antihydrogen annihilation detection plays a central role in the final
sensitivity of the experiments. We propose deep learning as a novel technique
to analyze antihydrogen annihilation data, and compare its performance with a
traditional track and vertex reconstruction method. We report that the deep
learning approach yields significant improvement, tripling event coverage while
simultaneously improving performance by over 5% in terms of Area Under Curve
(AUC). | [
1,
1,
0,
0,
0,
0
] | [
"Physics",
"Computer Science"
] |
Title: Nonparametric Cusum Charts for Angular Data with Applications in Health Science and Astrophysics,
Abstract: This paper develops non-parametric rotation invariant CUSUMs suited to the
detection of changes in the mean direction as well as changes in the
concentration parameter of angular data. The properties of the CUSUMs are
illustrated by theoretical calculations, Monte Carlo simulation and application
to sequentially observed angular data from health science and astrophysics. | [
0,
0,
0,
1,
0,
0
] | [
"Statistics",
"Physics"
] |
Title: Origin of Weak Turbulence in the Outer Regions of Protoplanetary Disks,
Abstract: The mechanism behind angular momentum transport in protoplanetary disks, and
whether this transport is turbulent in nature, is a fundamental issue in planet
formation studies. Recent ALMA observations have suggested that turbulent
velocities in the outer regions of these disks are less than ~5-10% of the
sound speed, contradicting theoretical predictions of turbulence driven by the
magnetorotational instability (MRI). These observations have generally been
interpreted to be consistent with a large-scale laminar magnetic wind driving
accretion. Here, we carry out local, shearing box simulations with varying
ionization levels and background magnetic field strengths in order to determine
which parameters produce results consistent with observations. We find that
even when the background magnetic field launches a strong largely laminar wind,
significant turbulence persists and is driven by localized regions of vertical
magnetic field (the result of zonal flows) that are unstable to the MRI. The
only conditions for which we find turbulent velocities below the observational
limits are weak background magnetic fields and ionization levels well below
that usually assumed in theoretical studies. We interpret these findings within
the context of a preliminary model in which a large scale magnetic field,
confined to the inner disk, hinders ionizing sources from reaching large radial
distances, e.g., through a sufficiently dense wind. Thus, in addition to such a
wind, this model predicts that for disks with weakly turbulent outer regions,
the outer disk will have significantly reduced ionization levels compared to
standard models and will harbor only a weak vertical magnetic field. | [
0,
1,
0,
0,
0,
0
] | [
"Physics"
] |
Title: Real-time Monocular Visual Odometry for Turbid and Dynamic Underwater Environments,
Abstract: In the context of robotic underwater operations, the visual degradations
induced by the medium properties make difficult the exclusive use of cameras
for localization purpose. Hence, most localization methods are based on
expensive navigational sensors associated with acoustic positioning. On the
other hand, visual odometry and visual SLAM have been exhaustively studied for
aerial or terrestrial applications, but state-of-the-art algorithms fail
underwater. In this paper we tackle the problem of using a simple low-cost
camera for underwater localization and propose a new monocular visual odometry
method dedicated to the underwater environment. We evaluate different tracking
methods and show that optical flow based tracking is more suited to underwater
images than classical approaches based on descriptors. We also propose a
keyframe-based visual odometry approach highly relying on nonlinear
optimization. The proposed algorithm has been assessed on both simulated and
real underwater datasets and outperforms state-of-the-art visual SLAM methods
under many of the most challenging conditions. The main application of this
work is the localization of Remotely Operated Vehicles (ROVs) used for
underwater archaeological missions but the developed system can be used in any
other applications as long as visual information is available. | [
1,
0,
0,
0,
0,
0
] | [
"Computer Science",
"Quantitative Biology"
] |
Title: Rating Protocol Design for Extortion and Cooperation in the Crowdsourcing Contest Dilemma,
Abstract: Crowdsourcing has emerged as a paradigm for leveraging human intelligence and
activity to solve a wide range of tasks. However, strategic workers will find
enticement in their self-interest to free-ride and attack in a crowdsourcing
contest dilemma game. Hence, incentive mechanisms are of great importance to
overcome the inefficiency of the socially undesirable equilibrium. Existing
incentive mechanisms are not effective in providing incentives for cooperation
in crowdsourcing competitions due to the following features: heterogeneous
workers compete against each other in a crowdsourcing platform with imperfect
monitoring. In this paper, we take these features into consideration, and
develop a novel game-theoretic design of rating protocols, which integrates
binary rating labels with differential pricing to maximize the requester's
utility, by extorting selfish workers and enforcing cooperation among them. By
quantifying necessary and sufficient conditions for the sustainable social
norm, we formulate the problem of maximizing the revenue of the requester among
all sustainable rating protocols, provide design guidelines for optimal rating
protocols, and design a low-complexity algorithm to select optimal design
parameters which are related to differential punishments and pricing schemes.
Simulation results demonstrate how intrinsic parameters impact on design
parameters, as well as the performance gain of the proposed rating protocol. | [
1,
0,
0,
0,
0,
0
] | [
"Computer Science",
"Mathematics"
] |
Title: On the $E$-polynomial of parabolic $\mathrm{Sp}_{2n}$-character varieties,
Abstract: We find the $E$-polynomials of a family of parabolic
$\mathrm{Sp}_{2n}$-character varieties $\mathcal{M}^{\xi}_{n}$ of Riemann
surfaces by constructing a stratification, proving that each stratum has
polynomial count, applying a result of Katz regarding the counting functions,
and finally adding up the resulting $E$-polynomials of the strata. To count the
number of $\mathbb{F}_{q}$-points of the strata, we invoke a formula due to
Frobenius. Our calculation make use of a formula for the evaluation of
characters on semisimple elements coming from Deligne-Lusztig theory, applied
to the character theory of $\mathrm{Sp}{\left(2n,\mathbb{F}_{q}\right)}$, and
Möbius inversion on the poset of set-partitions. We compute the Euler
characteristic of the $\mathcal{M}^{\xi}_{n}$ with these polynomials, and show
they are connected. | [
0,
0,
1,
0,
0,
0
] | [
"Mathematics"
] |
Title: Ivanov-Regularised Least-Squares Estimators over Large RKHSs and Their Interpolation Spaces,
Abstract: We study kernel least-squares estimation under a norm constraint. This form
of regularisation is known as Ivanov regularisation and it provides better
control of the norm of the estimator than the well-established Tikhonov
regularisation. This choice of regularisation allows us to dispose of the
standard assumption that the reproducing kernel Hilbert space (RKHS) has a
Mercer kernel, which is restrictive as it usually requires compactness of the
covariate set. Instead, we assume only that the RKHS is separable with a
bounded and measurable kernel. We provide rates of convergence for the expected
squared $L^2$ error of our estimator under the weak assumption that the
variance of the response variables is bounded and the unknown regression
function lies in an interpolation space between $L^2$ and the RKHS. We then
obtain faster rates of convergence when the regression function is bounded by
clipping the estimator. In fact, we attain the optimal rate of convergence.
Furthermore, we provide a high-probability bound under the stronger assumption
that the response variables have subgaussian errors and that the regression
function lies in an interpolation space between $L^\infty$ and the RKHS.
Finally, we derive adaptive results for the settings in which the regression
function is bounded. | [
0,
0,
1,
1,
0,
0
] | [
"Statistics",
"Mathematics",
"Computer Science"
] |
Title: Using Ice and Dust Lines to Constrain the Surface Densities of Protoplanetary Disks,
Abstract: We present a novel method for determining the surface density of
protoplanetary disks through consideration of disk 'dust lines' which indicate
the observed disk radial scale at different observational wavelengths. This
method relies on the assumption that the processes of particle growth and drift
control the radial scale of the disk at late stages of disk evolution such that
the lifetime of the disk is equal to both the drift timescale and growth
timescale of the maximum particle size at a given dust line. We provide an
initial proof of concept of our model through an application to the disk TW Hya
and are able to estimate the disk dust-to-gas ratio, CO abundance, and
accretion rate in addition to the total disk surface density. We find that our
derived surface density profile and dust-to-gas ratio are consistent with the
lower limits found through measurements of HD gas. The CO ice line also depends
on surface density through grain adsorption rates and drift and we find that
our theoretical CO ice line estimates have clear observational analogues. We
further apply our model to a large parameter space of theoretical disks and
find three observational diagnostics that may be used to test its validity.
First we predict that the dust lines of disks other than TW Hya will be
consistent with the normalized CO surface density profile shape for those
disks. Second, surface density profiles that we derive from disk ice lines
should match those derived from disk dust lines. Finally, we predict that disk
dust and ice lines will scale oppositely, as a function of surface density,
across a large sample of disks. | [
0,
1,
0,
0,
0,
0
] | [
"Physics",
"Astrophysics"
] |
Title: Quaternionic Projective Bundle Theorem and Gysin Triangle in MW-Motivic Cohomology,
Abstract: In this paper, we show that the motive $HP^n$ of the quaternionic
Grassmannian (as defined by I. Panin and C. Walter) splits in the category of
effective MW-motives (as defined by B. Calmès, F. Déglise and J. Fasel).
Moreover, we extend this result to an arbitrary symplectic bundle, obtaining
the so-called quaternionic projective bundle theorem. This enables us to define
Pontryagin classes of symplectic bundles in the Chow-Witt ring.
As an application, we prove that there is a Gysin triangle in MW-motivic
cohomology in case the normal bundle is symplectic. | [
0,
0,
1,
0,
0,
0
] | [
"Mathematics"
] |
Title: Summary of a Literature Review in Scalability of QoS-aware Service Composition,
Abstract: This paper shows that authors have no consistent way to characterize the
scalability of their solutions, and so consider only a limited number of
scaling characteristics. This review aimed at establishing the evidence that
the route for designing and evaluating the scalability of dynamic QoS-aware
service composition mechanisms has been lacking systematic guidance, and has
been informed by a very limited set of criteria. For such, we analyzed 47
papers, from 2004 to 2018. | [
1,
0,
0,
0,
0,
0
] | [
"Computer Science"
] |
Title: On the functional window of the avian compass,
Abstract: The functional window is an experimentally observed property of the avian
compass that refers to its selectivity around the geomagnetic field strength.
We show that the radical-pair model, using biologically feasible hyperfine
parameters, can qualitatively explain the salient features of the avian compass
as observed from behavioral experiments: its functional window, as well as
disruption of the compass action by an RF field of specific frequencies.
Further, we show that adjustment of the hyperfine parameters can tune the
functional window, suggesting a possible mechanism for its observed
adaptability to field variation. While these lend strong support to the
radical-pair model, we find it impossible to explain quantitatively the
observed width of the functional window within this model, or even with simple
augmentations thereto. This suggests that a deeper generalization of this model
may be called for; we conjecture that environmental coupling may be playing a
subtle role here that has not been captured accurately. Lastly, we examine a
possible biological purpose to the functional window; assuming evolutionary
benefit from radical-pair magnetoreception, we conjecture that the functional
window is simply a corollary thereof and brings no additional advantage. | [
0,
1,
0,
0,
0,
0
] | [
"Physics",
"Quantitative Biology"
] |
Title: Automated text summarisation and evidence-based medicine: A survey of two domains,
Abstract: The practice of evidence-based medicine (EBM) urges medical practitioners to
utilise the latest research evidence when making clinical decisions. Because of
the massive and growing volume of published research on various medical topics,
practitioners often find themselves overloaded with information. As such,
natural language processing research has recently commenced exploring
techniques for performing medical domain-specific automated text summarisation
(ATS) techniques-- targeted towards the task of condensing large medical texts.
However, the development of effective summarisation techniques for this task
requires cross-domain knowledge. We present a survey of EBM, the
domain-specific needs for EBM, automated summarisation techniques, and how they
have been applied hitherto. We envision that this survey will serve as a first
resource for the development of future operational text summarisation
techniques for EBM. | [
1,
0,
0,
0,
0,
0
] | [
"Computer Science",
"Quantitative Biology"
] |
Title: Generalized Moran sets Generated by Step-wise Adjustable Iterated Function Systems,
Abstract: In this article we provide a systematic way of creating generalized Moran
sets using an analogous iterated function system (IFS) procedure. We use a
step-wise adjustable IFS to introduce some variance (such as
non-self-similarity) in the fractal limit sets. The process retains the
computational simplicity of a standard IFS procedure. In our construction of
the generalized Moran sets, we also weaken the fourth Moran Structure Condition
that requires the same pattern of diameter ratios be used across a generation.
Moreover, we provide upper and lower bounds for the Hausdorff dimension of the
fractals created from this generalized process. Specific examples (Cantor-like
sets, Sierpinski-like Triangles, etc) with the calculations of their
corresponding dimensions are studied. | [
0,
1,
1,
0,
0,
0
] | [
"Mathematics"
] |
Title: Isospectrality For Orbifold Lens Spaces,
Abstract: We answer Mark Kac's famous question, "can one hear the shape of a drum?" in
the positive for orbifolds that are 3-dimensional and 4-dimensional lens
spaces; we thus complete the answer to this question for orbifold lens spaces
in all dimensions. We also show that the coefficients of the asymptotic
expansion of the trace of the heat kernel are not sufficient to determine the
above results. | [
0,
0,
1,
0,
0,
0
] | [
"Mathematics",
"Physics"
] |
Title: DCCO: Towards Deformable Continuous Convolution Operators,
Abstract: Discriminative Correlation Filter (DCF) based methods have shown competitive
performance on tracking benchmarks in recent years. Generally, DCF based
trackers learn a rigid appearance model of the target. However, this reliance
on a single rigid appearance model is insufficient in situations where the
target undergoes non-rigid transformations. In this paper, we propose a unified
formulation for learning a deformable convolution filter. In our framework, the
deformable filter is represented as a linear combination of sub-filters. Both
the sub-filter coefficients and their relative locations are inferred jointly
in our formulation. Experiments are performed on three challenging tracking
benchmarks: OTB- 2015, TempleColor and VOT2016. Our approach improves the
baseline method, leading to performance comparable to state-of-the-art. | [
1,
0,
0,
0,
0,
0
] | [
"Computer Science"
] |
Title: New X-ray bound on density of primordial black holes,
Abstract: We set a new upper limit on the abundance of primordial black holes (PBH)
based on existing X-ray data. PBH interactions with interstellar medium should
result in significant fluxes of X-ray photons, which would contribute to the
observed number density of compact X-ray objects in galaxies. The data
constrain PBH number density in the mass range from a few $M_\odot$ to $2\times
10^7 M_\odot$. PBH density needed to account for the origin of black holes
detected by LIGO is marginally allowed. | [
0,
1,
0,
0,
0,
0
] | [
"Physics"
] |
Title: Physics Informed Deep Learning (Part II): Data-driven Discovery of Nonlinear Partial Differential Equations,
Abstract: We introduce physics informed neural networks -- neural networks that are
trained to solve supervised learning tasks while respecting any given law of
physics described by general nonlinear partial differential equations. In this
second part of our two-part treatise, we focus on the problem of data-driven
discovery of partial differential equations. Depending on whether the available
data is scattered in space-time or arranged in fixed temporal snapshots, we
introduce two main classes of algorithms, namely continuous time and discrete
time models. The effectiveness of our approach is demonstrated using a wide
range of benchmark problems in mathematical physics, including conservation
laws, incompressible fluid flow, and the propagation of nonlinear shallow-water
waves. | [
1,
0,
0,
1,
0,
0
] | [
"Computer Science",
"Physics",
"Mathematics"
] |
Title: The HoTT reals coincide with the Escardó-Simpson reals,
Abstract: Escardó and Simpson defined a notion of interval object by a universal
property in any category with binary products. The Homotopy Type Theory book
defines a higher-inductive notion of reals, and suggests that the interval may
satisfy this universal property. We show that this is indeed the case in the
category of sets of any universe. We also show that the type of HoTT reals is
the least Cauchy complete subset of the Dedekind reals containing the
rationals. | [
1,
0,
1,
0,
0,
0
] | [
"Mathematics",
"Computer Science"
] |
Title: Noise2Noise: Learning Image Restoration without Clean Data,
Abstract: We apply basic statistical reasoning to signal reconstruction by machine
learning -- learning to map corrupted observations to clean signals -- with a
simple and powerful conclusion: it is possible to learn to restore images by
only looking at corrupted examples, at performance at and sometimes exceeding
training using clean data, without explicit image priors or likelihood models
of the corruption. In practice, we show that a single model learns photographic
noise removal, denoising synthetic Monte Carlo images, and reconstruction of
undersampled MRI scans -- all corrupted by different processes -- based on
noisy data only. | [
0,
0,
0,
1,
0,
0
] | [
"Computer Science",
"Statistics"
] |
Title: Jeffrey's prior sampling of deep sigmoidal networks,
Abstract: Neural networks have been shown to have a remarkable ability to uncover low
dimensional structure in data: the space of possible reconstructed images form
a reduced model manifold in image space. We explore this idea directly by
analyzing the manifold learned by Deep Belief Networks and Stacked Denoising
Autoencoders using Monte Carlo sampling. The model manifold forms an only
slightly elongated hyperball with actual reconstructed data appearing
predominantly on the boundaries of the manifold. In connection with the results
we present, we discuss problems of sampling high-dimensional manifolds as well
as recent work [M. Transtrum, G. Hart, and P. Qiu, Submitted (2014)] discussing
the relation between high dimensional geometry and model reduction. | [
1,
1,
0,
0,
0,
0
] | [
"Computer Science",
"Statistics"
] |
Title: Faster Learning by Reduction of Data Access Time,
Abstract: Nowadays, the major challenge in machine learning is the Big Data challenge.
The big data problems due to large number of data points or large number of
features in each data point, or both, the training of models have become very
slow. The training time has two major components: Time to access the data and
time to process (learn from) the data. So far, the research has focused only on
the second part, i.e., learning from the data. In this paper, we have proposed
one possible solution to handle the big data problems in machine learning. The
idea is to reduce the training time through reducing data access time by
proposing systematic sampling and cyclic/sequential sampling to select
mini-batches from the dataset. To prove the effectiveness of proposed sampling
techniques, we have used Empirical Risk Minimization, which is commonly used
machine learning problem, for strongly convex and smooth case. The problem has
been solved using SAG, SAGA, SVRG, SAAG-II and MBSGD (Mini-batched SGD), each
using two step determination techniques, namely, constant step size and
backtracking line search method. Theoretical results prove the same convergence
for systematic sampling, cyclic sampling and the widely used random sampling
technique, in expectation. Experimental results with bench marked datasets
prove the efficacy of the proposed sampling techniques and show up to six times
faster training. | [
0,
0,
0,
1,
0,
0
] | [
"Computer Science",
"Statistics"
] |
Title: ScaleSimulator: A Fast and Cycle-Accurate Parallel Simulator for Architectural Exploration,
Abstract: Design of next generation computer systems should be supported by simulation
infrastructure that must achieve a few contradictory goals such as fast
execution time, high accuracy, and enough flexibility to allow comparison
between large numbers of possible design points. Most existing architecture
level simulators are designed to be flexible and to execute the code in
parallel for greater efficiency, but at the cost of scarified accuracy. This
paper presents the ScaleSimulator simulation environment, which is based on a
new design methodology whose goal is to achieve near cycle accuracy while still
being flexible enough to simulate many different future system architectures
and efficient enough to run meaningful workloads. We achieve these goals by
making the parallelism a first-class citizen in our methodology. Thus, this
paper focuses mainly on the ScaleSimulator design points that enable better
parallel execution while maintaining the scalability and cycle accuracy of a
simulated architecture. The paper indicates that the new proposed
ScaleSimulator tool can (1) efficiently parallelize the execution of a
cycle-accurate architecture simulator, (2) efficiently simulate complex
architectures (e.g., out-of-order CPU pipeline, cache coherency protocol, and
network) and massive parallel systems, and (3) use meaningful workloads, such
as full simulation of OLTP benchmarks, to examine future architectural choices. | [
1,
0,
0,
0,
0,
0
] | [
"Computer Science"
] |
Title: Kahler-Einstein metrics and algebraic geometry,
Abstract: This is a survey article, based on the author's lectures in the 2015 Current
developments in Mathematics meeting; published in "Current developments in
Mathematics". Version 2, references corrected and added. | [
0,
0,
1,
0,
0,
0
] | [
"Mathematics"
] |
Title: AspEm: Embedding Learning by Aspects in Heterogeneous Information Networks,
Abstract: Heterogeneous information networks (HINs) are ubiquitous in real-world
applications. Due to the heterogeneity in HINs, the typed edges may not fully
align with each other. In order to capture the semantic subtlety, we propose
the concept of aspects with each aspect being a unit representing one
underlying semantic facet. Meanwhile, network embedding has emerged as a
powerful method for learning network representation, where the learned
embedding can be used as features in various downstream applications.
Therefore, we are motivated to propose a novel embedding learning
framework---AspEm---to preserve the semantic information in HINs based on
multiple aspects. Instead of preserving information of the network in one
semantic space, AspEm encapsulates information regarding each aspect
individually. In order to select aspects for embedding purpose, we further
devise a solution for AspEm based on dataset-wide statistics. To corroborate
the efficacy of AspEm, we conducted experiments on two real-words datasets with
two types of applications---classification and link prediction. Experiment
results demonstrate that AspEm can outperform baseline network embedding
learning methods by considering multiple aspects, where the aspects can be
selected from the given HIN in an unsupervised manner. | [
1,
0,
0,
0,
0,
0
] | [
"Computer Science",
"Statistics"
] |
Title: The energy-momentum tensor of electromagnetic fields in matter,
Abstract: We present a complete resolution of the Abraham-Minkowski controversy . This
is done by considering several new aspects which invalidate previous
discussions. We show that: 1)For polarized matter the center of mass theorem is
no longer valid in its usual form. A contribution related to microscopic spin
should be considered. 2)The electromagnetic dipolar energy density contributes
to the inertia of matter and should be incorporated covariantly to the the
energy-momentum tensor of matter. Then there is an electromagnetic component in
matter's momentum density whose variation explains the results of the only
experiment which supports Abraham's force. 3)Averaging the microscopic
Lorentz's force results in the unambiguos expression for the force density
exerted by the field. This force density is consistent with all the
experimental evidence. 4)Momentum conservation determines the electromagnetic
energy-momentum tensor. This tensor is different from Abraham's and Minkowski's
tensors, but one recovers Minkowski's expression for the momentum density. The
energy density is different from Poynting's expression but Poynting's vector
remains the same. Our tensor is non-symmetric which allows the field to exert a
distributed torque on matter. We use our results to discuss momentum and
angular momentum exchange in various situations of physical interest. We find
complete consistency of our equations in the description of the systems
considered. We also show that several alternative expressions of the field
energy-momentum tensor and force-density cannot be successfully used in all our
examples. In particular we verify in two of these examples that the center of
mass and spin introduced by us moves with constant velocity, but that the
standard center of mass does not. | [
0,
1,
0,
0,
0,
0
] | [
"Physics"
] |
Title: Tuning Free Orthogonal Matching Pursuit,
Abstract: Orthogonal matching pursuit (OMP) is a widely used compressive sensing (CS)
algorithm for recovering sparse signals in noisy linear regression models. The
performance of OMP depends on its stopping criteria (SC). SC for OMP discussed
in literature typically assumes knowledge of either the sparsity of the signal
to be estimated $k_0$ or noise variance $\sigma^2$, both of which are
unavailable in many practical applications. In this article we develop a
modified version of OMP called tuning free OMP or TF-OMP which does not require
a SC. TF-OMP is proved to accomplish successful sparse recovery under the usual
assumptions on restricted isometry constants (RIC) and mutual coherence of
design matrix. TF-OMP is numerically shown to deliver a highly competitive
performance in comparison with OMP having \textit{a priori} knowledge of $k_0$
or $\sigma^2$. Greedy algorithm for robust de-noising (GARD) is an OMP like
algorithm proposed for efficient estimation in classical overdetermined linear
regression models corrupted by sparse outliers. However, GARD requires the
knowledge of inlier noise variance which is difficult to estimate. We also
produce a tuning free algorithm (TF-GARD) for efficient estimation in the
presence of sparse outliers by extending the operating principle of TF-OMP to
GARD. TF-GARD is numerically shown to achieve a performance comparable to that
of the existing implementation of GARD. | [
1,
0,
0,
1,
0,
0
] | [
"Computer Science",
"Mathematics",
"Statistics"
] |
Title: New constructions of MDS codes with complementary duals,
Abstract: Linear complementary-dual (LCD for short) codes are linear codes that
intersect with their duals trivially. LCD codes have been used in certain
communication systems. It is recently found that LCD codes can be applied in
cryptography. This application of LCD codes renewed the interest in the
construction of LCD codes having a large minimum distance. MDS codes are
optimal in the sense that the minimum distance cannot be improved for given
length and code size. Constructing LCD MDS codes is thus of significance in
theory and practice. Recently, Jin (\cite{Jin}, IEEE Trans. Inf. Theory, 2016)
constructed several classes of LCD MDS codes through generalized Reed-Solomon
codes. In this paper, a different approach is proposed to obtain new LCD MDS
codes from generalized Reed-Solomon codes. Consequently, new code constructions
are provided and certain previously known results in \cite{Jin} are extended. | [
1,
0,
0,
0,
0,
0
] | [
"Computer Science",
"Mathematics"
] |
Title: Porosity and regularity in metric measure spaces,
Abstract: This is a report of a joint work with E. Järvenpää, M. Järvenpää,
T. Rajala, S. Rogovin, and V. Suomala. In [3], we characterized uniformly
porous sets in $s$-regular metric spaces in terms of regular sets by verifying
that a set $A$ is uniformly porous if and only if there is $t < s$ and a
$t$-regular set $F \supset A$. Here we outline the main idea of the proof and
also present an alternative proof for the crucial lemma needed in the proof of
the result. | [
0,
0,
1,
0,
0,
0
] | [
"Mathematics"
] |
Title: Strong-coupling superconductivity induced by calcium intercalation in bilayer transition-metal dichalcogenides,
Abstract: We theoretically investigate the possibility of achieving a superconducting
state in transition-metal dichalcogenide bilayers through intercalation, a
process previously and widely used to achieve metallization and superconducting
states in novel superconductors. For the Ca-intercalated bilayers MoS$_2$ and
WS$_2$, we find that the superconducting state is characterized by an
electron-phonon coupling constant larger than $1.0$ and a superconducting
critical temperature of $13.3$ and $9.3$ K, respectively. These results are
superior to other predicted or experimentally observed two-dimensional
conventional superconductors and suggest that the investigated materials may be
good candidates for nanoscale superconductors. More interestingly, we proved
that the obtained thermodynamic properties go beyond the predictions of the
mean-field Bardeen--Cooper--Schrieffer approximation and that the calculations
conducted within the framework of the strong-coupling Eliashberg theory should
be treated as those that yield quantitative results. | [
0,
1,
0,
0,
0,
0
] | [
"Physics"
] |
Title: Global existence for the nonlinear fractional Schrödinger equation with fractional dissipation,
Abstract: We consider the initial value problem for the fractional nonlinear
Schrödinger equation with a fractional dissipation. Global existence and
scattering are proved depending on the order of the fractional dissipation. | [
0,
0,
1,
0,
0,
0
] | [
"Mathematics",
"Physics"
] |
Title: Statistical properties of an enstrophy conserving discretisation for the stochastic quasi-geostrophic equation,
Abstract: A framework of variational principles for stochastic fluid dynamics was
presented by Holm (2015), and these stochastic equations were also derived by
Cotter et al. (2017). We present a conforming finite element discretisation for
the stochastic quasi-geostrophic equation that was derived from this framework.
The discretisation preserves the first two moments of potential vorticity, i.e.
the mean potential vorticity and the enstrophy. Following the work of Dubinkina
and Frank (2007), who investigated the statistical mechanics of discretisations
of the deterministic quasi-geostrophic equation, we investigate the statistical
mechanics of our discretisation of the stochastic quasi-geostrophic equation.
We compare the statistical properties of our discretisation with the Gibbs
distribution under assumption of these conserved quantities, finding that there
is agreement between the statistics under a wide range of set-ups. | [
0,
1,
0,
0,
0,
0
] | [
"Physics",
"Mathematics",
"Statistics"
] |
Title: Improved electronic structure and magnetic exchange interactions in transition metal oxides,
Abstract: We discuss the application of the Agapito Curtarolo and Buongiorno Nardelli
(ACBN0) pseudo-hybrid Hubbard density functional to several transition metal
oxides. ACBN0 is a fast, accurate and parameter-free alternative to traditional
DFT+$U$ and hybrid exact exchange methods. In ACBN0, the Hubbard energy of
DFT+$U$ is calculated via the direct evaluation of the local Coulomb and
exchange integrals in which the screening of the bare Coulomb potential is
accounted for by a renormalization of the density matrix. We demonstrate the
success of the ACBN0 approach for the electronic properties of a series
technologically relevant mono-oxides (MnO, CoO, NiO, FeO, both at equilibrium
and under pressure). We also present results on two mixed valence compounds,
Co$_3$O$_4$ and Mn$_3$O$_4$. Our results, obtained at the computational cost of
a standard LDA/PBE calculation, are in excellent agreement with hybrid
functionals, the GW approximation and experimental measurements. | [
0,
1,
0,
0,
0,
0
] | [
"Physics",
"Computer Science"
] |
Title: Neon2: Finding Local Minima via First-Order Oracles,
Abstract: We propose a reduction for non-convex optimization that can (1) turn an
stationary-point finding algorithm into an local-minimum finding one, and (2)
replace the Hessian-vector product computations with only gradient
computations. It works both in the stochastic and the deterministic settings,
without hurting the algorithm's performance.
As applications, our reduction turns Natasha2 into a first-order method
without hurting its performance. It also converts SGD, GD, SCSG, and SVRG into
algorithms finding approximate local minima, outperforming some best known
results. | [
1,
0,
0,
1,
0,
0
] | [
"Computer Science",
"Mathematics"
] |
Title: Simple Countermeasures to Mitigate the Effect of Pollution Attack in Network Coding Based Peer-to-Peer Live Streaming,
Abstract: Network coding based peer-to-peer streaming represents an effective solution
to aggregate user capacities and to increase system throughput in live
multimedia streaming. Nonetheless, such systems are vulnerable to pollution
attacks where a handful of malicious peers can disrupt the communication by
transmitting just a few bogus packets which are then recombined and relayed by
unaware honest nodes, further spreading the pollution over the network. Whereas
previous research focused on malicious nodes identification schemes and
pollution-resilient coding, in this paper we show pollution countermeasures
which make a standard network coding scheme resilient to pollution attacks.
Thanks to a simple yet effective analytical model of a reference node
collecting packets by malicious and honest neighbors, we demonstrate that i)
packets received earlier are less likely to be polluted and ii) short
generations increase the likelihood to recover a clean generation. Therefore,
we propose a recombination scheme where nodes draw packets to be recombined
according to their age in the input queue, paired with a decoding scheme able
to detect the reception of polluted packets early in the decoding process and
short generations. The effectiveness of our approach is experimentally
evaluated in a real system we developed and deployed on hundreds to thousands
peers. Experimental evidence shows that, thanks to our simple countermeasures,
the effect of a pollution attack is almost canceled and the video quality
experienced by the peers is comparable to pre-attack levels. | [
1,
0,
0,
0,
0,
0
] | [
"Computer Science"
] |
Title: Small-scale structure and the Lyman-$α$ forest baryon acoustic oscillation feature,
Abstract: The baryon-acoustic oscillation (BAO) feature in the Lyman-$\alpha$ forest is
one of the key probes of the cosmic expansion rate at redshifts z~2.5, well
before dark energy is believed to have become dynamically significant. A key
advantage of the BAO as a standard ruler is that it is a sharp feature and
hence is more robust against broadband systematic effects than other
cosmological probes. However, if the Lyman-$\alpha$ forest transmission is
sensitive to the initial streaming velocity of the baryons relative to the dark
matter, then the BAO peak position can be shifted. Here we investigate this
sensitivity using a suite of hydrodynamic simulations of small regions of the
intergalactic medium with a range of box sizes and physics assumptions; each
simulation starts from initial conditions at the kinematic decoupling era
(z~1059), undergoes a discrete change from neutral gas to ionized gas thermal
evolution at reionization (z~8), and is finally processed into a Lyman-$\alpha$
forest transmitted flux cube. Streaming velocities suppress small-scale
structure, leading to less violent relaxation after reionization. The changes
in the gas distribution and temperature-density relation at low redshift are
more subtle, due to the convergent temperature evolution in the ionized phase.
The change in the BAO scale is estimated to be of the order of 0.12% at z=2.5;
some of the major uncertainties and avenues for future improvement are
discussed. The predicted streaming velocity shift would be a subdominant but
not negligible effect (of order $0.26\sigma$) for the upcoming DESI
Lyman-$\alpha$ forest survey, and exceeds the cosmic variance floor. It is
hoped that this study will motivate additional theoretical work on the
magnitude of the BAO shift, both in the Lyman-$\alpha$ forest and in other
tracers of large-scale structure. | [
0,
1,
0,
0,
0,
0
] | [
"Physics"
] |
Title: InfoCatVAE: Representation Learning with Categorical Variational Autoencoders,
Abstract: This paper describes InfoCatVAE, an extension of the variational autoencoder
that enables unsupervised disentangled representation learning. InfoCatVAE uses
multimodal distributions for the prior and the inference network and then
maximizes the evidence lower bound objective (ELBO). We connect the new ELBO
derived for our model with a natural soft clustering objective which explains
the robustness of our approach. We then adapt the InfoGANs method to our
setting in order to maximize the mutual information between the categorical
code and the generated inputs and obtain an improved model. | [
0,
0,
0,
1,
0,
0
] | [
"Computer Science",
"Statistics"
] |
Title: Ulrich bundles on smooth projective varieties of minimal degree,
Abstract: We classify the Ulrich vector bundles of arbitrary rank on smooth projective
varieties of minimal degree. In the process, we prove the stability of the
sheaves of relative differentials on rational scrolls. | [
0,
0,
1,
0,
0,
0
] | [
"Mathematics"
] |
Title: $ε$-Regularity and Structure of 4-dimensional Shrinking Ricci Solitons,
Abstract: A closed four dimensional manifold cannot possess a non-flat Ricci soliton
metric with arbitrarily small $L^2$-norm of the curvature. In this paper, we
localize this fact in the case of shrinking Ricci solitons by proving an
$\varepsilon$-regularity theorem, thus confirming a conjecture of Cheeger-Tian.
As applications, we will also derive structural results concerning the
degeneration of the metrics on a family of complete non-compact four
dimensional shrinking Ricci solitons without a uniform entropy lower bound. In
the appendix, we provide a detailed account of the equivariant good chopping
theorem when collapsing with locally bounded curvature happens. | [
0,
0,
1,
0,
0,
0
] | [
"Mathematics"
] |
Title: Deep Memory Networks for Attitude Identification,
Abstract: We consider the task of identifying attitudes towards a given set of entities
from text. Conventionally, this task is decomposed into two separate subtasks:
target detection that identifies whether each entity is mentioned in the text,
either explicitly or implicitly, and polarity classification that classifies
the exact sentiment towards an identified entity (the target) into positive,
negative, or neutral.
Instead, we show that attitude identification can be solved with an
end-to-end machine learning architecture, in which the two subtasks are
interleaved by a deep memory network. In this way, signals produced in target
detection provide clues for polarity classification, and reversely, the
predicted polarity provides feedback to the identification of targets.
Moreover, the treatments for the set of targets also influence each other --
the learned representations may share the same semantics for some targets but
vary for others. The proposed deep memory network, the AttNet, outperforms
methods that do not consider the interactions between the subtasks or those
among the targets, including conventional machine learning methods and the
state-of-the-art deep learning models. | [
1,
0,
0,
0,
0,
0
] | [
"Computer Science"
] |
Title: Discrete flow posteriors for variational inference in discrete dynamical systems,
Abstract: Each training step for a variational autoencoder (VAE) requires us to sample
from the approximate posterior, so we usually choose simple (e.g. factorised)
approximate posteriors in which sampling is an efficient computation that fully
exploits GPU parallelism. However, such simple approximate posteriors are often
insufficient, as they eliminate statistical dependencies in the posterior.
While it is possible to use normalizing flow approximate posteriors for
continuous latents, some problems have discrete latents and strong statistical
dependencies. The most natural approach to model these dependencies is an
autoregressive distribution, but sampling from such distributions is inherently
sequential and thus slow. We develop a fast, parallel sampling procedure for
autoregressive distributions based on fixed-point iterations which enables
efficient and accurate variational inference in discrete state-space latent
variable dynamical systems. To optimize the variational bound, we considered
two ways to evaluate probabilities: inserting the relaxed samples directly into
the pmf for the discrete distribution, or converting to continuous logistic
latent variables and interpreting the K-step fixed-point iterations as a
normalizing flow. We found that converting to continuous latent variables gave
considerable additional scope for mismatch between the true and approximate
posteriors, which resulted in biased inferences, we thus used the former
approach. Using our fast sampling procedure, we were able to realize the
benefits of correlated posteriors, including accurate uncertainty estimates for
one cell, and accurate connectivity estimates for multiple cells, in an order
of magnitude less time. | [
0,
0,
0,
1,
1,
0
] | [
"Computer Science",
"Statistics"
] |
Title: Thermoelectric power factor enhancement by spin-polarized currents - a nanowire case study,
Abstract: Thermoelectric (TE) measurements have been performed on the workhorses of
today's data storage devices, exhibiting either the giant or the anisotropic
magnetoresistance effect (GMR and AMR). The temperature-dependent (50-300 K)
and magnetic field-dependent (up to 1 T) TE power factor (PF) has been
determined for several Co-Ni alloy nanowires with varying Co:Ni ratios as well
as for Co-Ni/Cu multilayered nanowires with various Cu layer thicknesses, which
were all synthesized via a template-assisted electrodeposition process. A
systematic investigation of the resistivity, as well as the Seebeck
coefficient, is performed for Co-Ni alloy nanowires and Co-Ni/Cu multilayered
nanowires. At room temperature, measured values of TE PFs up to 3.6 mWK-2m-1
for AMR samples and 2.0 mWK-2m-1 for GMR nanowires are obtained. Furthermore,
the TE PF is found to increase by up to 13.1 % for AMR Co-Ni alloy nanowires
and by up to 52 % for GMR Co-Ni/Cu samples in an external applied magnetic
field. The magnetic nanowires exhibit TE PFs that are of the same order of
magnitude as TE PFs of Bi-Sb-Se-Te based thermoelectric materials and,
additionally, give the opportunity to adjust the TE power output to changing
loads and hotspots through external magnetic fields. | [
0,
1,
0,
0,
0,
0
] | [
"Physics"
] |
Title: Risk-Sensitive Cooperative Games for Human-Machine Systems,
Abstract: Autonomous systems can substantially enhance a human's efficiency and
effectiveness in complex environments. Machines, however, are often unable to
observe the preferences of the humans that they serve. Despite the fact that
the human's and machine's objectives are aligned, asymmetric information, along
with heterogeneous sensitivities to risk by the human and machine, make their
joint optimization process a game with strategic interactions. We propose a
framework based on risk-sensitive dynamic games; the human seeks to optimize
her risk-sensitive criterion according to her true preferences, while the
machine seeks to adaptively learn the human's preferences and at the same time
provide a good service to the human. We develop a class of performance measures
for the proposed framework based on the concept of regret. We then evaluate
their dependence on the risk-sensitivity and the degree of uncertainty. We
present applications of our framework to self-driving taxis, and robo-financial
advising. | [
1,
0,
0,
1,
0,
0
] | [
"Computer Science",
"Quantitative Finance"
] |
Title: On Gauge Invariance and Covariant Derivatives in Metric Spaces,
Abstract: In this manuscript, we will discuss the construction of covariant derivative
operator in quantum gravity. We will find it is appropriate to use affine
connections more general than metric compatible connections in quantum gravity.
We will demonstrate this using the canonical quantization procedure. This is
valid irrespective of the presence and nature of sources. The standard Palatini
formalism, where metric and affine connections are the independent variables,
is not sufficient to construct a source-free theory of gravity with affine
connections more general than the metric compatible Levi-Civita connections.
This is also valid for minimally coupled interacting theories where sources
only couple with metric by using the metric compatible Levi-Civita connections
exclusively. We will discuss a potential formalism and possible extensions of
the action to introduce nonmetricity in these cases. This is also required to
construct a general interacting quantum theory with dynamical general affine
connections. We will have to use a modified Ricci tensor to state Einstein's
equation in the Palatini formalism. General affine connections can be described
by a third rank tensor with one contravariant and two covariant indices.
Antisymmetric part of this tensor in the lower indices gives torsion with a
half factor. In the Palatini formalism or its generalizations considered here,
symmetric part of this tensor in the lower indices is finite when torsion is
finite. This part can give a massless scalar field in a potential formalism. We
will have to extend the local conservation laws when we use general affine
connections. General affine connections can become significant to solve
cosmological problems. | [
0,
1,
0,
0,
0,
0
] | [
"Physics"
] |
Title: A Compressed Sensing Approach for Distribution Matching,
Abstract: In this work, we formulate the fixed-length distribution matching as a
Bayesian inference problem. Our proposed solution is inspired from the
compressed sensing paradigm and the sparse superposition (SS) codes. First, we
introduce sparsity in the binary source via position modulation (PM). We then
present a simple and exact matcher based on Gaussian signal quantization. At
the receiver, the dematcher exploits the sparsity in the source and performs
low-complexity dematching based on generalized approximate message-passing
(GAMP). We show that GAMP dematcher and spatial coupling lead to asymptotically
optimal performance, in the sense that the rate tends to the entropy of the
target distribution with vanishing reconstruction error in a proper limit.
Furthermore, we assess the performance of the dematcher on practical
Hadamard-based operators. A remarkable feature of our proposed solution is the
possibility to: i) perform matching at the symbol level (nonbinary); ii)
perform joint channel coding and matching. | [
0,
0,
0,
1,
0,
0
] | [
"Computer Science",
"Mathematics",
"Statistics"
] |
Title: Deviation from the dipole-ice model in the new spinel spin-ice candidate, MgEr$_2$Se$_4$,
Abstract: In spin ice research, small variations in structure or interactions drive a
multitude of different behaviors, yet the collection of known materials relies
heavily on the `227' pyrochlore structure. Here, we present thermodynamic,
structural and inelastic neutron scattering data on a new spin-ice material,
MgEr$_2$Se$_4$, which contributes to the relatively underexplored family of
rare-earth spinel chalcogenides. X-ray and neutron diffraction confirm a normal
spinel structure, and places Er$^{3+}$ moments on an ideal pyrochlore
sublattice. Measurement of crystal electric field excitations with neutron
inelastic scattering confirms that the moments have perfect Ising character,
and further identifies the ground state Kramers doublet as having
dipole-octupolar form with a significant multipolar character. Heat capacity
and magnetic neutron diffuse scattering have ice-like features, but are
inconsistent with Monte Carlo simulations of the nearest-neighbor and
next-nearest-neighbor dipolar spin-ice (DSI) models. A significant remnant
entropy is observed as T$\rightarrow$0 K, but again falls short of the full
Pauling expectation for DSI, unless significant disorder is added. We show that
these observations are fully in-line with what is recently reported for
CdEr$_2$Se$_4$, and point to the importance of quantum fluctuations in these
materials. | [
0,
1,
0,
0,
0,
0
] | [
"Physics"
] |
Title: Generating Nontrivial Melodies for Music as a Service,
Abstract: We present a hybrid neural network and rule-based system that generates pop
music. Music produced by pure rule-based systems often sounds mechanical. Music
produced by machine learning sounds better, but still lacks hierarchical
temporal structure. We restore temporal hierarchy by augmenting machine
learning with a temporal production grammar, which generates the music's
overall structure and chord progressions. A compatible melody is then generated
by a conditional variational recurrent autoencoder. The autoencoder is trained
with eight-measure segments from a corpus of 10,000 MIDI files, each of which
has had its melody track and chord progressions identified heuristically. The
autoencoder maps melody into a multi-dimensional feature space, conditioned by
the underlying chord progression. A melody is then generated by feeding a
random sample from that space to the autoencoder's decoder, along with the
chord progression generated by the grammar. The autoencoder can make musically
plausible variations on an existing melody, suitable for recurring motifs. It
can also reharmonize a melody to a new chord progression, keeping the rhythm
and contour. The generated music compares favorably with that generated by
other academic and commercial software designed for the music-as-a-service
industry. | [
1,
0,
0,
0,
0,
0
] | [
"Computer Science",
"Quantitative Biology"
] |
Title: Data-Driven Estimation of Travel Latency Cost Functions via Inverse Optimization in Multi-Class Transportation Networks,
Abstract: We develop a method to estimate from data travel latency cost functions in
multi-class transportation networks, which accommodate different types of
vehicles with very different characteristics (e.g., cars and trucks).
Leveraging our earlier work on inverse variational inequalities, we develop a
data-driven approach to estimate the travel latency cost functions. Extensive
numerical experiments using benchmark networks, ranging from moderate-sized to
large-sized, demonstrate the effectiveness and efficiency of our approach. | [
1,
0,
1,
0,
0,
0
] | [
"Computer Science",
"Mathematics",
"Statistics"
] |
Title: Guiding Chemical Synthesis: Computational Prediction of the Regioselectivity of CH Functionalization,
Abstract: We will develop a computational method (RegioSQM) for predicting the
regioselectivity of CH functionalization reactions that can be used by
synthetic chemists who are not experts in computational chemistry through a
simple web interface (regiosqm.org). CH functionalization, i.e. replacing the
hydrogen atom in a CH bond with another atom or molecule, is arguably single
most promising technique for increasing the efficiency of chemical synthesis,
but there are no generally applicable predictive tools that predict which CH
bond is most reactive. RegioSQM uses semiempirical quantum chemistry methods to
predict relative stabilities of reaction intermediates which correlates with
reaction rate and our goal is to determine which quantum method and
intermediate give the best result for each reaction type. Finally, we will
experimentally demonstrate how RegioSQM can be used to make the chemical
synthesis part of drug discovery more efficient. | [
0,
1,
0,
0,
0,
0
] | [
"Computer Science",
"Physics"
] |
Title: Potential-Function Proofs for First-Order Methods,
Abstract: This note discusses proofs for convergence of first-order methods based on
simple potential-function arguments. We cover methods like gradient descent
(for both smooth and non-smooth settings), mirror descent, and some accelerated
variants. | [
1,
0,
0,
0,
0,
0
] | [
"Computer Science",
"Mathematics"
] |
Title: Multidimensional $p$-adic continued fraction algorithms,
Abstract: We give a new class of multidimensional $p$-adic continued fraction
algorithms. We propose an algorithm in the class for which we can expect that
multidimensional $p$-adic version of Lagrange's Theorem holds. | [
0,
0,
1,
0,
0,
0
] | [
"Mathematics"
] |
Title: Expected Policy Gradients,
Abstract: We propose expected policy gradients (EPG), which unify stochastic policy
gradients (SPG) and deterministic policy gradients (DPG) for reinforcement
learning. Inspired by expected sarsa, EPG integrates across the action when
estimating the gradient, instead of relying only on the action in the sampled
trajectory. We establish a new general policy gradient theorem, of which the
stochastic and deterministic policy gradient theorems are special cases. We
also prove that EPG reduces the variance of the gradient estimates without
requiring deterministic policies and, for the Gaussian case, with no
computational overhead. Finally, we show that it is optimal in a certain sense
to explore with a Gaussian policy such that the covariance is proportional to
the exponential of the scaled Hessian of the critic with respect to the
actions. We present empirical results confirming that this new form of
exploration substantially outperforms DPG with the Ornstein-Uhlenbeck heuristic
in four challenging MuJoCo domains. | [
1,
0,
0,
1,
0,
0
] | [
"Computer Science",
"Statistics"
] |
Title: A new Hysteretic Nonlinear Energy Sink (HNES),
Abstract: The behavior of a new Hysteretic Nonlinear Energy Sink (HNES) coupled to a
linear primary oscillator is investigated in shock mitigation. Apart from a
small mass and a nonlinear elastic spring of the Duffing oscillator, the HNES
is also comprised of a purely hysteretic and a linear elastic spring of
potentially negative stiffness, connected in parallel. The Bouc-Wen model is
used to describe the force produced by both the purely hysteretic and linear
elastic springs. Coupling the primary oscillator with the HNES three nonlinear
equations of motion are derived, in terms of the two displacements and the
dimensionless hysteretic variable, which are integrated numerically using the
analog equation method. The performance of the HNES is examined by quantifying
the percentage of the initially induced energy in the primary system that is
passively transferred and dissipated by the HNES. Remarkable results are
achieved for a wide range of initial input energies. The great performance of
the HNES is mostly evidenced when the linear spring stiffness takes on negative
values. | [
0,
1,
0,
0,
0,
0
] | [
"Physics",
"Mathematics"
] |
Title: Measuring High-Energy Spectra with HAWC,
Abstract: The High-Altitude Water-Cherenkov (HAWC) experiment is a TeV $\gamma$-ray
observatory located \unit[4100]{m} above sea level on the Sierra Negra mountain
in Puebla, Mexico. The detector consists of 300 water-filled tanks, each
instrumented with 4 photomultiplier tubes that utilize the water-Cherenkov
technique to detect atmospheric air showers produced by cosmic $\gamma$ rays.
Construction of HAWC was completed in March of 2015. The experiment's wide
instantaneous field of view (\unit[2]{sr}) and high duty cycle (> 95\%) make it
a powerful survey instrument sensitive to pulsars, supernova remnants, and
other $\gamma$-ray sources. The mechanisms of particle acceleration at these
sources can be studied by analyzing their high-energy spectra. To this end, we
have developed an event-by-event energy-reconstruction algorithm using an
artificial neural network to estimate energies of primary $\gamma$ rays at
HAWC. We will present the details of this technique and its performance as well
as the current progress toward using it to measure energy spectra of
$\gamma$-ray sources. | [
0,
1,
0,
0,
0,
0
] | [
"Physics",
"Computer Science"
] |
Title: On the Three Properties of Stationary Populations and knotting with Non-Stationary Populations,
Abstract: We propose three properties that are related to the stationary population
identity (SPI) of population biology by connecting it with stationary
populations and non-stationary populations which are approaching stationarity.
These properties provide deeper insights into cohort formation in real-world
populations and the length of the duration for which stationary and
non-stationary conditions hold. The new concepts are based on the time gap
between the occurrence of stationary and non-stationary populations within the
SPI framework that we refer to as Oscillatory SPI and the Amplitude of SPI. | [
0,
0,
0,
0,
1,
0
] | [
"Quantitative Biology"
] |
Title: Generating and designing DNA with deep generative models,
Abstract: We propose generative neural network methods to generate DNA sequences and
tune them to have desired properties. We present three approaches: creating
synthetic DNA sequences using a generative adversarial network; a DNA-based
variant of the activation maximization ("deep dream") design method; and a
joint procedure which combines these two approaches together. We show that
these tools capture important structures of the data and, when applied to
designing probes for protein binding microarrays, allow us to generate new
sequences whose properties are estimated to be superior to those found in the
training data. We believe that these results open the door for applying deep
generative models to advance genomics research. | [
1,
0,
0,
1,
0,
0
] | [
"Computer Science",
"Quantitative Biology"
] |
Title: Asymptotic Theory for the Maximum of an Increasing Sequence of Parametric Functions,
Abstract: \cite{HillMotegi2017} present a new general asymptotic theory for the maximum
of a random array $\{\mathcal{X}_{n}(i)$ $:$ $1$ $\leq $ $i$ $\leq $
$\mathcal{L}\}_{n\geq 1}$, where each $\mathcal{X}_{n}(i)$ is assumed to
converge in probability as $n$ $\rightarrow $ $\infty $. The array dimension
$\mathcal{L}$ is allowed to increase with the sample size $n$. Existing extreme
value theory arguments focus on observed data $\mathcal{X}_{n}(i)$, and require
a well defined limit law for $\max_{1\leq i\leq
\mathcal{L}}|\mathcal{X}_{n}(i)|$ by restricting dependence across $i$. The
high dimensional central limit theory literature presumes approximability by a
Gaussian law, and also restricts attention to observed data.
\cite{HillMotegi2017} do not require $\max_{1\leq i\leq
\mathcal{L}_{n}}|\mathcal{X}_{n}(i)|$ to have a well defined limit nor be
approximable by a Gaussian random variable, and we do not make any assumptions
about dependence across $i$. We apply the theory to filtered data when the
variable of interest $\mathcal{X}_{n}(i,\theta _{0})$ is not observed, but its
sample counterpart $\mathcal{X}_{n}(i,\hat{\theta}_{n})$ is observed where
$\hat{\theta}_{n}$ estimates $\theta _{0}$. The main results are illustrated by
looking at unit root tests for a high dimensional random variable, and a
residuals white noise test. | [
0,
0,
1,
1,
0,
0
] | [
"Mathematics",
"Statistics"
] |
Title: Resilient Active Information Gathering with Mobile Robots,
Abstract: Applications of safety, security, and rescue in robotics, such as multi-robot
target tracking, involve the execution of information acquisition tasks by
teams of mobile robots. However, in failure-prone or adversarial environments,
robots get attacked, their communication channels get jammed, and their sensors
may fail, resulting in the withdrawal of robots from the collective task, and
consequently the inability of the remaining active robots to coordinate with
each other. As a result, traditional design paradigms become insufficient and,
in contrast, resilient designs against system-wide failures and attacks become
important. In general, resilient design problems are hard, and even though they
often involve objective functions that are monotone or submodular, scalable
approximation algorithms for their solution have been hitherto unknown. In this
paper, we provide the first algorithm, enabling the following capabilities:
minimal communication, i.e., the algorithm is executed by the robots based only
on minimal communication between them; system-wide resiliency, i.e., the
algorithm is valid for any number of denial-of-service attacks and failures;
and provable approximation performance, i.e., the algorithm ensures for all
monotone (and not necessarily submodular) objective functions a solution that
is finitely close to the optimal. We quantify our algorithm's approximation
performance using a notion of curvature for monotone set functions. We support
our theoretical analyses with simulated and real-world experiments, by
considering an active information gathering scenario, namely, multi-robot
target tracking. | [
1,
0,
0,
1,
0,
0
] | [
"Computer Science",
"Robotics"
] |
Title: Optical properties of a four-layer waveguiding nanocomposite structure in near-IR regime,
Abstract: The theoretical study of the optical properties of TE- and TM- modes in a
four-layer structure composed of the magneto-optical yttrium iron garnet
guiding layer on a dielectric substrate covered by planar nanocomposite guiding
multilayer is presented. The dispersion equation is obtained taking into
account the bigyrotropic properties of yttrium-iron garnet, and an original
algorithm for the guided modes identification is proposed. The dispersion
spectra are analyzed and the energy flux distributions across the structure are
calculated. The fourfold difference between the partial power fluxes within the
waveguide layers is achieved in the wavelength range of 200 nm. | [
0,
1,
0,
0,
0,
0
] | [
"Physics"
] |
Title: Source Forager: A Search Engine for Similar Source Code,
Abstract: Developers spend a significant amount of time searching for code: e.g., to
understand how to complete, correct, or adapt their own code for a new context.
Unfortunately, the state of the art in code search has not evolved much beyond
text search over tokenized source. Code has much richer structure and semantics
than normal text, and this property can be exploited to specialize the
code-search process for better querying, searching, and ranking of code-search
results.
We present a new code-search engine named Source Forager. Given a query in
the form of a C/C++ function, Source Forager searches a pre-populated code
database for similar C/C++ functions. Source Forager preprocesses the database
to extract a variety of simple code features that capture different aspects of
code. A search returns the $k$ functions in the database that are most similar
to the query, based on the various extracted code features.
We tested the usefulness of Source Forager using a variety of code-search
queries from two domains. Our experiments show that the ranked results returned
by Source Forager are accurate, and that query-relevant functions can be
reliably retrieved even when searching through a large code database that
contains very few query-relevant functions.
We believe that Source Forager is a first step towards much-needed tools that
provide a better code-search experience. | [
1,
0,
0,
0,
0,
0
] | [
"Computer Science"
] |
Title: A New Test of Multivariate Nonlinear Causality,
Abstract: The multivariate nonlinear Granger causality developed by Bai et al. (2010)
plays an important role in detecting the dynamic interrelationships between two
groups of variables. Following the idea of Hiemstra-Jones (HJ) test proposed by
Hiemstra and Jones (1994), they attempt to establish a central limit theorem
(CLT) of their test statistic by applying the asymptotical property of
multivariate $U$-statistic. However, Bai et al. (2016) revisit the HJ test and
find that the test statistic given by HJ is NOT a function of $U$-statistics
which implies that the CLT neither proposed by Hiemstra and Jones (1994) nor
the one extended by Bai et al. (2010) is valid for statistical inference. In
this paper, we re-estimate the probabilities and reestablish the CLT of the new
test statistic. Numerical simulation shows that our new estimates are
consistent and our new test performs decent size and power. | [
0,
0,
0,
1,
0,
0
] | [
"Statistics",
"Mathematics"
] |
Title: Nonlinear dynamics of polar regions in paraelectric phase of (Ba1-x,Srx)TiO3 ceramics,
Abstract: The dynamic dielectric nonlinearity of barium strontium titanate
(Ba1-x,Srx)TiO3 ceramics is investigated in their paraelectric phase. With the
goal to contribute to the identification of the mechanisms that govern the
dielectric nonlinearity in this family, we analyze the amplitude and the phase
angles of the first and the third harmonics of polarization. Our study shows
that an interpretation of the field-dependent polarization in paraelectric
(Ba1-x,Srx)TiO3 ceramics in terms of the Rayleigh-type dynamics is inadequate
for our samples and that their nonlinear response rather resembles that
observed in canonical relaxor Pb(Mg1/3Nb2/3)O3. | [
0,
1,
0,
0,
0,
0
] | [
"Physics"
] |
Title: Finite Sample Analysis of Two-Timescale Stochastic Approximation with Applications to Reinforcement Learning,
Abstract: Two-timescale Stochastic Approximation (SA) algorithms are widely used in
Reinforcement Learning (RL). Their iterates have two parts that are updated
using distinct stepsizes. In this work, we develop a novel recipe for their
finite sample analysis. Using this, we provide a concentration bound, which is
the first such result for a two-timescale SA. The type of bound we obtain is
known as `lock-in probability'. We also introduce a new projection scheme, in
which the time between successive projections increases exponentially. This
scheme allows one to elegantly transform a lock-in probability into a
convergence rate result for projected two-timescale SA. From this latter
result, we then extract key insights on stepsize selection. As an application,
we finally obtain convergence rates for the projected two-timescale RL
algorithms GTD(0), GTD2, and TDC. | [
1,
0,
0,
0,
0,
0
] | [
"Computer Science",
"Mathematics",
"Statistics"
] |
Title: Existence and uniqueness of solutions to Y-systems and TBA equations,
Abstract: We consider Y-system functional equations of the form $$
Y_n(x+i)Y_n(x-i)=\prod_{m=1}^N (1+Y_m(x))^{G_{nm}}$$ and the corresponding
nonlinear integral equations of the Thermodynamic Bethe Ansatz. We prove an
existence and uniqueness result for solutions of these equations, subject to
appropriate conditions on the analytical properties of the $Y_n$, in particular
the absence of zeros in a strip around the real axis. The matrix $G_{nm}$ must
have non-negative real entries, and be irreducible and diagonalisable over
$\mathbb{R}$ with spectral radius less than 2. This includes the adjacency
matrices of finite Dynkin diagrams, but covers much more as we do not require
$G_{nm}$ to be integers. Our results specialise to the constant Y-system,
proving existence and uniqueness of a strictly positive solution in that case. | [
0,
1,
0,
0,
0,
0
] | [
"Mathematics",
"Physics"
] |
Title: Normalization of Neural Networks using Analytic Variance Propagation,
Abstract: We address the problem of estimating statistics of hidden units in a neural
network using a method of analytic moment propagation. These statistics are
useful for approximate whitening of the inputs in front of saturating
non-linearities such as a sigmoid function. This is important for
initialization of training and for reducing the accumulated scale and bias
dependencies (compensating covariate shift), which presumably eases the
learning. In batch normalization, which is currently a very widely applied
technique, sample estimates of statistics of hidden units over a batch are
used. The proposed estimation uses an analytic propagation of mean and variance
of the training set through the network. The result depends on the network
structure and its current weights but not on the specific batch input. The
estimates are suitable for initialization and normalization, efficient to
compute and independent of the batch size. The experimental verification well
supports these claims. However, the method does not share the generalization
properties of BN, to which our experiments give some additional insight. | [
0,
0,
0,
1,
0,
0
] | [
"Computer Science",
"Statistics"
] |
Title: Delta sets for symmetric numerical semigroups with embedding dimension three,
Abstract: This work extends the results known for the Delta sets of non-symmetric
numerical semigroups with embedding dimension three to the symmetric case.
Thus, we have a fast algorithm to compute the Delta set of any embedding
dimension three numerical semigroup. Also, as a consequence of these resutls,
the sets that can be realized as Delta sets of numerical semigroups of
embedding dimension three are fully characterized. | [
0,
0,
1,
0,
0,
0
] | [
"Mathematics"
] |
Title: Unsure When to Stop? Ask Your Semantic Neighbors,
Abstract: In iterative supervised learning algorithms it is common to reach a point in
the search where no further induction seems to be possible with the available
data. If the search is continued beyond this point, the risk of overfitting
increases significantly. Following the recent developments in inductive
semantic stochastic methods, this paper studies the feasibility of using
information gathered from the semantic neighborhood to decide when to stop the
search. Two semantic stopping criteria are proposed and experimentally assessed
in Geometric Semantic Genetic Programming (GSGP) and in the Semantic Learning
Machine (SLM) algorithm (the equivalent algorithm for neural networks). The
experiments are performed on real-world high-dimensional regression datasets.
The results show that the proposed semantic stopping criteria are able to
detect stopping points that result in a competitive generalization for both
GSGP and SLM. This approach also yields computationally efficient algorithms as
it allows the evolution of neural networks in less than 3 seconds on average,
and of GP trees in at most 10 seconds. The usage of the proposed semantic
stopping criteria in conjunction with the computation of optimal
mutation/learning steps also results in small trees and neural networks. | [
1,
0,
0,
1,
0,
0
] | [
"Computer Science",
"Statistics"
] |
Title: Warming trend in cold season of the Yangtze River Delta and its correlation with Siberian high,
Abstract: Based on the meteorological data from 1960 to 2010, we investigated the
temperature variation in the Yangtze River Delta (YRD) by using Mann-Kendall
nonparametric test and explored the correlation between the temperature in the
cold season and the Siberian high intensity (SHI) by using correlation analysis
method. The main results are that (a) the temperature in YRD increased
remarkably during the study period, (b) the warming trend in the cold season
made the higher contribution to annual warming, and (c) there was a significant
negative correlation between the temperature in the cold season and the SHI. | [
0,
0,
0,
1,
0,
0
] | [
"Physics",
"Statistics"
] |
Title: Modeling and Quantifying the Forces Driving Online Video Popularity Evolution,
Abstract: Video popularity is an essential reference for optimizing resource allocation
and video recommendation in online video services. However, there is still no
convincing model that can accurately depict a video's popularity evolution. In
this paper, we propose a dynamic popularity model by modeling the video
information diffusion process driven by various forms of recommendation.
Through fitting the model with real traces collected from a practical system,
we can quantify the strengths of the recommendation forces. Such quantification
can lead to characterizing video popularity patterns, user behaviors and
recommendation strategies, which is illustrated by a case study of TV episodes. | [
1,
0,
0,
0,
0,
0
] | [
"Computer Science",
"Statistics"
] |
Title: Measurement of the Lorentz-FitzGerald Body Contraction,
Abstract: A complete foundational discussion of acceleration in context of Special
Relativity is presented. Acceleration allows the measurement of a
Lorentz-FitzGerald body contraction created. It is argued that in the back
scattering of a probing laser beam from a relativistic flying electron cloud
mirror generated by an ultra-intense laser pulse, a first measurement of a
Lorentz-FitzGerald body contraction is feasible. | [
0,
1,
0,
0,
0,
0
] | [
"Physics"
] |
Title: Information Directed Sampling for Stochastic Bandits with Graph Feedback,
Abstract: We consider stochastic multi-armed bandit problems with graph feedback, where
the decision maker is allowed to observe the neighboring actions of the chosen
action. We allow the graph structure to vary with time and consider both
deterministic and Erdős-Rényi random graph models. For such a graph
feedback model, we first present a novel analysis of Thompson sampling that
leads to tighter performance bound than existing work. Next, we propose new
Information Directed Sampling based policies that are graph-aware in their
decision making. Under the deterministic graph case, we establish a Bayesian
regret bound for the proposed policies that scales with the clique cover number
of the graph instead of the number of actions. Under the random graph case, we
provide a Bayesian regret bound for the proposed policies that scales with the
ratio of the number of actions over the expected number of observations per
iteration. To the best of our knowledge, this is the first analytical result
for stochastic bandits with random graph feedback. Finally, using numerical
evaluations, we demonstrate that our proposed IDS policies outperform existing
approaches, including adaptions of upper confidence bound, $\epsilon$-greedy
and Exp3 algorithms. | [
1,
0,
0,
1,
0,
0
] | [
"Computer Science",
"Statistics"
] |
Title: Hausdorff operators on holomorphic Hardy spaces and applications,
Abstract: The aim of this paper is to characterize the nonnegative functions $\varphi$
defined on $(0,\infty)$ for which the Hausdorff operator
$$\mathscr H_\varphi f(z)= \int_0^\infty
f\left(\frac{z}{t}\right)\frac{\varphi(t)}{t}dt$$ is bounded on the Hardy
spaces of the upper half-plane $\mathcal H_a^p(\mathbb C_+)$, $p\in[1,\infty]$.
The corresponding operator norms and their applications are also given. | [
0,
0,
1,
0,
0,
0
] | [
"Mathematics"
] |
Title: Does Your Phone Know Your Touch?,
Abstract: This paper explores supervised techniques for continuous anomaly detection
from biometric touch screen data. A capacitive sensor array used to mimic a
touch screen as used to collect touch and swipe gestures from participants. The
gestures are recorded over fixed segments of time, with position and force
measured for each gesture. Support Vector Machine, Logistic Regression, and
Gaussian mixture models were tested to learn individual touch patterns. Test
results showed true negative and true positive scores of over 95% accuracy for
all gesture types, with logistic regression models far outperforming the other
methods. A more expansive and varied data collection over longer periods of
time is needed to determine pragmatic usage of these results. | [
0,
0,
0,
1,
0,
0
] | [
"Computer Science",
"Statistics"
] |
Title: Nucleus: A Pilot Project,
Abstract: Early in 2016, an environmental scan was conducted by the Research Library
Data Working Group for three purposes:
1.) Perform a survey of the data management landscape at Los Alamos National
Laboratory in order to identify local gaps in data management services.
2.) Conduct an environmental scan of external institutions to benchmark
budgets, infrastructure, and personnel dedicated to data management.
3.) Draft a research data infrastructure model that aligns with the current
workflow and classification restrictions at Los Alamos National Laboratory.
This report is a summary of those activities and the draft for a pilot data
management project. | [
1,
0,
0,
0,
0,
0
] | [
"Computer Science",
"Statistics"
] |
Title: Non Volatile MoS$_{2}$ Field Effect Transistors Directly Gated By Single Crystalline Epitaxial Ferroelectric,
Abstract: We demonstrate non-volatile, n-type, back-gated, MoS$_{2}$ transistors,
placed directly on an epitaxial grown, single crystalline,
PbZr$_{0.2}$Ti$_{0.8}$O$_{3}$ (PZT) ferroelectric. The transistors show decent
ON current (19 ${\mu}A/{\mu}$m), high on-off ratio (10$^{7}$), and a
subthreshold swing of (SS ~ 92 mV/dec) with a 100 nm thick PZT layer as the
back gate oxide. Importantly, the ferroelectric polarization can directly
control the channel charge, showing a clear anti-clockwise hysteresis. We have
selfconsistently confirmed the switching of the ferroelectric and corresponding
change in channel current from a direct time-dependent measurement. Our results
demonstrate that it is possible to obtain transistor operation directly on
polar surfaces and therefore it should be possible to integrate 2D electronics
with single crystalline functional oxides. | [
0,
1,
0,
0,
0,
0
] | [
"Physics"
] |
Title: Integral Equations and Machine Learning,
Abstract: As both light transport simulation and reinforcement learning are ruled by
the same Fredholm integral equation of the second kind, reinforcement learning
techniques may be used for photorealistic image synthesis: Efficiency may be
dramatically improved by guiding light transport paths by an approximate
solution of the integral equation that is learned during rendering. In the
light of the recent advances in reinforcement learning for playing games, we
investigate the representation of an approximate solution of an integral
equation by artificial neural networks and derive a loss function for that
purpose. The resulting Monte Carlo and quasi-Monte Carlo methods train neural
networks with standard information instead of linear information and naturally
are able to generate an arbitrary number of training samples. The methods are
demonstrated for applications in light transport simulation. | [
1,
0,
0,
0,
0,
0
] | [
"Computer Science",
"Mathematics"
] |
Title: Experiments on bright field and dark field high energy electron imaging with thick target material,
Abstract: Using a high energy electron beam for the imaging of high density matter with
both high spatial-temporal and areal density resolution under extreme states of
temperature and pressure is one of the critical challenges in high energy
density physics . When a charged particle beam passes through an opaque target,
the beam will be scattered with a distribution that depends on the thickness of
the material. By collecting the scattered beam either near or off axis,
so-called bright field or dark field images can be obtained. Here we report on
an electron radiography experiment using 45 MeV electrons from an S-band
photo-injector, where scattered electrons, after interacting with a sample, are
collected and imaged by a quadrupole imaging system. We achieved a few
micrometers (about 4 micrometers) spatial resolution and about 10 micrometers
thickness resolution for a silicon target of 300-600 micron thickness. With
addition of dark field images that are captured by selecting electrons with
large scattering angle, we show that more useful information in determining
external details such as outlines, boundaries and defects can be obtained. | [
0,
1,
0,
0,
0,
0
] | [
"Physics"
] |
Title: SOLAR: Deep Structured Latent Representations for Model-Based Reinforcement Learning,
Abstract: Model-based reinforcement learning (RL) methods can be broadly categorized as
global model methods, which depend on learning models that provide sensible
predictions in a wide range of states, or local model methods, which
iteratively refit simple models that are used for policy improvement. While
predicting future states that will result from the current actions is
difficult, local model methods only attempt to understand system dynamics in
the neighborhood of the current policy, making it possible to produce local
improvements without ever learning to predict accurately far into the future.
The main idea in this paper is that we can learn representations that make it
easy to retrospectively infer simple dynamics given the data from the current
policy, thus enabling local models to be used for policy learning in complex
systems. To that end, we focus on learning representations with probabilistic
graphical model (PGM) structure, which allows us to devise an efficient local
model method that infers dynamics from real-world rollouts with the PGM as a
global prior. We compare our method to other model-based and model-free RL
methods on a suite of robotics tasks, including manipulation tasks on a real
Sawyer robotic arm directly from camera images. Videos of our results are
available at this https URL | [
1,
0,
0,
1,
0,
0
] | [
"Computer Science",
"Statistics"
] |
Title: Robust and Fast Decoding of High-Capacity Color QR Codes for Mobile Applications,
Abstract: The use of color in QR codes brings extra data capacity, but also inflicts
tremendous challenges on the decoding process due to chromatic distortion,
cross-channel color interference and illumination variation. Particularly, we
further discover a new type of chromatic distortion in high-density color QR
codes, cross-module color interference, caused by the high density which also
makes the geometric distortion correction more challenging. To address these
problems, we propose two approaches, namely, LSVM-CMI and QDA-CMI, which
jointly model these different types of chromatic distortion. Extended from SVM
and QDA, respectively, both LSVM-CMI and QDA-CMI optimize over a particular
objective function to learn a color classifier. Furthermore, a robust geometric
transformation method and several pipeline refinements are proposed to boost
the decoding performance for mobile applications. We put forth and implement a
framework for high-capacity color QR codes equipped with our methods, called
HiQ. To evaluate the performance of HiQ, we collect a challenging large-scale
color QR code dataset, CUHK-CQRC, which consists of 5390 high-density color QR
code samples. The comparison with the baseline method [2] on CUHK-CQRC shows
that HiQ at least outperforms [2] by 188% in decoding success rate and 60% in
bit error rate. Our implementation of HiQ in iOS and Android also demonstrates
the effectiveness of our framework in real-world applications. | [
1,
0,
0,
0,
0,
0
] | [
"Computer Science"
] |
Title: The quest for H$_3^+$ at Neptune: deep burn observations with NASA IRTF iSHELL,
Abstract: Emission from the molecular ion H$_3^+$ is a powerful diagnostic of the upper
atmosphere of Jupiter, Saturn, and Uranus, but it remains undetected at
Neptune. In search of this emission, we present near-infrared spectral
observations of Neptune between 3.93 and 4.00 $\mu$m taken with the newly
commissioned iSHELL instrument on the NASA Infrared Telescope Facility in
Hawaii, obtained 17-20 August 2017. We spent 15.4 h integrating across the disk
of the planet, yet were unable to unambiguously identify any H$_3^+$ line
emissions. Assuming a temperature of 550 K, we derive an upper limit on the
column integrated density of $1.0^{+1.2}_{-0.8}\times10^{13}$ m$^{-2}$, which
is an improvement of 30\% on the best previous observational constraint. This
result means that models are over-estimating the density by at least a factor
of 5, highlighting the need for renewed modelling efforts. A potential solution
is strong vertical mixing of polyatomic neutral species from Neptune's upper
stratosphere to the thermosphere, reacting with H$_3^+$, thus greatly reducing
the column integrated H$_3^+$ densities. This upper limit also provide
constraints on future attempts at detecting H$_3^+$ using the James Webb Space
Telescope. | [
0,
1,
0,
0,
0,
0
] | [
"Physics"
] |
Title: Redistributing Funds across Charitable Crowdfunding Campaigns,
Abstract: On Kickstarter only 36% of crowdfunding campaigns successfully raise
sufficient funds for their projects. In this paper, we explore the possibility
of redistribution of crowdfunding donations to increase the chances of success.
We define several intuitive redistribution policies and, using data from a real
crowdfunding platform, LaunchGood, we assess the potential improvement in
campaign fundraising success rates. We find that an aggressive redistribution
scheme can boost campaign success rates from 37% to 79%, but such
choice-agnostic redistribution schemes come at the cost of disregarding donor
preferences. Taking inspiration from offline giving societies and donor clubs,
we build a case for choice preserving redistribution schemes that strike a
balance between increasing the number of successful campaigns and respecting
giving preference. We find that choice-preserving redistribution can easily
achieve campaign success rates of 48%. Finally, we discuss the implications of
these different redistribution schemes for the various stakeholders in the
crowdfunding ecosystem. | [
1,
0,
0,
0,
0,
0
] | [
"Quantitative Finance",
"Statistics"
] |
Title: Far-HO: A Bilevel Programming Package for Hyperparameter Optimization and Meta-Learning,
Abstract: In (Franceschi et al., 2018) we proposed a unified mathematical framework,
grounded on bilevel programming, that encompasses gradient-based hyperparameter
optimization and meta-learning. We formulated an approximate version of the
problem where the inner objective is solved iteratively, and gave sufficient
conditions ensuring convergence to the exact problem. In this work we show how
to optimize learning rates, automatically weight the loss of single examples
and learn hyper-representations with Far-HO, a software package based on the
popular deep learning framework TensorFlow that allows to seamlessly tackle
both HO and ML problems. | [
0,
0,
0,
1,
0,
0
] | [
"Computer Science",
"Mathematics"
] |
Title: Femtosecond mega-electron-volt electron microdiffraction,
Abstract: Instruments to visualize transient structural changes of inhomogeneous
materials on the nanometer scale with atomic spatial and temporal resolution
are demanded to advance materials science, bioscience, and fusion sciences. One
such technique is femtosecond electron microdiffraction, in which a short pulse
of electrons with femtosecond-scale duration is focused into a micron-scale
spot and used to obtain diffraction images to resolve ultrafast structural
dynamics over localized crystalline domain. In this letter, we report the
experimental demonstration of time-resolved mega-electron-volt electron
microdiffraction which achieves a 5 {\mu}m root-mean-square (rms) beam size on
the sample and a 100 fs rms temporal resolution. Using pulses of 10k electrons
at 4.2 MeV energy with a normalized emittance 3 nm-rad, we obtained high
quality diffraction from a single 10 {\mu}m paraffin (C_44 H_90) crystal. The
phonon softening mode in optical-pumped polycrystalline Bi was also
time-resolved, demonstrating the temporal resolution limits of our instrument
design. This new characterization capability will open many research
opportunities in material and biological sciences. | [
0,
1,
0,
0,
0,
0
] | [
"Physics"
] |
Title: Deep Recurrent Neural Network for Protein Function Prediction from Sequence,
Abstract: As high-throughput biological sequencing becomes faster and cheaper, the need
to extract useful information from sequencing becomes ever more paramount,
often limited by low-throughput experimental characterizations. For proteins,
accurate prediction of their functions directly from their primary amino-acid
sequences has been a long standing challenge. Here, machine learning using
artificial recurrent neural networks (RNN) was applied towards classification
of protein function directly from primary sequence without sequence alignment,
heuristic scoring or feature engineering. The RNN models containing
long-short-term-memory (LSTM) units trained on public, annotated datasets from
UniProt achieved high performance for in-class prediction of four important
protein functions tested, particularly compared to other machine learning
algorithms using sequence-derived protein features. RNN models were used also
for out-of-class predictions of phylogenetically distinct protein families with
similar functions, including proteins of the CRISPR-associated nuclease,
ferritin-like iron storage and cytochrome P450 families. Applying the trained
RNN models on the partially unannotated UniRef100 database predicted not only
candidates validated by existing annotations but also currently unannotated
sequences. Some RNN predictions for the ferritin-like iron sequestering
function were experimentally validated, even though their sequences differ
significantly from known, characterized proteins and from each other and cannot
be easily predicted using popular bioinformatics methods. As sequencing and
experimental characterization data increases rapidly, the machine-learning
approach based on RNN could be useful for discovery and prediction of
homologues for a wide range of protein functions. | [
1,
0,
0,
1,
0,
0
] | [
"Computer Science",
"Quantitative Biology"
] |
Title: CubemapSLAM: A Piecewise-Pinhole Monocular Fisheye SLAM System,
Abstract: We present a real-time feature-based SLAM (Simultaneous Localization and
Mapping) system for fisheye cameras featured by a large field-of-view (FoV).
Large FoV cameras are beneficial for large-scale outdoor SLAM applications,
because they increase visual overlap between consecutive frames and capture
more pixels belonging to the static parts of the environment. However, current
feature-based SLAM systems such as PTAM and ORB-SLAM limit their camera model
to pinhole only. To compensate for the vacancy, we propose a novel SLAM system
with the cubemap model that utilizes the full FoV without introducing
distortion from the fisheye lens, which greatly benefits the feature matching
pipeline. In the initialization and point triangulation stages, we adopt a
unified vector-based representation to efficiently handle matches across
multiple faces, and based on this representation we propose and analyze a novel
inlier checking metric. In the optimization stage, we design and test a novel
multi-pinhole reprojection error metric that outperforms other metrics by a
large margin. We evaluate our system comprehensively on a public dataset as
well as a self-collected dataset that contains real-world challenging
sequences. The results suggest that our system is more robust and accurate than
other feature-based fisheye SLAM approaches. The CubemapSLAM system has been
released into the public domain. | [
1,
0,
0,
0,
0,
0
] | [
"Computer Science"
] |
Title: $J$-holomorphic disks with pre-Lagrangian boundary conditions,
Abstract: The purpose of this paper is to carry out a classical construction of a
non-constant holomorphic disk with boundary on (the suspension of) a Lagrangian
submanifold in $\mathbb{R}^{2 n}$ in the case the Lagrangian is the lift of a
coisotropic (a.k.a. pre-Lagrangian) submanifold in (a subset $U$ of)
$\mathbb{R}^{2 n - 1}$. We show that the positive lower and finite upper bounds
for the area of such a disk (which are due to M. Gromov and J.-C. Sikorav and
F. Laudenbach-Sikorav for general Lagrangians) depend on the coisotropic
submanifold only but not on its lift to the symplectization. The main
application is to a $C^0$-characterization of contact embeddings in terms of
coisotropic embeddings in another paper by the present author. Moreover, we
prove a version of Gromov's non-existence of exact Lagrangian embeddings into
standard $\mathbb{R}^{2 n}$ for coisotropic embeddings into $S^1 \times
\mathbb{R}^{2 n}$. This allows us to distinguish different contact structures
on the latter by means of the (modified) contact shape invariant. As in the
general Lagrangian case, all of the existence results are based on Gromov's
theory of $J$-holomorphic curves and his compactness theorem (or persistence
principle). Analytical difficulties arise mainly at the ends of the cone
$\mathbb{R}_+ \times U$. | [
0,
0,
1,
0,
0,
0
] | [
"Mathematics",
"Physics"
] |
Title: Neutron activation and prompt gamma intensity in Ar/CO$_{2}$-filled neutron detectors at the European Spallation Source,
Abstract: Monte Carlo simulations using MCNP6.1 were performed to study the effect of
neutron activation in Ar/CO$_{2}$ neutron detector counting gas. A general MCNP
model was built and validated with simple analytical calculations. Simulations
and calculations agree that only the $^{40}$Ar activation can have a
considerable effect. It was shown that neither the prompt gamma intensity from
the $^{40}$Ar neutron capture nor the produced $^{41}$Ar activity have an
impact in terms of gamma dose rate around the detector and background level. | [
0,
1,
0,
0,
0,
0
] | [
"Physics"
] |
Title: Solving 1ODEs with functions,
Abstract: Here we present a new approach to deal with first order ordinary differential
equations (1ODEs), presenting functions. This method is an alternative to the
one we have presented in [1]. In [2], we have establish the theoretical
background to deal, in the extended Prelle-Singer approach context, with
systems of 1ODEs. In this present paper, we will apply these results in order
to produce a method that is more efficient in a great number of cases.
Directly, the solving of 1ODEs is applicable to any problem presenting
parameters to which the rate of change is related to the parameter itself.
Apart from that, the solving of 1ODEs can be a part of larger mathematical
processes vital to dealing with many problems. | [
0,
1,
1,
0,
0,
0
] | [
"Mathematics"
] |
Title: Strong Consistency of Spectral Clustering for Stochastic Block Models,
Abstract: In this paper we prove the strong consistency of several methods based on the
spectral clustering techniques that are widely used to study the community
detection problem in stochastic block models (SBMs). We show that under some
weak conditions on the minimal degree, the number of communities, and the
eigenvalues of the probability block matrix, the K-means algorithm applied to
the eigenvectors of the graph Laplacian associated with its first few largest
eigenvalues can classify all individuals into the true community uniformly
correctly almost surely. Extensions to both regularized spectral clustering and
degree-corrected SBMs are also considered. We illustrate the performance of
different methods on simulated networks. | [
0,
0,
0,
1,
0,
0
] | [
"Statistics",
"Mathematics",
"Computer Science"
] |
Title: Machine learning out-of-equilibrium phases of matter,
Abstract: Neural network based machine learning is emerging as a powerful tool for
obtaining phase diagrams when traditional regression schemes using local
equilibrium order parameters are not available, as in many-body localized or
topological phases. Nevertheless, instances of machine learning offering new
insights have been rare up to now. Here we show that a single feed-forward
neural network can decode the defining structures of two distinct MBL phases
and a thermalizing phase, using entanglement spectra obtained from individual
eigenstates. For this, we introduce a simplicial geometry based method for
extracting multi-partite phase boundaries. We find that this method outperforms
conventional metrics (like the entanglement entropy) for identifying MBL phase
transitions, revealing a sharper phase boundary and shedding new insight into
the topology of the phase diagram. Furthermore, the phase diagram we acquire
from a single disorder configuration confirms that the machine-learning based
approach we establish here can enable speedy exploration of large phase spaces
that can assist with the discovery of new MBL phases. To our knowledge this
work represents the first example of a machine learning approach revealing new
information beyond conventional knowledge. | [
0,
1,
0,
0,
0,
0
] | [
"Physics",
"Computer Science"
] |
Title: Maximal solutions for the Infinity-eigenvalue problem,
Abstract: In this article we prove that the first eigenvalue of the $\infty-$Laplacian
$$ \left\{ \begin{array}{rclcl}
\min\{ -\Delta_\infty v,\, |\nabla v|-\lambda_{1, \infty}(\Omega) v \} & = &
0 & \text{in} & \Omega v & = & 0 & \text{on} & \partial \Omega, \end{array}
\right. $$ has a unique (up to scalar multiplication) maximal solution. This
maximal solution can be obtained as the limit as $\ell \nearrow 1$ of concave
problems of the form $$ \left\{ \begin{array}{rclcl}
\min\{ -\Delta_\infty v_{\ell},\, |\nabla v_{\ell}|-\lambda_{1,
\infty}(\Omega) v_{\ell}^{\ell} \} & = & 0 & \text{in} & \Omega v_{\ell} & = &
0 & \text{on} & \partial \Omega. \end{array} \right. $$ In this way we obtain
that the maximal eigenfunction is the unique one that is the limit of the
concave problems as happens for the usual eigenvalue problem for the
$p-$Laplacian for a fixed $1<p<\infty$. | [
0,
0,
1,
0,
0,
0
] | [
"Mathematics"
] |
Title: BL-MNE: Emerging Heterogeneous Social Network Embedding through Broad Learning with Aligned Autoencoder,
Abstract: Network embedding aims at projecting the network data into a low-dimensional
feature space, where the nodes are represented as a unique feature vector and
network structure can be effectively preserved. In recent years, more and more
online application service sites can be represented as massive and complex
networks, which are extremely challenging for traditional machine learning
algorithms to deal with. Effective embedding of the complex network data into
low-dimension feature representation can both save data storage space and
enable traditional machine learning algorithms applicable to handle the network
data. Network embedding performance will degrade greatly if the networks are of
a sparse structure, like the emerging networks with few connections. In this
paper, we propose to learn the embedding representation for a target emerging
network based on the broad learning setting, where the emerging network is
aligned with other external mature networks at the same time. To solve the
problem, a new embedding framework, namely "Deep alIgned autoencoder based
eMbEdding" (DIME), is introduced in this paper. DIME handles the diverse link
and attribute in a unified analytic based on broad learning, and introduces the
multiple aligned attributed heterogeneous social network concept to model the
network structure. A set of meta paths are introduced in the paper, which
define various kinds of connections among users via the heterogeneous link and
attribute information. The closeness among users in the networks are defined as
the meta proximity scores, which will be fed into DIME to learn the embedding
vectors of users in the emerging network. Extensive experiments have been done
on real-world aligned social networks, which have demonstrated the
effectiveness of DIME in learning the emerging network embedding vectors. | [
1,
0,
0,
0,
0,
0
] | [
"Computer Science"
] |
Title: Multi-channel discourse as an indicator for Bitcoin price and volume movements,
Abstract: This research aims to identify how Bitcoin-related news publications and
online discourse are expressed in Bitcoin exchange movements of price and
volume. Being inherently digital, all Bitcoin-related fundamental data (from
exchanges, as well as transactional data directly from the blockchain) is
available online, something that is not true for traditional businesses or
currencies traded on exchanges. This makes Bitcoin an interesting subject for
such research, as it enables the mapping of sentiment to fundamental events
that might otherwise be inaccessible. Furthermore, Bitcoin discussion largely
takes place on online forums and chat channels. In stock trading, the value of
sentiment data in trading decisions has been demonstrated numerous times [1]
[2] [3], and this research aims to determine whether there is value in such
data for Bitcoin trading models. To achieve this, data over the year 2015 has
been collected from Bitcointalk.org, (the biggest Bitcoin forum in post
volume), established news sources such as Bloomberg and the Wall Street
Journal, the complete /r/btc and /r/Bitcoin subreddits, and the bitcoin-otc and
bitcoin-dev IRC channels. By analyzing this data on sentiment and volume, we
find weak to moderate correlations between forum, news, and Reddit sentiment
and movements in price and volume from 1 to 5 days after the sentiment was
expressed. A Granger causality test confirms the predictive causality of the
sentiment on the daily percentage price and volume movements, and at the same
time underscores the predictive causality of market movements on sentiment
expressions in online communities | [
0,
0,
0,
0,
0,
1
] | [
"Quantitative Finance",
"Statistics"
] |
Title: Fano Resonances in a Photonic Crystal Covered with a Perforated Gold Film and its Application to Biosensing,
Abstract: Optical properties of the photonic crystal covered with a perforated metal
film were investigated and the existence of the Fano-type resonances was shown.
The Fano resonances originate from the interaction between the optical Tamm
state and the waveguide modes of the photonic crystal. It manifests itself as a
narrow dip in a broad peak in the transmission spectrum related to the optical
Tamm state. The design of a sensor based on this Fano resonance that is
sensitive to the change of the environment refractive index is suggested. | [
0,
1,
0,
0,
0,
0
] | [
"Physics",
"Quantitative Biology"
] |
Title: Non-Stationary Bandits with Habituation and Recovery Dynamics,
Abstract: Many settings involve sequential decision-making where a set of actions can
be chosen at each time step, each action provides a stochastic reward, and the
distribution for the reward of each action is initially unknown. However,
frequent selection of a specific action may reduce its expected reward, while
abstaining from choosing an action may cause its expected reward to increase.
Such non-stationary phenomena are observed in many real world settings such as
personalized healthcare-adherence improving interventions and targeted online
advertising. Though finding an optimal policy for general models with
non-stationarity is PSPACE-complete, we propose and analyze a new class of
models called ROGUE (Reducing or Gaining Unknown Efficacy) bandits, which we
show in this paper can capture these phenomena and are amenable to the design
of effective policies. We first present a consistent maximum likelihood
estimator for the parameters of these models. Next, we construct finite sample
concentration bounds that lead to an upper confidence bound policy called the
ROGUE Upper Confidence Bound (ROGUE-UCB) algorithm. We prove that under proper
conditions the ROGUE-UCB algorithm achieves logarithmic in time regret, unlike
existing algorithms which result in linear regret. We conclude with a numerical
experiment using real data from a personalized healthcare-adherence improving
intervention to increase physical activity. In this intervention, the goal is
to optimize the selection of messages (e.g., confidence increasing vs.
knowledge increasing) to send to each individual each day to increase adherence
and physical activity. Our results show that ROGUE-UCB performs better in terms
of regret and average reward as compared to state of the art algorithms, and
the use of ROGUE-UCB increases daily step counts by roughly 1,000 steps a day
(about a half-mile more of walking) as compared to other algorithms. | [
1,
0,
1,
0,
0,
0
] | [
"Computer Science",
"Statistics"
] |
Title: An Approach to Controller Design Based on the Generalized Cloud Model,
Abstract: In this paper, an approach to controller design based on the cloud models,
without using the analog plant model is presented. | [
1,
0,
0,
0,
0,
0
] | [
"Computer Science",
"Mathematics"
] |
Subsets and Splits