text
stringlengths 57
2.88k
| labels
sequencelengths 6
6
|
---|---|
Title: Looping and Clustering model for the organization of protein-DNA complexes on the bacterial genome,
Abstract: The bacterial genome is organized in a structure called the nucleoid by a
variety of associated proteins. These proteins can form complexes on DNA that
play a central role in various biological processes, including chromosome
segregation. A prominent example is the large ParB-DNA complex, which forms an
essential component of the segregation machinery in many bacteria. ChIP-Seq
experiments show that ParB proteins localize around centromere-like parS sites
on the DNA to which ParB binds specifically, and spreads from there over large
sections of the chromosome. Recent theoretical and experimental studies suggest
that DNA-bound ParB proteins can interact with each other to condense into a
coherent 3D complex on the DNA. However, the structural organization of this
protein-DNA complex remains unclear, and a predictive quantitative theory for
the distribution of ParB proteins on DNA is lacking. Here, we propose the
Looping and Clustering (LC) model, which employs a statistical physics approach
to describe protein-DNA complexes. The LC model accounts for the extrusion of
DNA loops from a cluster of interacting DNA-bound proteins. Conceptually, the
structure of the protein-DNA complex is determined by a competition between
attractive protein interactions and the configurational and loop entropy of
this protein-DNA cluster. Indeed, we show that the protein interaction strength
determines the "tightness" of the loopy protein-DNA complex. With this approach
we consider the genomic organization of such a protein-DNA cluster around a
single high-affinity binding site. Thus, our model provides a theoretical
framework to quantitatively compute the binding profiles of ParB-like proteins
around a cognate (parS) binding site. | [
0,
1,
0,
0,
0,
0
] |
Title: Mean field limits for nonlinear spatially extended hawkes processes with exponential memory kernels,
Abstract: We consider spatially extended systems of interacting nonlinear Hawkes
processes modeling large systems of neurons placed in Rd and study the
associated mean field limits. As the total number of neurons tends to infinity,
we prove that the evolution of a typical neuron, attached to a given spatial
position, can be described by a nonlinear limit differential equation driven by
a Poisson random measure. The limit process is described by a neural field
equation. As a consequence, we provide a rigorous derivation of the neural
field equation based on a thorough mean field analysis. | [
0,
0,
1,
0,
0,
0
] |
Title: Bias correction in daily maximum and minimum temperature measurements through Gaussian process modeling,
Abstract: The Global Historical Climatology Network-Daily database contains, among
other variables, daily maximum and minimum temperatures from weather stations
around the globe. It is long known that climatological summary statistics based
on daily temperature minima and maxima will not be accurate, if the bias due to
the time at which the observations were collected is not accounted for. Despite
some previous work, to our knowledge, there does not exist a satisfactory
solution to this important problem. In this paper, we carefully detail the
problem and develop a novel approach to address it. Our idea is to impute the
hourly temperatures at the location of the measurements by borrowing
information from the nearby stations that record hourly temperatures, which
then can be used to create accurate summaries of temperature extremes. The key
difficulty is that these imputations of the temperature curves must satisfy the
constraint of falling between the observed daily minima and maxima, and
attaining those values at least once in a twenty-four hour period. We develop a
spatiotemporal Gaussian process model for imputing the hourly measurements from
the nearby stations, and then develop a novel and easy to implement Markov
Chain Monte Carlo technique to sample from the posterior distribution
satisfying the above constraints. We validate our imputation model using hourly
temperature data from four meteorological stations in Iowa, of which one is
hidden and the data replaced with daily minima and maxima, and show that the
imputed temperatures recover the hidden temperatures well. We also demonstrate
that our model can exploit information contained in the data to infer the time
of daily measurements. | [
0,
0,
0,
1,
0,
0
] |
Title: Wave-induced vortex recoil and nonlinear refraction,
Abstract: When a vortex refracts surface waves, the momentum flux carried by the waves
changes direction and the waves induce a reaction force on the vortex. We study
experimentally the resulting vortex distortion. Incoming surface gravity waves
impinge on a steady vortex of velocity $U_0$ driven magneto-hydrodynamically at
the bottom of a fluid layer. The waves induce a shift of the vortex center in
the direction transverse to wave propagation, together with a decrease in
surface vorticity. We interpret these two phenomena in the framework introduced
by Craik and Leibovich (1976): we identify the dimensionless Stokes drift
$S=U_s/U_0$ as the relevant control parameter, $U_s$ being the Stokes drift
velocity of the waves. We propose a simple vortex line model which indicates
that the shift of the vortex center originates from a balance between vorticity
advection by the Stokes drift and self-advection of the vortex. The decrease in
surface vorticity is interpreted as a consequence of vorticity expulsion by the
fast Stokes drift, which confines it at depth. This purely hydrodynamic process
is analogous to the magnetohydrodynamic expulsion of magnetic field by a
rapidly moving conductor through the electromagnetic skin effect. We study
vorticity expulsion in the limit of fast Stokes drift and deduce that the
surface vorticity decreases as $1/S$, a prediction which is compatible with the
experimental data. Such wave-induced vortex distortions have important
consequences for the nonlinear regime of wave refraction: the refraction angle
rapidly decreases with wave intensity. | [
0,
1,
0,
0,
0,
0
] |
Title: First-principles prediction of the stacking fault energy of gold at finite temperature,
Abstract: The intrinsic stacking fault energy (ISFE) $\gamma$ is a material parameter
fundamental to the discussion of plastic deformation mechanisms in metals.
Here, we scrutinize the temperature dependence of the ISFE of Au through
accurate first-principles derived Helmholtz free energies employing both the
super cell approach and the axial Ising model (AIM). A significant decrease of
the ISFE with temperature, $-(36$-$39)$\,\% from 0 to 890\,K depending on the
treatment of thermal expansion, is revealed, which matches the estimate based
on the experimental temperature coefficient $d \gamma / d T $ closely. We make
evident that this decrease predominantly originates from the excess vibrational
entropy at the stacking fault layer, although the contribution arising from the
static lattice expansion compensates it by approximately 60\,\%. Electronic
excitations are found to be of minor importance for the ISFE change with
temperature. We show that the Debye model in combination with the AIM captures
the correct sign but significantly underestimates the magnitude of the
vibrational contribution to $\gamma(T)$. The hexagonal close-packed (hcp) and
double hcp structures are established as metastable phases of Au. Our results
demonstrate that quantitative agreement with experiments can be obtained if all
relevant temperature-induced excitations are considered in first-principles
modeling and that the temperature dependence of the ISFE is substantial enough
to be taken into account in crystal plasticity modeling. | [
0,
1,
0,
0,
0,
0
] |
Title: Throughput-Optimal Broadcast in Wireless Networks with Point-to-Multipoint Transmissions,
Abstract: We consider the problem of efficient packet dissemination in wireless
networks with point-to-multi-point wireless broadcast channels. We propose a
dynamic policy, which achieves the broadcast capacity of the network. This
policy is obtained by first transforming the original multi-hop network into a
precedence-relaxed virtual single-hop network and then finding an optimal
broadcast policy for the relaxed network. The resulting policy is shown to be
throughput-optimal for the original wireless network using a sample-path
argument. We also prove the NP-completeness of the finite-horizon broadcast
problem, which is in contrast with the polynomial time solvability of the
problem with point-to-point channels. Illustrative simulation results
demonstrate the efficacy of the proposed broadcast policy in achieving the full
broadcast capacity with low delay. | [
1,
0,
1,
0,
0,
0
] |
Title: Explaining Aviation Safety Incidents Using Deep Temporal Multiple Instance Learning,
Abstract: Although aviation accidents are rare, safety incidents occur more frequently
and require a careful analysis to detect and mitigate risks in a timely manner.
Analyzing safety incidents using operational data and producing event-based
explanations is invaluable to airline companies as well as to governing
organizations such as the Federal Aviation Administration (FAA) in the United
States. However, this task is challenging because of the complexity involved in
mining multi-dimensional heterogeneous time series data, the lack of
time-step-wise annotation of events in a flight, and the lack of scalable tools
to perform analysis over a large number of events. In this work, we propose a
precursor mining algorithm that identifies events in the multidimensional time
series that are correlated with the safety incident. Precursors are valuable to
systems health and safety monitoring and in explaining and forecasting safety
incidents. Current methods suffer from poor scalability to high dimensional
time series data and are inefficient in capturing temporal behavior. We propose
an approach by combining multiple-instance learning (MIL) and deep recurrent
neural networks (DRNN) to take advantage of MIL's ability to learn using weakly
supervised data and DRNN's ability to model temporal behavior. We describe the
algorithm, the data, the intuition behind taking a MIL approach, and a
comparative analysis of the proposed algorithm with baseline models. We also
discuss the application to a real-world aviation safety problem using data from
a commercial airline company and discuss the model's abilities and
shortcomings, with some final remarks about possible deployment directions. | [
1,
0,
0,
1,
0,
0
] |
Title: The Challenge of Spin-Orbit-Tuned Ground States in Iridates,
Abstract: Effects of spin-orbit interactions in condensed matter are an important and
rapidly evolving topic. Strong competition between spin-orbit, on-site Coulomb
and crystalline electric field interactions in iridates drives exotic quantum
states that are unique to this group of materials. In particular, the Jeff =
1/2 Mott state served as an early signal that the combined effect of strong
spin-orbit and Coulomb interactions in iridates has unique, intriguing
consequences. In this Key Issues Review, we survey some current experimental
studies of iridates. In essence, these materials tend to defy conventional
wisdom: absence of conventional correlations between magnetic and insulating
states, avoidance of metallization at high pressures, S-shaped I-V
characteristic, emergence of an odd-parity hidden order, etc. It is
particularly intriguing that there exist conspicuous discrepancies between
current experimental results and theoretical proposals that address
superconducting, topological and quantum spin liquid phases. This class of
materials, in which the lattice degrees of freedom play a critical role seldom
seen in other materials, evidently presents some profound intellectual
challenges that call for more investigations both experimentally and
theoretically. Physical properties unique to these materials may help unlock a
world of possibilities for functional materials and devices. We emphasize that,
given the rapidly developing nature of this field, this Key Issues Review is by
no means an exhaustive report of the current state of experimental studies of
iridates. | [
0,
1,
0,
0,
0,
0
] |
Title: Derived Picard groups of preprojective algebras of Dynkin type,
Abstract: In this paper, we study two-sided tilting complexes of preprojective algebras
of Dynkin type. We construct the most fundamental class of two-sided tilting
complexes, which has a group structure by derived tensor products and induces a
group of auto-equivalences of the derived category. We show that the group
structure of the two-sided tilting complexes is isomorphic to the braid group
of the corresponding folded graph. Moreover we show that these two-sided
tilting complexes induce tilting mutation and any tilting complex is given as
the derived tensor products of them. Using these results, we determine the
derived Picard group of preprojective algebras for type $A$ and $D$. | [
0,
0,
1,
0,
0,
0
] |
Title: Semi-supervised Learning in Network-Structured Data via Total Variation Minimization,
Abstract: We propose and analyze a method for semi-supervised learning from
partially-labeled network-structured data. Our approach is based on a graph
signal recovery interpretation under a clustering hypothesis that labels of
data points belonging to the same well-connected subset (cluster) are similar
valued. This lends naturally to learning the labels by total variation (TV)
minimization, which we solve by applying a recently proposed primal-dual method
for non-smooth convex optimization. The resulting algorithm allows for a highly
scalable implementation using message passing over the underlying empirical
graph, which renders the algorithm suitable for big data applications. By
applying tools of compressed sensing, we derive a sufficient condition on the
underlying network structure such that TV minimization recovers clusters in the
empirical graph of the data. In particular, we show that the proposed
primal-dual method amounts to maximizing network flows over the empirical graph
of the dataset. Moreover, the learning accuracy of the proposed algorithm is
linked to the set of network flows between data points having known labels. The
effectiveness and scalability of our approach is verified by numerical
experiments. | [
1,
0,
0,
1,
0,
0
] |
Title: Susceptibility of Methicillin Resistant Staphylococcus aureus to Vancomycin using Liposomal Drug Delivery System,
Abstract: Staphylococcus aureus responsible for nosocomial infections is a significant
threat to the public health. The increasing resistance of S.aureus to various
antibiotics has drawn it to a prime focus for research on designing an
appropriate drug delivery system. Emergence of Methicillin Resistant
Staphylococcus aureus (MRSA) in 1961, necessitated the use of vancomycin "the
drug of last resort" to treat these infections. Unfortunately, S.aureus has
already started gaining resistances to vancomycin. Liposome encapsulation of
drugs have been earlier shown to provide an efficient method of microbial
inhibition in many cases. We have studied the effect of liposome encapsulated
vancomycin on MRSA and evaluated the antibacterial activity of the
liposome-entrapped drug in comparison to that of the free drug based on the
minimum inhibitory concentration (MIC) of the drug. The MIC for liposomal
vancomycin was found to be about half of that of free vancomycin. The growth
response of MRSA showed that the liposomal vancomycin induced the culture to go
into bacteriostatic state and phagocytic killing was enhanced. Administration
of the antibiotic encapsulated in liposome thus was shown to greatly improve
the drug delivery as well as the drug resistance caused by MRSA. | [
0,
0,
0,
0,
1,
0
] |
Title: Pseudospectral Model Predictive Control under Partially Learned Dynamics,
Abstract: Trajectory optimization of a controlled dynamical system is an essential part
of autonomy, however many trajectory optimization techniques are limited by the
fidelity of the underlying parametric model. In the field of robotics, a lack
of model knowledge can be overcome with machine learning techniques, utilizing
measurements to build a dynamical model from the data. This paper aims to take
the middle ground between these two approaches by introducing a semi-parametric
representation of the underlying system dynamics. Our goal is to leverage the
considerable information contained in a traditional physics based model and
combine it with a data-driven, non-parametric regression technique known as a
Gaussian Process. Integrating this semi-parametric model with model predictive
pseudospectral control, we demonstrate this technique on both a cart pole and
quadrotor simulation with unmodeled damping and parametric error. In order to
manage parametric uncertainty, we introduce an algorithm that utilizes Sparse
Spectrum Gaussian Processes (SSGP) for online learning after each rollout. We
implement this online learning technique on a cart pole and quadrator, then
demonstrate the use of online learning and obstacle avoidance for the dubin
vehicle dynamics. | [
1,
0,
0,
0,
0,
0
] |
Title: Manifold Regularization for Kernelized LSTD,
Abstract: Policy evaluation or value function or Q-function approximation is a key
procedure in reinforcement learning (RL). It is a necessary component of policy
iteration and can be used for variance reduction in policy gradient methods.
Therefore its quality has a significant impact on most RL algorithms. Motivated
by manifold regularized learning, we propose a novel kernelized policy
evaluation method that takes advantage of the intrinsic geometry of the state
space learned from data, in order to achieve better sample efficiency and
higher accuracy in Q-function approximation. Applying the proposed method in
the Least-Squares Policy Iteration (LSPI) framework, we observe superior
performance compared to widely used parametric basis functions on two standard
benchmarks in terms of policy quality. | [
1,
0,
0,
1,
0,
0
] |
Title: Radio Weak Lensing Shear Measurement in the Visibility Domain - II. Source Extraction,
Abstract: This paper extends the method introduced in Rivi et al. (2016b) to measure
galaxy ellipticities in the visibility domain for radio weak lensing surveys.
In that paper we focused on the development and testing of the method for the
simple case of individual galaxies located at the phase centre, and proposed to
extend it to the realistic case of many sources in the field of view by
isolating visibilities of each source with a faceting technique. In this second
paper we present a detailed algorithm for source extraction in the visibility
domain and show its effectiveness as a function of the source number density by
running simulations of SKA1-MID observations in the band 950-1150 MHz and
comparing original and measured values of galaxies' ellipticities. Shear
measurements from a realistic population of 10^4 galaxies randomly located in a
field of view of 1 deg^2 (i.e. the source density expected for the current
radio weak lensing survey proposal with SKA1) are also performed. At SNR >= 10,
the multiplicative bias is only a factor 1.5 worse than what found when
analysing individual sources, and is still comparable to the bias values
reported for similar measurement methods at optical wavelengths. The additive
bias is unchanged from the case of individual sources, but is significantly
larger than typically found in optical surveys. This bias depends on the shape
of the uv coverage and we suggest that a uv-plane weighting scheme to produce a
more isotropic shape could reduce and control additive bias. | [
0,
1,
0,
0,
0,
0
] |
Title: MOA Data Reveal a New Mass, Distance, and Relative Proper Motion for Planetary System OGLE-2015-BLG-0954L,
Abstract: We present the MOA Collaboration light curve data for planetary microlensing
event OGLE-2015-BLG-0954, which was previously announced in a paper by the
KMTNet and OGLE Collaborations. The MOA data cover the caustic exit, which was
not covered by the KMTNet or OGLE data, and they provide a more reliable
measurement of the finite source effect. The MOA data also provide a new source
color measurement that reveals a lens-source relative proper motion of
$\mu_{\rm rel} = 11.8\pm 0.8\,$mas/yr, which compares to the value of $\mu_{\rm
rel} = 18.4\pm 1.7\,$mas/yr reported in the KMTNet-OGLE paper. This new MOA
value for $\mu_{\rm rel}$ has an a priori probability that is a factor of $\sim
100$ times larger than the previous value, and it does not require a lens
system distance of $D_L < 1\,$kpc. Based on the corrected source color, we find
that the lens system consists of a planet of mass $3.4^{+3.7}_{-1.6} M_{\rm
Jup}$ orbiting a $0.30^{+0.34}_{-0.14}M_\odot$ star at an orbital separation of
$2.1^{+2.2}_{-1.0}\,$AU and a distance of $1.2^{+1.1}_{-0.5}\,$kpc. | [
0,
1,
0,
0,
0,
0
] |
Title: Single crystal polarized neutron diffraction study of the magnetic structure of HoFeO$_3$,
Abstract: Polarised neutron diffraction measurements have been made on HoFeO$_3$ single
crystals magnetised in both the [001] and [100] directions ($Pbnm$ setting).
The polarisation dependencies of Bragg reflection intensities were measured
both with a high field of H = 9 T parallel to [001] at T = 70 K and with the
lower field H = 0.5 T parallel to [100] at T = 5, 15, 25~K. A Fourier
projection of magnetization induced parallel to [001], made using the $hk0$
reflections measured in 9~T, indicates that almost all of it is due to
alignment of Ho moments. Further analysis of the asymmetries of general
reflections in these data showed that although, at 70~K, 9~T applied parallel
to [001] hardly perturbs the antiferromagnetic order of the Fe sublattices, it
induces significant antiferromagnetic order of the Ho sublattices in the
$x\mhyphen y$ plane, with the antiferromagnetic components of moment having the
same order of magnitude as the induced ferromagnetic ones. Strong intensity
asymmetries measured in the low temperature $\Gamma_2$ structure with a lower
field, 0.5 T $\parallel$ [100] allowed the variation of the ordered components
of the Ho and Fe moments to be followed. Their absolute orientations, in the
180\degree\ domain stabilised by the field were determined relative to the
distorted perovskite structure,. This relationship fixes the sign of the
Dzyalshinski-Moriya (D-M) interaction which leads to the weak ferromagnetism.
Our results indicate that the combination of strong y-axis anisotropy of the Ho
moments and Ho-Fe exchange interactions breaks the centrosymmetry of the
structure and could lead to ferroelectric polarization. | [
0,
1,
0,
0,
0,
0
] |
Title: Rigidity of volume-minimizing hypersurfaces in Riemannian 5-manifolds,
Abstract: In this paper we generalize the main result of [4] for manifolds that are not
necessarily Einstein. In fact, we obtain an upper bound for the volume of a
locally volume-minimizing closed hypersurface $\Sigma$ of a Riemannian
5-manifold $M$ with scalar curvature bounded from below by a positive constant
in terms of the total traceless Ricci curvature of $\Sigma$. Furthermore, if
$\Sigma$ saturates the respective upper bound and $M$ has nonnegative Ricci
curvature, then $\Sigma$ is isometric to $\mathbb{S}^4$ up to scaling and $M$
splits in a neighborhood of $\Sigma$. Also, we obtain a rigidity result for the
Riemannian cover of $M$ when $\Sigma$ minimizes the volume in its homotopy
class and saturates the upper bound. | [
0,
0,
1,
0,
0,
0
] |
Title: Dolha - an Efficient and Exact Data Structure for Streaming Graphs,
Abstract: A streaming graph is a graph formed by a sequence of incoming edges with time
stamps. Unlike static graphs, the streaming graph is highly dynamic and time
related. In the real world, the high volume and velocity streaming graphs such
as internet traffic data, social network communication data and financial
transfer data are bringing challenges to the classic graph data structures. We
present a new data structure: double orthogonal list in hash table (Dolha)
which is a high speed and high memory efficiency graph structure applicable to
streaming graph. Dolha has constant time cost for single edge and near linear
space cost that we can contain billions of edges information in memory size and
process an incoming edge in nanoseconds. Dolha also has linear time cost for
neighborhood queries, which allow it to support most algorithms in graphs
without extra cost. We also present a persistent structure based on Dolha that
has the ability to handle the sliding window update and time related queries. | [
1,
0,
0,
0,
0,
0
] |
Title: Minimal Representations of Lie Algebras With Non-Trivial Levi Decomposition,
Abstract: We obtain minimal dimension matrix representations for each of the Lie
algebras of dimension five, six, seven, and eight obtained by Turkowski that
have a non-trivial Levi decomposition. The Key technique involves using
subspace associated to a particular representation of semi-simple Lie algebra
to help in the construction of the radical in the putative Levi decomposition. | [
0,
0,
1,
0,
0,
0
] |
Title: A Longitudinal Higher-Order Diagnostic Classification Model,
Abstract: Providing diagnostic feedback about growth is crucial to formative decisions
such as targeted remedial instructions or interventions. This paper proposed a
longitudinal higher-order diagnostic classification modeling approach for
measuring growth. The new modeling approach is able to provide quantitative
values of overall and individual growth by constructing a multidimensional
higher-order latent structure to take into account the correlations among
multiple latent attributes that are examined across different occasions. In
addition, potential local item dependence among anchor (or repeated) items can
also be taken into account. Model parameter estimation is explored in a
simulation study. An empirical example is analyzed to illustrate the
applications and advantages of the proposed modeling approach. | [
0,
0,
0,
1,
0,
0
] |
Title: On Fairness and Calibration,
Abstract: The machine learning community has become increasingly concerned with the
potential for bias and discrimination in predictive models. This has motivated
a growing line of work on what it means for a classification procedure to be
"fair." In this paper, we investigate the tension between minimizing error
disparity across different population groups while maintaining calibrated
probability estimates. We show that calibration is compatible only with a
single error constraint (i.e. equal false-negatives rates across groups), and
show that any algorithm that satisfies this relaxation is no better than
randomizing a percentage of predictions for an existing classifier. These
unsettling findings, which extend and generalize existing results, are
empirically confirmed on several datasets. | [
1,
0,
0,
1,
0,
0
] |
Title: Role of 1-D finite size Heisenberg chain in increasing metal to insulator transition temperature in hole rich VO2,
Abstract: VO2 samples are grown with different oxygen concentrations leading to
different monoclinic, M1 and triclinic, T insulating phases which undergo a
first order metal to insulator transition (MIT) followed by a structural phase
transition (SPT) to rutile tetragonal phase. The metal insulator transition
temperature (Tc) was found to be increased with increasing native defects.
Vanadium vacancy (VV) is envisaged to create local strains in the lattice which
prevents twisting of the V-V dimers promoting metastable monoclinic, M2 and T
phases at intermediate temperatures. It is argued that MIT is driven by strong
electronic correlation. The low temperature insulating phase can be considered
as a collection of one-dimensional (1-D) half-filled band, which undergoes Mott
transition to 1-D infinitely long Heisenberg spin 1/2 chains leading to
structural distortion due to spin-phonon coupling. Presence of VV creates
localized holes (d0) in the nearest neighbor, thereby fragmenting the spin 1/2
chains at nanoscale, which in turn increase the Tc value more than that of an
infinitely long one. The Tc value scales inversely with the average size of
fragmented Heisenberg spin 1/2 chains following a critical exponent of 2/3,
which is exactly the same predicted theoretically for Heisenberg spin 1/2 chain
at nanoscale undergoing SPT (spin-Peierls transition). Thus, the observation of
MIT and SPT at the same time in VO2 can be explained from our phenomenological
model of reduced 1-D Heisenberg spin 1/2 chains. The reported increase
(decrease) in Tc value of VO2 by doping with metal having valency less (more)
than four, can also be understood easily with our unified model, for the first
time, considering finite size scaling of Heisenberg chains. | [
0,
1,
0,
0,
0,
0
] |
Title: A multilayer multiconfiguration time-dependent Hartree study of the nonequilibrium Anderson impurity model at zero temperature,
Abstract: Quantum transport is studied for the nonequilibrium Anderson impurity model
at zero temperature employing the multilayer multiconfiguration time-dependent
Hartree theory within the second quantization representation (ML-MCTDH-SQR) of
Fock space. To adress both linear and nonlinear conductance in the Kondo
regime, two new techniques of the ML-MCTDH-SQR simulation methodology are
introduced: (i) the use of correlated initial states, which is achieved by
imaginary time propagation of the overall Hamiltonian at zero voltage and (ii)
the adoption of the logarithmic discretization of the electronic continuum.
Employing the improved methodology, the signature of the Kondo effect is
analyzed. | [
0,
1,
0,
0,
0,
0
] |
Title: Semi-Lagrangian one-step methods for two classes of time-dependent partial differential systems,
Abstract: Semi-Lagrangian methods are numerical methods designed to find approximate
solutions to particular time-dependent partial differential equations (PDEs)
that describe the advection process. We propose semi-Lagrangian one-step
methods for numerically solving initial value problems for two general systems
of partial differential equations. Along the characteristic lines of the PDEs,
we use ordinary differential equation (ODE) numerical methods to solve the
PDEs. The main benefit of our methods is the efficient achievement of high
order local truncation error through the use of Runge-Kutta methods along the
characteristics. In addition, we investigate the numerical analysis of
semi-Lagrangian methods applied to systems of PDEs: stability, convergence, and
maximum error bounds. | [
0,
0,
1,
0,
0,
0
] |
Title: Efficient Nearest-Neighbor Search for Dynamical Systems with Nonholonomic Constraints,
Abstract: Nearest-neighbor search dominates the asymptotic complexity of sampling-based
motion planning algorithms and is often addressed with k-d tree data
structures. While it is generally believed that the expected complexity of
nearest-neighbor queries is $O(log(N))$ in the size of the tree, this paper
reveals that when a classic k-d tree approach is used with sub-Riemannian
metrics, the expected query complexity is in fact $\Theta(N^p \log(N))$ for a
number $p \in [0, 1)$ determined by the degree of nonholonomy of the system.
These metrics arise naturally in nonholonomic mechanical systems, including
classic wheeled robot models. To address this negative result, we propose novel
k-d tree build and query strategies tailored to sub-Riemannian metrics and
demonstrate significant improvements in the running time of nearest-neighbor
search queries. | [
1,
0,
0,
0,
0,
0
] |
Title: Primitivity, Uniform Minimality and State Complexity of Boolean Operations,
Abstract: A minimal deterministic finite automaton (DFA) is uniformly minimal if it
always remains minimal when the final state set is replaced by a non-empty
proper subset of the state set. We prove that a permutation DFA is uniformly
minimal if and only if its transition monoid is a primitive group. We use this
to study boolean operations on group languages, which are recognized by direct
products of permutation DFAs. A direct product cannot be uniformly minimal,
except in the trivial case where one of the DFAs in the product is a one-state
DFA. However, non-trivial direct products can satisfy a weaker condition we
call uniform boolean minimality, where only final state sets used to recognize
boolean operations are considered. We give sufficient conditions for a direct
product of two DFAs to be uniformly boolean minimal, which in turn gives
sufficient conditions for pairs of group languages to have maximal state
complexity under all binary boolean operations ("maximal boolean complexity").
In the case of permutation DFAs with one final state, we give necessary and
sufficient conditions for pairs of group languages to have maximal boolean
complexity. Our results demonstrate a connection between primitive groups and
automata with strong minimality properties. | [
1,
0,
1,
0,
0,
0
] |
Title: Summary of Topological Study of Chaotic CBC Mode of Operation,
Abstract: In cryptography, block ciphers are the most fundamental elements in many
symmetric-key encryption systems. The Cipher Block Chaining, denoted CBC,
presents one of the most famous mode of operation that uses a block cipher to
provide confidentiality or authenticity. In this research work, we intend to
summarize our results that have been detailed in our previous series of
articles. The goal of this series has been to obtain a complete topological
study of the CBC block cipher mode of operation after proving his chaotic
behavior according to the reputed definition of Devaney. | [
1,
1,
0,
0,
0,
0
] |
Title: MSO+nabla is undecidable,
Abstract: This paper is about an extension of monadic second-order logic over infinite
trees, which adds a quantifier that says "the set of branches \pi which satisfy
a formula \phi(\pi) has probability one". This logic was introduced by
Michalewski and Mio; we call it MSO+nabla following Shelah and Lehmann. The
logic MSO+nabla subsumes many qualitative probabilistic formalisms, including
qualitative probabilistic CTL, probabilistic LTL, or parity tree automata with
probabilistic acceptance conditions. We consider the decision problem: decide
if a sentence of MSO+nabla is true in the infinite binary tree? For sentences
from the weak variant of this logic (set quantifiers range only over finite
sets) the problem was known to be decidable, but the question for the full
logic remained open. In this paper we show that the problem for the full logic
MSO+nabla is undecidable. | [
1,
0,
0,
0,
0,
0
] |
Title: Computer-aided implant design for the restoration of cranial defects,
Abstract: Patient-specific cranial implants are important and necessary in the surgery
of cranial defect restoration. However, traditional methods of manual design of
cranial implants are complicated and time-consuming. Our purpose is to develop
a novel software named EasyCrania to design the cranial implants conveniently
and efficiently. The process can be divided into five steps, which are
mirroring model, clipping surface, surface fitting, the generation of the
initial implant and the generation of the final implant. The main concept of
our method is to use the geometry information of the mirrored model as the base
to generate the final implant. The comparative studies demonstrated that the
EasyCrania can improve the efficiency of cranial implant design significantly.
And, the intra- and inter-rater reliability of the software were stable, which
were 87.07+/-1.6% and 87.73+/-1.4% respectively. | [
1,
0,
0,
0,
0,
0
] |
Title: Dimension-free Information Concentration via Exp-Concavity,
Abstract: Information concentration of probability measures have important implications
in learning theory. Recently, it is discovered that the information content of
a log-concave distribution concentrates around their differential entropy,
albeit with an unpleasant dependence on the ambient dimension. In this work, we
prove that if the potentials of the log-concave distribution are exp-concave,
which is a central notion for fast rates in online and statistical learning,
then the concentration of information can be further improved to depend only on
the exp-concavity parameter, and hence, it can be dimension independent.
Central to our proof is a novel yet simple application of the variance
Brascamp-Lieb inequality. In the context of learning theory, our
concentration-of-information result immediately implies high-probability
results to many of the previous bounds that only hold in expectation. | [
0,
0,
0,
1,
0,
0
] |
Title: Transfer results for Frobenius extensions,
Abstract: We study Frobenius extensions which are free-filtered by a totally ordered,
finitely generated abelian group, and their free-graded counterparts. First we
show that the Frobenius property passes up from a free-graded extension to a
free-filtered extension, then also from a free-filtered extension to the
extension of their Rees algebras. Our main theorem states that, under some
natural hypotheses, a free-filtered extension of algebras is Frobenius if and
only if the associated graded extension is Frobenius. In the final section we
apply this theorem to provide new examples and non-examples of Frobenius
extensions. | [
0,
0,
1,
0,
0,
0
] |
Title: Recency-weighted Markovian inference,
Abstract: We describe a Markov latent state space (MLSS) model, where the latent state
distribution is a decaying mixture over multiple past states. We present a
simple sampling algorithm that allows to approximate such high-order MLSS with
fixed time and memory costs. | [
1,
0,
0,
1,
0,
0
] |
Title: A revisit on the compactness of commutators,
Abstract: A new characterization of CMO(R^n) is established by the local mean
oscillation. Some characterizations of iterated compact commutators on weighted
Lebesgue spaces are given, which are new even in the unweighted setting for the
first order commutators. | [
0,
0,
1,
0,
0,
0
] |
Title: Question Answering through Transfer Learning from Large Fine-grained Supervision Data,
Abstract: We show that the task of question answering (QA) can significantly benefit
from the transfer learning of models trained on a different large, fine-grained
QA dataset. We achieve the state of the art in two well-studied QA datasets,
WikiQA and SemEval-2016 (Task 3A), through a basic transfer learning technique
from SQuAD. For WikiQA, our model outperforms the previous best model by more
than 8%. We demonstrate that finer supervision provides better guidance for
learning lexical and syntactic information than coarser supervision, through
quantitative results and visual analysis. We also show that a similar transfer
learning procedure achieves the state of the art on an entailment task. | [
1,
0,
0,
0,
0,
0
] |
Title: On mesoprimary decomposition of monoid congruences,
Abstract: We prove two main results concerning mesoprimary decomposition of monoid
congruences, as introduced by Kahle and Miller. First, we identify which
associated prime congruences appear in every mesoprimary decomposition, thereby
completing the theory of mesoprimary decomposition of monoid congruences as a
more faithful analog of primary decomposition. Second, we answer a question
posed by Kahle and Miller by characterizing which finite posets arise as the
set of associated prime congruences of monoid congruences. | [
0,
0,
1,
0,
0,
0
] |
Title: Assessing the state of e-Readiness for Small and Medium Companies in Mexico: a Proposed Taxonomy and Adoption Model,
Abstract: Emerging economies frequently show a large component of their Gross Domestic
Product to be dependant on the economic activity of small and medium
enterprises. Nevertheless, e-business solutions are more likely designed for
large companies. SMEs seem to follow a classical family-based management, used
to traditional activities, rather than seeking new ways of adding value to
their business strategy. Thus, a large portion of a nations economy may be at
disadvantage for competition. This paper aims at assessing the state of
e-business readiness of Mexican SMEs based on already published e-business
evolution models and by means of a survey research design. Data is being
collected in three cities with differing sizes and infrastructure conditions.
Statistical results are expected to be presented. A second part of this
research aims at applying classical adoption models to suggest potential causal
relationships, as well as more suitable recommendations for development. | [
0,
0,
0,
0,
0,
1
] |
Title: Data Augmentation for Robust Keyword Spotting under Playback Interference,
Abstract: Accurate on-device keyword spotting (KWS) with low false accept and false
reject rate is crucial to customer experience for far-field voice control of
conversational agents. It is particularly challenging to maintain low false
reject rate in real world conditions where there is (a) ambient noise from
external sources such as TV, household appliances, or other speech that is not
directed at the device (b) imperfect cancellation of the audio playback from
the device, resulting in residual echo, after being processed by the Acoustic
Echo Cancellation (AEC) system. In this paper, we propose a data augmentation
strategy to improve keyword spotting performance under these challenging
conditions. The training set audio is artificially corrupted by mixing in music
and TV/movie audio, at different signal to interference ratios. Our results
show that we get around 30-45% relative reduction in false reject rates, at a
range of false alarm rates, under audio playback from such devices. | [
0,
0,
0,
1,
0,
0
] |
Title: A Canonical-based NPN Boolean Matching Algorithm Utilizing Boolean Difference and Cofactor Signature,
Abstract: This paper presents a new compact canonical-based algorithm to solve the
problem of single-output completely specified NPN Boolean matching. We propose
a new signature vector Boolean difference and cofactor (DC) signature vector.
Our algorithm utilizes the Boolean difference, cofactor signature and symmetry
properties to search for canonical transformations. The use of symmetry and
Boolean difference notably reduces the search space and speeds up the Boolean
matching process compared to the algorithm proposed in [1]. We tested our
algorithm on a large number of circuits. The experimental results showed that
the average runtime of our algorithm 37% higher and its average search space
67% smaller compared to [1] when tested on general circuits. | [
1,
0,
0,
0,
0,
0
] |
Title: Cross ratios on boundaries of symmetric spaces and Euclidean buildings,
Abstract: We generalize the natural cross ratio on the ideal boundary of a rank one
symmetric spaces, or even $\mathrm{CAT}(-1)$ space, to higher rank symmetric
spaces and (non-locally compact) Euclidean buildings - we obtain vector valued
cross ratios defined on simplices of the building at infinity. We show several
properties of those cross ratios; for example that (under some restrictions)
periods of hyperbolic isometries give back the translation vector. In addition,
we show that cross ratio preserving maps on the chamber set are induced by
isometries and vice versa - motivating that the cross ratios bring the geometry
of the symmetric space/Euclidean building to the boundary. | [
0,
0,
1,
0,
0,
0
] |
Title: Heterogeneous inputs to central pattern generators can shape insect gaits,
Abstract: In our previous work, we studied an interconnected bursting neuron model for
insect locomotion, and its corresponding phase oscillator model, which at high
speed can generate stable tripod gaits with three legs off the ground
simultaneously in swing, and at low speed can generate stable tetrapod gaits
with two legs off the ground simultaneously in swing. However, at low speed
several other stable locomotion patterns, that are not typically observed as
insect gaits, may coexist. In the present paper, by adding heterogeneous
external input to each oscillator, we modify the bursting neuron model so that
its corresponding phase oscillator model produces only one stable gait at each
speed, specifically: a unique stable tetrapod gait at low speed, a unique
stable tripod gait at high speed, and a unique branch of stable transition
gaits connecting them. This suggests that control signals originating in the
brain and central nervous system can modify gait patterns. | [
0,
0,
0,
0,
1,
0
] |
Title: Pinning down the mass of Kepler-10c: the importance of sampling and model comparison,
Abstract: Initial RV characterisation of the enigmatic planet Kepler-10c suggested a
mass of $\sim17$ M$_\oplus$, which was remarkably high for a planet with radius
$2.32$ R$_\oplus$; further observations and subsequent analysis hinted at a
(possibly much) lower mass, but masses derived using RVs from two different
spectrographs (HARPS-N and HIRES) were incompatible at a $3\sigma$-level. We
demonstrate here how such mass discrepancies may readily arise from sub-optimal
sampling and/or neglecting to model even a single coherent signal (stellar,
planetary, or otherwise) that may be present in RVs. We then present a
plausible resolution of the mass discrepancy, and ultimately characterise
Kepler-10c as having mass $7.37_{-1.19}^{+1.32}$ M$_\oplus$, and mean density
$3.14^{+0.63}_{-0.55}$ g cm$^{-3}$. | [
0,
1,
0,
0,
0,
0
] |
Title: Human-Centered Autonomous Vehicle Systems: Principles of Effective Shared Autonomy,
Abstract: Building effective, enjoyable, and safe autonomous vehicles is a lot harder
than has historically been considered. The reason is that, simply put, an
autonomous vehicle must interact with human beings. This interaction is not a
robotics problem nor a machine learning problem nor a psychology problem nor an
economics problem nor a policy problem. It is all of these problems put into
one. It challenges our assumptions about the limitations of human beings at
their worst and the capabilities of artificial intelligence systems at their
best. This work proposes a set of principles for designing and building
autonomous vehicles in a human-centered way that does not run away from the
complexity of human nature but instead embraces it. We describe our development
of the Human-Centered Autonomous Vehicle (HCAV) as an illustrative case study
of implementing these principles in practice. | [
1,
0,
0,
0,
0,
0
] |
Title: A Relaxation-based Network Decomposition Algorithm for Parallel Transient Stability Simulation with Improved Convergence,
Abstract: Transient stability simulation of a large-scale and interconnected electric
power system involves solving a large set of differential algebraic equations
(DAEs) at every simulation time-step. With the ever-growing size and complexity
of power grids, dynamic simulation becomes more time-consuming and
computationally difficult using conventional sequential simulation techniques.
To cope with this challenge, this paper aims to develop a fully distributed
approach intended for implementation on High Performance Computer (HPC)
clusters. A novel, relaxation-based domain decomposition algorithm known as
Parallel-General-Norton with Multiple-port Equivalent (PGNME) is proposed as
the core technique of a two-stage decomposition approach to divide the overall
dynamic simulation problem into a set of subproblems that can be solved
concurrently to exploit parallelism and scalability. While the convergence
property has traditionally been a concern for relaxation-based decomposition,
an estimation mechanism based on multiple-port network equivalent is adopted as
the preconditioner to enhance the convergence of the proposed algorithm. The
proposed algorithm is illustrated using rigorous mathematics and validated both
in terms of speed-up and capability. Moreover, a complexity analysis is
performed to support the observation that PGNME scales well when the size of
the subproblems are sufficiently large. | [
1,
0,
0,
0,
0,
0
] |
Title: Random group cobordisms of rank 7/4,
Abstract: We construct a model of random groups of rank 7/4, and show that in this
model the random group has the exponential mesoscopic rank property. | [
0,
0,
1,
0,
0,
0
] |
Title: Transport properties across the many-body localization transition in quasiperiodic and random systems,
Abstract: We theoretically study transport properties in one-dimensional interacting
quasiperiodic systems at infinite temperature. We compare and contrast the
dynamical transport properties across the many-body localization (MBL)
transition in quasiperiodic and random models. Using exact diagonalization we
compute the optical conductivity $\sigma(\omega)$ and the return probability
$R(\tau)$ and study their average low-frequency and long-time power-law
behavior, respectively. We show that the low-energy transport dynamics is
markedly distinct in both the thermal and MBL phases in quasiperiodic and
random models and find that the diffusive and MBL regimes of the quasiperiodic
model are more robust than those in the random system. Using the distribution
of the DC conductivity, we quantify the contribution of sample-to-sample and
state-to-state fluctuations of $\sigma(\omega)$ across the MBL transition. We
find that the activated dynamical scaling ansatz works poorly in the
quasiperiodic model but holds in the random model with an estimated activation
exponent $\psi\approx 0.9$. We argue that near the MBL transition in
quasiperiodic systems, critical eigenstates give rise to a subdiffusive
crossover regime on finite-size systems. | [
0,
1,
0,
0,
0,
0
] |
Title: Analytic solution and stationary phase approximation for the Bayesian lasso and elastic net,
Abstract: The lasso and elastic net linear regression models impose a
double-exponential prior distribution on the model parameters to achieve
regression shrinkage and variable selection, allowing the inference of robust
models from large data sets. However, there has been limited success in
deriving estimates for the full posterior distribution of regression
coefficients in these models, due to a need to evaluate analytically
intractable partition function integrals. Here, the Fourier transform is used
to express these integrals as complex-valued oscillatory integrals over
"regression frequencies". This results in an analytic expansion and stationary
phase approximation for the partition functions of the Bayesian lasso and
elastic net, where the non-differentiability of the double-exponential prior
has so far eluded such an approach. Use of this approximation leads to highly
accurate numerical estimates for the expectation values and marginal posterior
distributions of the regression coefficients, and allows for Bayesian inference
of much higher dimensional models than previously possible. | [
0,
0,
1,
1,
0,
0
] |
Title: Deep Learning Assisted Heuristic Tree Search for the Container Pre-marshalling Problem,
Abstract: One of the key challenges for operations researchers solving real-world
problems is designing and implementing high-quality heuristics to guide their
search procedures. In the past, machine learning techniques have failed to play
a major role in operations research approaches, especially in terms of guiding
branching and pruning decisions. We integrate deep neural networks into a
heuristic tree search procedure to decide which branch to choose next and to
estimate a bound for pruning the search tree of an optimization problem. We
call our approach Deep Learning assisted heuristic Tree Search (DLTS) and apply
it to a well-known problem from the container terminals literature, the
container pre-marshalling problem (CPMP). Our approach is able to learn
heuristics customized to the CPMP solely through analyzing the solutions to
CPMP instances, and applies this knowledge within a heuristic tree search to
produce the highest quality heuristic solutions to the CPMP to date. | [
1,
0,
0,
0,
0,
0
] |
Title: Efficient Benchmarking of Algorithm Configuration Procedures via Model-Based Surrogates,
Abstract: The optimization of algorithm (hyper-)parameters is crucial for achieving
peak performance across a wide range of domains, ranging from deep neural
networks to solvers for hard combinatorial problems. The resulting algorithm
configuration (AC) problem has attracted much attention from the machine
learning community. However, the proper evaluation of new AC procedures is
hindered by two key hurdles. First, AC benchmarks are hard to set up. Second
and even more significantly, they are computationally expensive: a single run
of an AC procedure involves many costly runs of the target algorithm whose
performance is to be optimized in a given AC benchmark scenario. One common
workaround is to optimize cheap-to-evaluate artificial benchmark functions
(e.g., Branin) instead of actual algorithms; however, these have different
properties than realistic AC problems. Here, we propose an alternative
benchmarking approach that is similarly cheap to evaluate but much closer to
the original AC problem: replacing expensive benchmarks by surrogate benchmarks
constructed from AC benchmarks. These surrogate benchmarks approximate the
response surface corresponding to true target algorithm performance using a
regression model, and the original and surrogate benchmark share the same
(hyper-)parameter space. In our experiments, we construct and evaluate
surrogate benchmarks for hyperparameter optimization as well as for AC problems
that involve performance optimization of solvers for hard combinatorial
problems, drawing training data from the runs of existing AC procedures. We
show that our surrogate benchmarks capture overall important characteristics of
the AC scenarios, such as high- and low-performing regions, from which they
were derived, while being much easier to use and orders of magnitude cheaper to
evaluate. | [
1,
0,
0,
1,
0,
0
] |
Title: An open-source platform to study uniaxial stress effects on nanoscale devices,
Abstract: We present an automatic measurement platform that enables the
characterization of nanodevices by electrical transport and optical
spectroscopy as a function of uniaxial stress. We provide insights into and
detailed descriptions of the mechanical device, the substrate design and
fabrication, and the instrument control software, which is provided under
open-source license. The capability of the platform is demonstrated by
characterizing the piezo-resistance of an InAs nanowire device using a
combination of electrical transport and Raman spectroscopy. The advantages of
this measurement platform are highlighted by comparison with state-of-the-art
piezo-resistance measurements in InAs nanowires. We envision that the
systematic application of this methodology will provide new insights into the
physics of nanoscale devices and novel materials for electronics, and thus
contribute to the assessment of the potential of strain as a technology booster
for nanoscale electronics. | [
0,
1,
0,
0,
0,
0
] |
Title: Parametrised second-order complexity theory with applications to the study of interval computation,
Abstract: We extend the framework for complexity of operators in analysis devised by
Kawamura and Cook (2012) to allow for the treatment of a wider class of
representations. The main novelty is to endow represented spaces of interest
with an additional function on names, called a parameter, which measures the
complexity of a given name. This parameter generalises the size function which
is usually used in second-order complexity theory and therefore also central to
the framework of Kawamura and Cook. The complexity of an algorithm is measured
in terms of its running time as a second-order function in the parameter, as
well as in terms of how much it increases the complexity of a given name, as
measured by the parameters on the input and output side.
As an application we develop a rigorous computational complexity theory for
interval computation. In the framework of Kawamura and Cook the representation
of real numbers based on nested interval enclosures does not yield a reasonable
complexity theory. In our new framework this representation is polytime
equivalent to the usual Cauchy representation based on dyadic rational
approximation. By contrast, the representation of continuous real functions
based on interval enclosures is strictly smaller in the polytime reducibility
lattice than the usual representation, which encodes a modulus of continuity.
Furthermore, the function space representation based on interval enclosures is
optimal in the sense that it contains the minimal amount of information amongst
those representations which render evaluation polytime computable. | [
1,
0,
0,
0,
0,
0
] |
Title: VEEGAN: Reducing Mode Collapse in GANs using Implicit Variational Learning,
Abstract: Deep generative models provide powerful tools for distributions over
complicated manifolds, such as those of natural images. But many of these
methods, including generative adversarial networks (GANs), can be difficult to
train, in part because they are prone to mode collapse, which means that they
characterize only a few modes of the true distribution. To address this, we
introduce VEEGAN, which features a reconstructor network, reversing the action
of the generator by mapping from data to noise. Our training objective retains
the original asymptotic consistency guarantee of GANs, and can be interpreted
as a novel autoencoder loss over the noise. In sharp contrast to a traditional
autoencoder over data points, VEEGAN does not require specifying a loss
function over the data, but rather only over the representations, which are
standard normal by assumption. On an extensive set of synthetic and real world
image datasets, VEEGAN indeed resists mode collapsing to a far greater extent
than other recent GAN variants, and produces more realistic samples. | [
0,
0,
0,
1,
0,
0
] |
Title: Adaptive posterior contraction rates for the horseshoe,
Abstract: We investigate the frequentist properties of Bayesian procedures for
estimation based on the horseshoe prior in the sparse multivariate normal means
model. Previous theoretical results assumed that the sparsity level, that is,
the number of signals, was known. We drop this assumption and characterize the
behavior of the maximum marginal likelihood estimator (MMLE) of a key parameter
of the horseshoe prior. We prove that the MMLE is an effective estimator of the
sparsity level, in the sense that it leads to (near) minimax optimal estimation
of the underlying mean vector generating the data. Besides this empirical Bayes
procedure, we consider the hierarchical Bayes method of putting a prior on the
unknown sparsity level as well. We show that both Bayesian techniques lead to
rate-adaptive optimal posterior contraction, which implies that the horseshoe
posterior is a good candidate for generating rate-adaptive credible sets. | [
0,
0,
1,
1,
0,
0
] |
Title: Investigating Enactive Learning for Autonomous Intelligent Agents,
Abstract: The enactive approach to cognition is typically proposed as a viable
alternative to traditional cognitive science. Enactive cognition displaces the
explanatory focus from the internal representations of the agent to the direct
sensorimotor interaction with its environment. In this paper, we investigate
enactive learning through means of artificial agent simulations. We compare the
performances of the enactive agent to an agent operating on classical
reinforcement learning in foraging tasks within maze environments. The
characteristics of the agents are analysed in terms of the accessibility of the
environmental states, goals, and exploration/exploitation tradeoffs. We confirm
that the enactive agent can successfully interact with its environment and
learn to avoid unfavourable interactions using intrinsically defined goals. The
performance of the enactive agent is shown to be limited by the number of
affordable actions. | [
1,
0,
0,
0,
0,
0
] |
Title: Predicting Pulsar Scintillation from Refractive Plasma Sheets,
Abstract: The dynamic and secondary spectra of many pulsars show evidence for
long-lived, aligned images of the pulsar that are stationary on a thin
scattering sheet. One explanation for this phenomenon considers the effects of
wave crests along sheets in the ionized interstellar medium, such as those due
to Alfvén waves propagating along current sheets. If these sheets are closely
aligned to our line-of-sight to the pulsar, high bending angles arise at the
wave crests and a selection effect causes alignment of images produced at
different crests, similar to grazing reflection off of a lake. Using geometric
optics, we develop a simple parameterized model of these corrugated sheets that
can be constrained with a single observation and that makes observable
predictions for variations in the scintillation of the pulsar over time and
frequency. This model reveals qualitative differences between lensing from
overdense and underdense corrugated sheets: Only if the sheet is overdense
compared to the surrounding interstellar medium can the lensed images be
brighter than the line-of-sight image to the pulsar, and the faint lensed
images are closer to the pulsar at higher frequencies if the sheet is
underdense, but at lower frequencies if the sheet is overdense. | [
0,
1,
0,
0,
0,
0
] |
Title: Quantum Monte Carlo with variable spins: fixed-phase and fixed-node approximations,
Abstract: We study several aspects of the recently introduced fixed-phase spin-orbit
diffusion Monte Carlo (FPSODMC) method, in particular, its relation to the
fixed-node method and its potential use as a general approach for electronic
structure calculations. We illustrate constructions of spinor-based wave
functions with the full space-spin symmetry without assigning up or down spin
labels to particular electrons, effectively "complexifying" even ordinary
real-valued wave functions. Interestingly, with proper choice of the simulation
parameters and spin variables, such fixed-phase calculations enable one to
reach also the fixed-node limit. The fixed-phase solution provides a
straightforward interpretation as the lowest bosonic state in a given effective
potential generated by the many-body approximate phase. In addition, the
divergences present at real wave function nodes are smoothed out to lower
dimensionality, decreasing thus the variation of sampled quantities and making
the sampling also more straightforward. We illustrate some of these properties
on calculations of selected first-row systems that recover the fixed-node
results with quantitatively similar levels of the corresponding biases. At the
same time, the fixed-phase approach opens new possibilities for more general
trial wave functions with further opportunities for increasing accuracy in
practical calculations. | [
0,
1,
0,
0,
0,
0
] |
Title: A Tutorial on Canonical Correlation Methods,
Abstract: Canonical correlation analysis is a family of multivariate statistical
methods for the analysis of paired sets of variables. Since its proposition,
canonical correlation analysis has for instance been extended to extract
relations between two sets of variables when the sample size is insufficient in
relation to the data dimensionality, when the relations have been considered to
be non-linear, and when the dimensionality is too large for human
interpretation. This tutorial explains the theory of canonical correlation
analysis including its regularised, kernel, and sparse variants. Additionally,
the deep and Bayesian CCA extensions are briefly reviewed. Together with the
numerical examples, this overview provides a coherent compendium on the
applicability of the variants of canonical correlation analysis. By bringing
together techniques for solving the optimisation problems, evaluating the
statistical significance and generalisability of the canonical correlation
model, and interpreting the relations, we hope that this article can serve as a
hands-on tool for applying canonical correlation methods in data analysis. | [
1,
0,
0,
1,
0,
0
] |
Title: A note on $p^λ$-convex set in a complete Riemannian manifold,
Abstract: In this paper we have generalized the notion of $\lambda$-radial contraction
in complete Riemannian manifold and developed the concept of $p^\lambda$-convex
function. We have also given a counter example proving the fact that in general
$\lambda$-radial contraction of a geodesic is not necessarily a geodesic. We
have also deduced some relations between geodesic convex sets and
$p^\lambda$-convex sets and showed that under certain conditions they are
equivalent. | [
0,
0,
1,
0,
0,
0
] |
Title: Transpiling Programmable Computable Functions to Answer Set Programs,
Abstract: Programming Computable Functions (PCF) is a simplified programming language
which provides the theoretical basis of modern functional programming
languages. Answer set programming (ASP) is a programming paradigm focused on
solving search problems. In this paper we provide a translation from PCF to
ASP. Using this translation it becomes possible to specify search problems
using PCF. | [
1,
0,
0,
0,
0,
0
] |
Title: Scattering polarization of the $d$-states of ions and solar magnetic field: Effects of isotropic collisions,
Abstract: Analysis of solar magnetic fields using observations as well as theoretical
interpretations of the scattering polarization is commonly designated as a high
priority area of the solar research. The interpretation of the observed
polarization raises a serious theoretical challenge to the researchers involved
in this field. In fact, realistic interpretations need detailed investigations
of the depolarizing role of isotropic collisions with neutral hydrogen. The
goal of this paper is to determine new relationships which allow the
calculation of any collisional rates of the d-levels of ions by simply
determining the value of n^* and $E_p$ without the need of determining the
interaction potentials and treating the dynamics of collisions. The
determination of n^* and E_p is easy and based on atomic data usually available
online. Accurate collisional rates allow a reliable diagnostics of solar
magnetic fields. In this work we applied our collisional FORTRAN code to a
large number of cases involving complex and simple ions. After that, the
results are utilized and injected in a genetic programming code developed with
C-langugae in order to infer original relationships which will be of great help
to solar applications. We discussed the accurarcy of our collisional rates in
the cases of polarized complex atoms and atoms with hyperfine structure. The
relationships are expressed on the tensorial basis and we explain how to
include their contributions in the master equation giving the variation of the
density matrix elements. As a test, we compared the results obtained through
the general relationships provided in this work with the results obtained
directly by running our code of collisions. These comparisons show a percentage
of error of about 10% in the average value. | [
0,
1,
0,
0,
0,
0
] |
Title: Persistence Codebooks for Topological Data Analysis,
Abstract: Topological data analysis, such as persistent homology has shown beneficial
properties for machine learning in many tasks. Topological representations,
such as the persistence diagram (PD), however, have a complex structure
(multiset of intervals) which makes it difficult to combine with typical
machine learning workflows. We present novel compact fixed-size vectorial
representations of PDs based on clustering and bag of words encodings that cope
well with the inherent sparsity of PDs. Our novel representations outperform
state-of-the-art approaches from topological data analysis and are
computationally more efficient. | [
0,
0,
0,
1,
0,
0
] |
Title: Tonic activation of extrasynaptic NMDA receptors decreases intrinsic excitability and promotes bistability in a model of neuronal activity,
Abstract: NMDA receptors (NMDA-R) typically contribute to excitatory synaptic
transmission in the central nervous system. While calcium influx through NMDA-R
plays a critical role in synaptic plasticity, indirect experimental evidence
also exists demonstrating actions of NMDAR-mediated calcium influx on neuronal
excitability through the activation of calcium-activated potassium channels.
But, so far, this mechanism has not been studied theoretically. Our theoretical
model provide a simple description of neuronal electrical activity including
the tonic activity of NMDA receptors and a cytosolic calcium compartment. We
show that calcium influx through NMDA-R can directly be coupled to activation
of calcium-activated potassium channels providing an overall inhibitory effect
on neuronal excitability. Furthermore, the presence of tonic NMDA-R activity
promotes bistability in electrical activity by dramatically increasing the
stimulus interval where both a stable steady state and repetitive firing can
exist. This results could provide an intrinsic mechanism for the constitution
of memory traces in neuronal circuits. They also shed light on the way by which
beta-amyloids can decrease neuronal activity when interfering with NMDA-R in
Alzheimer's disease. | [
0,
0,
0,
0,
1,
0
] |
Title: Using Minimum Path Cover to Boost Dynamic Programming on DAGs: Co-Linear Chaining Extended,
Abstract: Aligning sequencing reads on graph representations of genomes is an important
ingredient of pan-genomics. Such approaches typically find a set of local
anchors that indicate plausible matches between substrings of a read to
subpaths of the graph. These anchor matches are then combined to form a
(semi-local) alignment of the complete read on a subpath. Co-linear chaining is
an algorithmically rigorous approach to combine the anchors. It is a well-known
approach for the case of two sequences as inputs. Here we extend the approach
so that one of the inputs can be a directed acyclic graph (DAGs), e.g. a
splicing graph in transcriptomics or a variant graph in pan-genomics.
This extension to DAGs turns out to have a tight connection to the minimum
path cover problem, asking for a minimum-cardinality set of paths that cover
all the nodes of a DAG. We study the case when the size $k$ of a minimum path
cover is small, which is often the case in practice. First, we propose an
algorithm for finding a minimum path cover of a DAG $(V,E)$ in $O(k|E|\log|V|)$
time, improving all known time-bounds when $k$ is small and the DAG is not too
dense. Second, we introduce a general technique for extending dynamic
programming (DP) algorithms from sequences to DAGs. This is enabled by our
minimum path cover algorithm, and works by mimicking the DP algorithm for
sequences on each path of the minimum path cover. This technique generally
produces algorithms that are slower than their counterparts on sequences only
by a factor $k$. Our technique can be applied, for example, to the classical
longest increasing subsequence and longest common subsequence problems,
extended to labeled DAGs. Finally, we apply this technique to the co-linear
chaining problem. We also implemented the new co-linear chaining approach.
Experiments on splicing graphs show that the new method is efficient also in
practice. | [
1,
0,
0,
0,
0,
0
] |
Title: Steady States of Rotating Stars and Galaxies,
Abstract: A rotating continuum of particles attracted to each other by gravity may be
modeled by the Euler-Poisson system. The existence of solutions is a very
classical problem. Here it is proven that a curve of solutions exists,
parametrized by the rotation speed, with a fixed mass independent of the speed.
The rotation is allowed to vary with the distance to the axis. A special case
is when the equation of state is $p=\rho^\gamma,\ 6/5<\gamma<2,\ \gamma\ne4/3$,
in contrast to previous variational methods which have required $4/3 < \gamma$.
The continuum of particles may alternatively be modeled microscopically by
the Vlasov-Poisson system. The kinetic density is a prescribed function. We
prove an analogous theorem asserting the existence of a curve of solutions with
constant mass. In this model the whole range $(6/5,2)$ is allowed, including
$\gamma=4/3$. | [
0,
0,
1,
0,
0,
0
] |
Title: Critical properties of the contact process with quenched dilution,
Abstract: We have studied the critical properties of the contact process on a square
lattice with quenched site dilution by Monte Carlo simulations. This was
achieved by generating in advance the percolating cluster, through the use of
an appropriate epidemic model, and then by the simulation of the contact
process on the top of the percolating cluster. The dynamic critical exponents
were calculated by assuming an activated scaling relation and the static
exponents by the usual power law behavior. Our results are in agreement with
the prediction that the quenched diluted contact process belongs to the
universality class of the random transverse-field Ising model. We have also
analyzed the model and determined the phase diagram by the use of a mean-field
theory that takes into account the correlation between neighboring sites. | [
0,
1,
0,
0,
0,
0
] |
Title: On Data-Dependent Random Features for Improved Generalization in Supervised Learning,
Abstract: The randomized-feature approach has been successfully employed in large-scale
kernel approximation and supervised learning. The distribution from which the
random features are drawn impacts the number of features required to
efficiently perform a learning task. Recently, it has been shown that employing
data-dependent randomization improves the performance in terms of the required
number of random features. In this paper, we are concerned with the
randomized-feature approach in supervised learning for good generalizability.
We propose the Energy-based Exploration of Random Features (EERF) algorithm
based on a data-dependent score function that explores the set of possible
features and exploits the promising regions. We prove that the proposed score
function with high probability recovers the spectrum of the best fit within the
model class. Our empirical results on several benchmark datasets further verify
that our method requires smaller number of random features to achieve a certain
generalization error compared to the state-of-the-art while introducing
negligible pre-processing overhead. EERF can be implemented in a few lines of
code and requires no additional tuning parameters. | [
1,
0,
0,
1,
0,
0
] |
Title: Anomalous electron spectrum and its relation to peak structure of electron scattering rate in cuprate superconductors,
Abstract: The recent discovery of a direct link between the sharp peak in the electron
quasiparticle scattering rate of cuprate superconductors and the well-known
peak-dip-hump structure in the electron quasiparticle excitation spectrum is
calling for an explanation. Within the framework of the kinetic-energy driven
superconducting mechanism, the complicated line-shape in the electron
quasiparticle excitation spectrum of cuprate superconductors is investigated.
It is shown that the interaction between electrons by the exchange of spin
excitations generates a notable peak structure in the electron quasiparticle
scattering rate around the antinodal and nodal regions. However, this peak
structure disappears at the hot spots, which leads to that the striking
peak-dip-hump structure is developed around the antinodal and nodal regions,
and vanishes at the hot spots. The theory also confirms that the sharp peak
observed in the electron quasiparticle scattering rate is directly responsible
for the remarkable peak-dip-hump structure in the electron quasiparticle
excitation spectrum of cuprate superconductors. | [
0,
1,
0,
0,
0,
0
] |
Title: A comparison theorem for MW-motivic cohomology,
Abstract: We prove that for a finitely generated field over an infinite perfect field
k, and for any integer n, the (n,n)-th MW-motivic cohomology group identifies
with the n-th Milnor-Witt K-theory group of that field | [
0,
0,
1,
0,
0,
0
] |
Title: Wiener Filtering for Passive Linear Quantum Systems,
Abstract: This paper considers a version of the Wiener filtering problem for
equalization of passive quantum linear quantum systems. We demonstrate that
taking into consideration the quantum nature of the signals involved leads to
features typically not encountered in classical equalization problems. Most
significantly, finding a mean-square optimal quantum equalizing filter amounts
to solving a nonconvex constrained optimization problem. We discuss two
approaches to solving this problem, both involving a relaxation of the
constraint. In both cases, unlike classical equalization, there is a threshold
on the variance of the noise below which an improvement of the mean-square
error cannot be guaranteed. | [
1,
0,
0,
0,
0,
0
] |
Title: Density Estimation with Contaminated Data: Minimax Rates and Theory of Adaptation,
Abstract: This paper studies density estimation under pointwise loss in the setting of
contamination model. The goal is to estimate $f(x_0)$ at some
$x_0\in\mathbb{R}$ with i.i.d. observations, $$ X_1,\dots,X_n\sim
(1-\epsilon)f+\epsilon g, $$ where $g$ stands for a contamination distribution.
In the context of multiple testing, this can be interpreted as estimating the
null density at a point. We carefully study the effect of contamination on
estimation through the following model indices: contamination proportion
$\epsilon$, smoothness of target density $\beta_0$, smoothness of contamination
density $\beta_1$, and level of contamination $m$ at the point to be estimated,
i.e. $g(x_0)\leq m$. It is shown that the minimax rate with respect to the
squared error loss is of order $$
[n^{-\frac{2\beta_0}{2\beta_0+1}}]\vee[\epsilon^2(1\wedge
m)^2]\vee[n^{-\frac{2\beta_1}{2\beta_1+1}}\epsilon^{\frac{2}{2\beta_1+1}}], $$
which characterizes the exact influence of contamination on the difficulty of
the problem. We then establish the minimal cost of adaptation to contamination
proportion, to smoothness and to both of the numbers. It is shown that some
small price needs to be paid for adaptation in any of the three cases.
Variations of Lepski's method are considered to achieve optimal adaptation.
The problem is also studied when there is no smoothness assumption on the
contamination distribution. This setting that allows for an arbitrary
contamination distribution is recognized as Huber's $\epsilon$-contamination
model. The minimax rate is shown to be $$
[n^{-\frac{2\beta_0}{2\beta_0+1}}]\vee [\epsilon^{\frac{2\beta_0}{\beta_0+1}}].
$$ The adaptation theory is also different from the smooth contamination case.
While adaptation to either contamination proportion or smoothness only costs a
logarithmic factor, adaptation to both numbers is proved to be impossible. | [
0,
0,
1,
1,
0,
0
] |
Title: Composable Deep Reinforcement Learning for Robotic Manipulation,
Abstract: Model-free deep reinforcement learning has been shown to exhibit good
performance in domains ranging from video games to simulated robotic
manipulation and locomotion. However, model-free methods are known to perform
poorly when the interaction time with the environment is limited, as is the
case for most real-world robotic tasks. In this paper, we study how maximum
entropy policies trained using soft Q-learning can be applied to real-world
robotic manipulation. The application of this method to real-world manipulation
is facilitated by two important features of soft Q-learning. First, soft
Q-learning can learn multimodal exploration strategies by learning policies
represented by expressive energy-based models. Second, we show that policies
learned with soft Q-learning can be composed to create new policies, and that
the optimality of the resulting policy can be bounded in terms of the
divergence between the composed policies. This compositionality provides an
especially valuable tool for real-world manipulation, where constructing new
policies by composing existing skills can provide a large gain in efficiency
over training from scratch. Our experimental evaluation demonstrates that soft
Q-learning is substantially more sample efficient than prior model-free deep
reinforcement learning methods, and that compositionality can be performed for
both simulated and real-world tasks. | [
1,
0,
0,
1,
0,
0
] |
Title: Stimulated Raman Scattering Imposes Fundamental Limits to the Duration and Bandwidth of Temporal Cavity Solitons,
Abstract: Temporal cavity solitons (CS) are optical pulses that can persist in passive
resonators, and they play a key role in the generation of coherent
microresonator frequency combs. In resonators made of amorphous materials, such
as fused silica, they can exhibit a spectral red-shift due to stimulated Raman
scattering. Here we show that this Raman-induced self-frequency-shift imposes a
fundamental limit on the duration and bandwidth of temporal CSs. Specifically,
we theoretically predict that stimulated Raman scattering introduces a
previously unidentified Hopf bifurcation that leads to destabilization of CSs
at large pump-cavity detunings, limiting the range of detunings over which they
can exist. We have confirmed our theoretical predictions by performing
extensive experiments in several different synchronously-driven fiber ring
resonators, obtaining results in excellent agreement with numerical
simulations. Our results could have significant implications for the future
design of Kerr frequency comb systems based on amorphous microresonators. | [
0,
1,
0,
0,
0,
0
] |
Title: A Unifying Framework for Convergence Analysis of Approximate Newton Methods,
Abstract: Many machine learning models are reformulated as optimization problems. Thus,
it is important to solve a large-scale optimization problem in big data
applications. Recently, subsampled Newton methods have emerged to attract much
attention for optimization due to their efficiency at each iteration, rectified
a weakness in the ordinary Newton method of suffering a high cost in each
iteration while commanding a high convergence rate. Other efficient stochastic
second order methods are also proposed. However, the convergence properties of
these methods are still not well understood. There are also several important
gaps between the current convergence theory and the performance in real
applications. In this paper, we aim to fill these gaps. We propose a unifying
framework to analyze local convergence properties of second order methods.
Based on this framework, our theoretical analysis matches the performance in
real applications. | [
1,
0,
0,
0,
0,
0
] |
Title: A Newman property for BLD-mappings,
Abstract: We define a Newman property for BLD-mappings and study its connections to the
porosity of the branch set in the setting of generalized manifolds equipped
with complete path metrics. | [
0,
0,
1,
0,
0,
0
] |
Title: Fast and Accurate Time Series Classification with WEASEL,
Abstract: Time series (TS) occur in many scientific and commercial applications,
ranging from earth surveillance to industry automation to the smart grids. An
important type of TS analysis is classification, which can, for instance,
improve energy load forecasting in smart grids by detecting the types of
electronic devices based on their energy consumption profiles recorded by
automatic sensors. Such sensor-driven applications are very often characterized
by (a) very long TS and (b) very large TS datasets needing classification.
However, current methods to time series classification (TSC) cannot cope with
such data volumes at acceptable accuracy; they are either scalable but offer
only inferior classification quality, or they achieve state-of-the-art
classification quality but cannot scale to large data volumes.
In this paper, we present WEASEL (Word ExtrAction for time SEries
cLassification), a novel TSC method which is both scalable and accurate. Like
other state-of-the-art TSC methods, WEASEL transforms time series into feature
vectors, using a sliding-window approach, which are then analyzed through a
machine learning classifier. The novelty of WEASEL lies in its specific method
for deriving features, resulting in a much smaller yet much more discriminative
feature set. On the popular UCR benchmark of 85 TS datasets, WEASEL is more
accurate than the best current non-ensemble algorithms at orders-of-magnitude
lower classification and training times, and it is almost as accurate as
ensemble classifiers, whose computational complexity makes them inapplicable
even for mid-size datasets. The outstanding robustness of WEASEL is also
confirmed by experiments on two real smart grid datasets, where it
out-of-the-box achieves almost the same accuracy as highly tuned,
domain-specific methods. | [
1,
0,
0,
1,
0,
0
] |
Title: Stack Overflow Considered Harmful? The Impact of Copy&Paste on Android Application Security,
Abstract: Online programming discussion platforms such as Stack Overflow serve as a
rich source of information for software developers. Available information
include vibrant discussions and oftentimes ready-to-use code snippets.
Anecdotes report that software developers copy and paste code snippets from
those information sources for convenience reasons. Such behavior results in a
constant flow of community-provided code snippets into production software. To
date, the impact of this behaviour on code security is unknown. We answer this
highly important question by quantifying the proliferation of security-related
code snippets from Stack Overflow in Android applications available on Google
Play. Access to the rich source of information available on Stack Overflow
including ready-to-use code snippets provides huge benefits for software
developers. However, when it comes to code security there are some caveats to
bear in mind: Due to the complex nature of code security, it is very difficult
to provide ready-to-use and secure solutions for every problem. Hence,
integrating a security-related code snippet from Stack Overflow into production
software requires caution and expertise. Unsurprisingly, we observed insecure
code snippets being copied into Android applications millions of users install
from Google Play every day. To quantitatively evaluate the extent of this
observation, we scanned Stack Overflow for code snippets and evaluated their
security score using a stochastic gradient descent classifier. In order to
identify code reuse in Android applications, we applied state-of-the-art static
analysis. Our results are alarming: 15.4% of the 1.3 million Android
applications we analyzed, contained security-related code snippets from Stack
Overflow. Out of these 97.9% contain at least one insecure code snippet. | [
1,
0,
0,
0,
0,
0
] |
Title: K-theory of group Banach algebras and Banach property RD,
Abstract: We investigate Banach algebras of convolution operators on the $L^p$ spaces
of a locally compact group, and their K-theory. We show that for a discrete
group, the corresponding K-theory groups depend continuously on $p$ in an
inductive sense. Via a Banach version of property RD, we show that for a large
class of groups, the K-theory groups of the Banach algebras are independent of
$p$. | [
0,
0,
1,
0,
0,
0
] |
Title: Learning from various labeling strategies for suicide-related messages on social media: An experimental study,
Abstract: Suicide is an important but often misunderstood problem, one that researchers
are now seeking to better understand through social media. Due in large part to
the fuzzy nature of what constitutes suicidal risks, most supervised approaches
for learning to automatically detect suicide-related activity in social media
require a great deal of human labor to train. However, humans themselves have
diverse or conflicting views on what constitutes suicidal thoughts. So how to
obtain reliable gold standard labels is fundamentally challenging and, we
hypothesize, depends largely on what is asked of the annotators and what slice
of the data they label. We conducted multiple rounds of data labeling and
collected annotations from crowdsourcing workers and domain experts. We
aggregated the resulting labels in various ways to train a series of supervised
models. Our preliminary evaluations show that using unanimously agreed labels
from multiple annotators is helpful to achieve robust machine models. | [
1,
0,
0,
0,
0,
0
] |
Title: Modeling Information Flow Through Deep Neural Networks,
Abstract: This paper proposes a principled information theoretic analysis of
classification for deep neural network structures, e.g. convolutional neural
networks (CNN). The output of convolutional filters is modeled as a random
variable Y conditioned on the object class C and network filter bank F. The
conditional entropy (CENT) H(Y |C,F) is shown in theory and experiments to be a
highly compact and class-informative code, that can be computed from the filter
outputs throughout an existing CNN and used to obtain higher classification
results than the original CNN itself. Experiments demonstrate the effectiveness
of CENT feature analysis in two separate CNN classification contexts. 1) In the
classification of neurodegeneration due to Alzheimer's disease (AD) and natural
aging from 3D magnetic resonance image (MRI) volumes, 3 CENT features result in
an AUC=94.6% for whole-brain AD classification, the highest reported accuracy
on the public OASIS dataset used and 12% higher than the softmax output of the
original CNN trained for the task. 2) In the context of visual object
classification from 2D photographs, transfer learning based on a small set of
CENT features identified throughout an existing CNN leads to AUC values
comparable to the 1000-feature softmax output of the original network when
classifying previously unseen object categories. The general information
theoretical analysis explains various recent CNN design successes, e.g. densely
connected CNN architectures, and provides insights for future research
directions in deep learning. | [
1,
0,
0,
1,
0,
0
] |
Title: Deep Learning applied to Road Traffic Speed forecasting,
Abstract: In this paper, we propose deep learning architectures (FNN, CNN and LSTM) to
forecast a regression model for time dependent data. These algorithm's are
designed to handle Floating Car Data (FCD) historic speeds to predict road
traffic data. For this we aggregate the speeds into the network inputs in an
innovative way. We compare the RMSE thus obtained with the results of a simpler
physical model, and show that the latter achieves better RMSE accuracy. We also
propose a new indicator, which evaluates the algorithms improvement when
compared to a benchmark prediction. We conclude by questioning the interest of
using deep learning methods for this specific regression task. | [
1,
0,
0,
1,
0,
0
] |
Title: Wait For It: Identifying "On-Hold" Self-Admitted Technical Debt,
Abstract: Self-admitted technical debt refers to situations where a software developer
knows that their current implementation is not optimal and indicates this using
a source code comment. In this work, we hypothesize that it is possible to
develop automated techniques to understand a subset of these comments in more
detail, and to propose tool support that can help developers manage
self-admitted technical debt more effectively. Based on a qualitative study of
335 comments indicating self-admitted technical debt, we first identify one
particular class of debt amenable to automated management: "on-hold"
self-admitted technical debt, i.e., debt which contains a condition to indicate
that a developer is waiting for a certain event or an updated functionality
having been implemented elsewhere. We then design and evaluate an automated
classifier which can automatically identify these "on-hold" instances with a
precision of 0.81 as well as detect the specific conditions that developers are
waiting for. Our work presents a first step towards automated tool support that
is able to indicate when certain instances of self-admitted technical debt are
ready to be addressed. | [
1,
0,
0,
0,
0,
0
] |
Title: Real time observation of granular rock analogue material deformation and failure using nonlinear laser interferometry,
Abstract: A better understanding and anticipation of natural processes such as
landsliding or seismic fault activity requires detailed theoretical and
experimental analysis of rock mechanics and geomaterial dynamics. These last
decades, considerable progress has been made towards understanding deformation
and fracture process in laboratory experiment on granular rock materials, as
the well-known shear banding experiment. One of the reasons for this progress
is the continuous improvement in the instrumental techniques of observation.
But the lack of real time methods does not allow the detection of indicators of
the upcoming fracture process and thus to anticipate the phenomenon. Here, we
have performed uniaxial compression experiments to analyse the response of a
granular rock material sample to different shocks. We use a novel
interferometric laser sensor based on the nonlinear self-mixing interferometry
technique to observe in real time the deformations of the sample and assess its
usefulness as a diagnostic tool for the analysis of geomaterial dynamics. Due
to the high spatial and temporal resolution of this approach, we observe both
vibrations processes in response to a dynamic loading and the onset of failure.
The latter is preceded by a continuous variation of vibration period of the
material. After several shocks, the material response is no longer reversible
and we detect a progressive accumulation of irreversible deformation leading to
the fracture process. We demonstrate that material failure is anticipated by
the critical slowing down of the surface vibrational motion, which may
therefore be envisioned as an early warning signal or predictor to the
macroscopic failure of the sample. The nonlinear self-mixing interferometry
technique is readily extensible to fault propagation measurements. As such, it
opens a new window of observation for the study of geomaterial deformation and
failure. | [
0,
1,
0,
0,
0,
0
] |
Title: CREATE: Multimodal Dataset for Unsupervised Learning, Generative Modeling and Prediction of Sensory Data from a Mobile Robot in Indoor Environments,
Abstract: The CREATE database is composed of 14 hours of multimodal recordings from a
mobile robotic platform based on the iRobot Create. The various sensors cover
vision, audition, motors and proprioception. The dataset has been designed in
the context of a mobile robot that can learn multimodal representations of its
environment, thanks to its ability to navigate the environment. This ability
can also be used to learn the dependencies and relationships between the
different modalities of the robot (e.g. vision, audition), as they reflect both
the external environment and the internal state of the robot. The provided
multimodal dataset is expected to have multiple usages, such as multimodal
unsupervised object learning, multimodal prediction and egomotion/causality
detection. | [
1,
0,
0,
0,
0,
0
] |
Title: Deep learning enhanced mobile-phone microscopy,
Abstract: Mobile-phones have facilitated the creation of field-portable, cost-effective
imaging and sensing technologies that approach laboratory-grade instrument
performance. However, the optical imaging interfaces of mobile-phones are not
designed for microscopy and produce spatial and spectral distortions in imaging
microscopic specimens. Here, we report on the use of deep learning to correct
such distortions introduced by mobile-phone-based microscopes, facilitating the
production of high-resolution, denoised and colour-corrected images, matching
the performance of benchtop microscopes with high-end objective lenses, also
extending their limited depth-of-field. After training a convolutional neural
network, we successfully imaged various samples, including blood smears,
histopathology tissue sections, and parasites, where the recorded images were
highly compressed to ease storage and transmission for telemedicine
applications. This method is applicable to other low-cost, aberrated imaging
systems, and could offer alternatives for costly and bulky microscopes, while
also providing a framework for standardization of optical images for clinical
and biomedical applications. | [
1,
1,
0,
0,
0,
0
] |
Title: Minimal Controllability of Conjunctive Boolean Networks is NP-Complete,
Abstract: Given a conjunctive Boolean network (CBN) with $n$ state-variables, we
consider the problem of finding a minimal set of state-variables to directly
affect with an input so that the resulting conjunctive Boolean control network
(CBCN) is controllable. We give a necessary and sufficient condition for
controllability of a CBCN; an $O(n^2)$-time algorithm for testing
controllability; and prove that nonetheless the minimal controllability problem
for CBNs is NP-hard. | [
1,
0,
1,
0,
0,
0
] |
Title: D-optimal designs for complex Ornstein-Uhlenbeck processes,
Abstract: Complex Ornstein-Uhlenbeck (OU) processes have various applications in
statistical modelling. They play role e.g. in the description of the motion of
a charged test particle in a constant magnetic field or in the study of
rotating waves in time-dependent reaction diffusion systems, whereas Kolmogorov
used such a process to model the so-called Chandler wobble, small deviation in
the Earth's axis of rotation. In these applications parameter estimation and
model fitting is based on discrete observations of the underlying stochastic
process, however, the accuracy of the results strongly depend on the
observation points.
This paper studies the properties of D-optimal designs for estimating the
parameters of a complex OU process with a trend. We show that in contrast with
the case of the classical real OU process, a D-optimal design exists not only
for the trend parameter, but also for joint estimation of the covariance
parameters, moreover, these optimal designs are equidistant. | [
0,
0,
1,
1,
0,
0
] |
Title: Self-adjointness and spectral properties of Dirac operators with magnetic links,
Abstract: We define Dirac operators on $\mathbb{S}^3$ (and $\mathbb{R}^3$) with
magnetic fields supported on smooth, oriented links and prove self-adjointness
of certain (natural) extensions. We then analyze their spectral properties and
show, among other things, that these operators have discrete spectrum. Certain
examples, such as circles in $\mathbb{S}^3$, are investigated in detail and we
compute the dimension of the zero-energy eigenspace. | [
0,
0,
1,
0,
0,
0
] |
Title: Latent tree models,
Abstract: Latent tree models are graphical models defined on trees, in which only a
subset of variables is observed. They were first discussed by Judea Pearl as
tree-decomposable distributions to generalise star-decomposable distributions
such as the latent class model. Latent tree models, or their submodels, are
widely used in: phylogenetic analysis, network tomography, computer vision,
causal modeling, and data clustering. They also contain other well-known
classes of models like hidden Markov models, Brownian motion tree model, the
Ising model on a tree, and many popular models used in phylogenetics. This
article offers a concise introduction to the theory of latent tree models. We
emphasise the role of tree metrics in the structural description of this model
class, in designing learning algorithms, and in understanding fundamental
limits of what and when can be learned. | [
0,
0,
1,
1,
0,
0
] |
Title: Towards a Knowledge Graph based Speech Interface,
Abstract: Applications which use human speech as an input require a speech interface
with high recognition accuracy. The words or phrases in the recognised text are
annotated with a machine-understandable meaning and linked to knowledge graphs
for further processing by the target application. These semantic annotations of
recognised words can be represented as a subject-predicate-object triples which
collectively form a graph often referred to as a knowledge graph. This type of
knowledge representation facilitates to use speech interfaces with any spoken
input application, since the information is represented in logical, semantic
form, retrieving and storing can be followed using any web standard query
languages. In this work, we develop a methodology for linking speech input to
knowledge graphs and study the impact of recognition errors in the overall
process. We show that for a corpus with lower WER, the annotation and linking
of entities to the DBpedia knowledge graph is considerable. DBpedia Spotlight,
a tool to interlink text documents with the linked open data is used to link
the speech recognition output to the DBpedia knowledge graph. Such a
knowledge-based speech recognition interface is useful for applications such as
question answering or spoken dialog systems. | [
1,
0,
0,
0,
0,
0
] |
Title: Gradient Masking Causes CLEVER to Overestimate Adversarial Perturbation Size,
Abstract: A key problem in research on adversarial examples is that vulnerability to
adversarial examples is usually measured by running attack algorithms. Because
the attack algorithms are not optimal, the attack algorithms are prone to
overestimating the size of perturbation needed to fool the target model. In
other words, the attack-based methodology provides an upper-bound on the size
of a perturbation that will fool the model, but security guarantees require a
lower bound. CLEVER is a proposed scoring method to estimate a lower bound.
Unfortunately, an estimate of a bound is not a bound. In this report, we show
that gradient masking, a common problem that causes attack methodologies to
provide only a very loose upper bound, causes CLEVER to overestimate the size
of perturbation needed to fool the model. In other words, CLEVER does not
resolve the key problem with the attack-based methodology, because it fails to
provide a lower bound. | [
0,
0,
0,
1,
0,
0
] |
Title: Quantum capacitance of double-layer graphene,
Abstract: We study the ground-state properties of a double layer graphene system with
the Coulomb interlayer electron-electron interaction modeled within the random
phase approximation. We first obtain an expression of the quantum capacitance
of a two layer system. In addition, we calculate the many-body
exchange-correlation energy and quantum capacitance of the hybrid double layer
graphene system at zero-temperature. We show an enhancement of the majority
density layer thermodynamic density-of-states owing to an increasing interlayer
interaction between two layers near the Dirac point. The quantum capacitance
near the neutrality point behaves like square root of the total density,
$\alpha \sqrt{n}$, where the coefficient $\alpha$ decreases by increasing the
charge density imbalance between two layers. Furthermore, we show that the
quantum capacitance changes linearly by the gate voltage. Our results can be
verified by current experiments. | [
0,
1,
0,
0,
0,
0
] |
Title: Distributed Time Synchronization for Networks with Random Delays and Measurement Noise,
Abstract: In this paper a new distributed asynchronous algorithm is proposed for time
synchronization in networks with random communication delays, measurement noise
and communication dropouts. Three different types of the drift correction
algorithm are introduced, based on different kinds of local time increments.
Under nonrestrictive conditions concerning network properties, it is proved
that all the algorithm types provide convergence in the mean square sense and
with probability one (w.p.1) of the corrected drifts of all the nodes to the
same value (consensus). An estimate of the convergence rate of these algorithms
is derived. For offset correction, a new algorithm is proposed containing a
compensation parameter coping with the influence of random delays and special
terms taking care of the influence of both linearly increasing time and drift
correction. It is proved that the corrected offsets of all the nodes converge
in the mean square sense and w.p.1. An efficient offset correction algorithm
based on consensus on local compensation parameters is also proposed. It is
shown that the overall time synchronization algorithm can also be implemented
as a flooding algorithm with one reference node. It is proved that it is
possible to achieve bounded error between local corrected clocks in the mean
square sense and w.p.1. Simulation results provide an additional practical
insight into the algorithm properties and show its advantage over the existing
methods. | [
1,
0,
0,
0,
0,
0
] |
Title: Controller Synthesis for Discrete-time Hybrid Polynomial Systems via Occupation Measures,
Abstract: We present a novel controller synthesis approach for discrete-time hybrid
polynomial systems, a class of systems that can model a wide variety of
interactions between robots and their environment. The approach is rooted in
recently developed techniques that use occupation measures to formulate the
controller synthesis problem as an infinite-dimensional linear program. The
relaxation of the linear program as a finite-dimensional semidefinite program
can be solved to generate a control law. The approach has several advantages
including that the formulation is convex, that the formulation and the
extracted controllers are simple, and that the computational complexity is
polynomial in the state and control input dimensions. We illustrate our
approach on some robotics examples. | [
1,
0,
0,
0,
0,
0
] |
Title: Semiflat Orbifold Projections,
Abstract: We compute the semiflat positive cone $K_0^{+SF}(A_\theta^\sigma)$ of the
$K_0$-group of the irrational rotation orbifold $A_\theta^\sigma$ under the
noncommutative Fourier transform $\sigma$ and show that it is determined by
classes of positive trace and the vanishing of two topological invariants. The
semiflat orbifold projections are 3-dimensional and come in three basic
topological genera: $(2,0,0)$, $(1,1,2)$, $(0,0,2)$. (A projection is called
semiflat when it has the form $h + \sigma(h)$ where $h$ is a flip-invariant
projection such that $h\sigma(h)=0$.) Among other things, we also show that
every number in $(0,1) \cap (2\mathbb Z + 2\mathbb Z\theta)$ is the trace of a
semiflat projection in $A_\theta$. The noncommutative Fourier transform is the
order 4 automorphism $\sigma: V \to U \to V^{-1}$ (and the flip is $\sigma^2$:
$U \to U^{-1},\ V \to V^{-1}$), where $U,V$ are the canonical unitary
generators of the rotation algebra $A_\theta$ satisfying $VU = e^{2\pi i\theta}
UV$. | [
0,
0,
1,
0,
0,
0
] |
Title: Magnetization process of the S = 1/2 two-leg organic spin-ladder compound BIP-BNO,
Abstract: We have measured the magnetization of the organic compound BIP-BNO
(3,5'-bis(N-tert-butylaminoxyl)-3',5-dibromobiphenyl) up to 76 T where the
magnetization is saturated. The S = 1/2 antiferromagnetic Heisenberg two-leg
spin-ladder model accounts for the obtained experimental data regarding the
magnetization curve, which is clarified using the quantum Monte Carlo method.
The exchange constants on the rung and the side rail of the ladder are
estimated to be J(rung)/kB = 65.7 K and J(leg)/kB = 14.1 K, respectively,
deeply in the strong coupling region: J(rung)/J(leg) > 1. | [
0,
1,
0,
0,
0,
0
] |
Title: Deep Over-sampling Framework for Classifying Imbalanced Data,
Abstract: Class imbalance is a challenging issue in practical classification problems
for deep learning models as well as traditional models. Traditionally
successful countermeasures such as synthetic over-sampling have had limited
success with complex, structured data handled by deep learning models. In this
paper, we propose Deep Over-sampling (DOS), a framework for extending the
synthetic over-sampling method to exploit the deep feature space acquired by a
convolutional neural network (CNN). Its key feature is an explicit, supervised
representation learning, for which the training data presents each raw input
sample with a synthetic embedding target in the deep feature space, which is
sampled from the linear subspace of in-class neighbors. We implement an
iterative process of training the CNN and updating the targets, which induces
smaller in-class variance among the embeddings, to increase the discriminative
power of the deep representation. We present an empirical study using public
benchmarks, which shows that the DOS framework not only counteracts class
imbalance better than the existing method, but also improves the performance of
the CNN in the standard, balanced settings. | [
1,
0,
0,
1,
0,
0
] |
Title: The Double Galaxy Cluster Abell 2465 III. X-ray and Weak-lensing Observations,
Abstract: We report Chandra X-ray observations and optical weak-lensing measurements
from Subaru/Suprime-Cam images of the double galaxy cluster Abell 2465
(z=0.245). The X-ray brightness data are fit to a beta-model to obtain the
radial gas density profiles of the northeast (NE) and southwest (SW)
sub-components, which are seen to differ in structure. We determine core radii,
central temperatures, the gas masses within $r_{500c}$, and the total masses
for the broader NE and sharper SW components assuming hydrostatic equilibrium.
The central entropy of the NE clump is about two times higher than the SW.
Along with its structural properties, this suggests that it has undergone
merging on its own. The weak-lensing analysis gives virial masses for each
substructure, which compare well with earlier dynamical results. The derived
outer mass contours of the SW sub-component from weak lensing are more
irregular and extended than those of the NE. Although there is a weak
enhancement and small offsets between X-ray gas and mass centers from weak
lensing, the lack of large amounts of gas between the two sub-clusters
indicates that Abell 2465 is in a pre-merger state. A dynamical model that is
consistent with the observed cluster data, based on the FLASH program and the
radial infall model, is constructed, where the subclusters currently separated
by ~1.2Mpc are approaching each other at ~2000km/s and will meet in ~0.4Gyr. | [
0,
1,
0,
0,
0,
0
] |
Title: Monte Carlo study of magnetic nanoparticles adsorbed on halloysite $Al_2Si_2O_5(OH)_4$ nanotubes,
Abstract: We study properties of magnetic nanoparticles adsorbed on the halloysite
surface. For that a distinct magnetic Hamiltonian with random distribution of
spins on a cylindrical surface was solved by using a nonequilibrium Monte Carlo
method. The parameters for our simulations: anisotropy constant, nanoparticle
size distribution, saturated magnetization and geometrical parameters of the
halloysite template were taken from recent experiments. We calculate the
hysteresis loops and temperature dependence of the zero field cooling (ZFC)
susceptibility, which maximum determines the blocking temperature. It is shown
that the dipole-dipole interaction between nanoparticles moderately increases
the blocking temperature and weakly increases the coercive force. The obtained
hysteresis loops (e.g., the value of the coercive force) for Ni nanoparticles
are in reasonable agreement with the experimental data. We also discuss the
sensitivity of the hysteresis loops and ZFC susceptibilities to the change of
anisotropy and dipole-dipole interaction, as well as the 3d-shell occupation of
the metallic nanoparticles; in particular we predict larger coercive force for
Fe, than for Ni nanoparticles. | [
0,
1,
0,
0,
0,
0
] |
Title: Conversion of Mersenne Twister to double-precision floating-point numbers,
Abstract: The 32-bit Mersenne Twister generator MT19937 is a widely used random number
generator. To generate numbers with more than 32 bits in bit length, and
particularly when converting into 53-bit double-precision floating-point
numbers in $[0,1)$ in the IEEE 754 format, the typical implementation
concatenates two successive 32-bit integers and divides them by a power of $2$.
In this case, the 32-bit MT19937 is optimized in terms of its equidistribution
properties (the so-called dimension of equidistribution with $v$-bit accuracy)
under the assumption that one will mainly be using 32-bit output values, and
hence the concatenation sometimes degrades the dimension of equidistribution
compared with the simple use of 32-bit outputs. In this paper, we analyze such
phenomena by investigating hidden $\mathbb{F}_2$-linear relations among the
bits of high-dimensional outputs. Accordingly, we report that MT19937 with a
specific lag set fails several statistical tests, such as the overlapping
collision test, matrix rank test, and Hamming independence test. | [
1,
0,
0,
1,
0,
0
] |
Title: Computational Study of Halide Perovskite-Derived A$_2$BX$_6$ Inorganic Compounds: Chemical Trends in Electronic Structure and Structural Stability,
Abstract: The electronic structure and energetic stability of A$_2$BX$_6$ halide
compounds with the cubic and tetragonal variants of the perovskite-derived
K$_2$PtCl$_6$ prototype structure are investigated computationally within the
frameworks of density-functional-theory (DFT) and hybrid (HSE06) functionals.
The HSE06 calculations are undertaken for seven known A$_2$BX$_6$ compounds
with A = K, Rb and Cs, and B = Sn, Pd, Pt, Te, and X = I. Trends in band gaps
and energetic stability are identified, which are explored further employing
DFT calculations over a larger range of chemistries, characterized by A = K,
Rb, Cs, B = Si, Ge, Sn, Pb, Ni, Pd, Pt, Se and Te and X = Cl, Br, I. For the
systems investigated in this work, the band gap increases from iodide to
bromide to chloride. Further, variations in the A site cation influences the
band gap as well as the preferred degree of tetragonal distortion. Smaller A
site cations such as K and Rb favor tetragonal structural distortions,
resulting in a slightly larger band gap. For variations in the B site in the
(Ni, Pd, Pt) group and the (Se, Te) group, the band gap increases with
increasing cation size. However, no observed chemical trend with respect to
cation size for band gap was found for the (Si, Sn, Ge, Pb) group. The findings
in this work provide guidelines for the design of halide A$_2$BX$_6$ compounds
for potential photovoltaic applications. | [
0,
1,
0,
0,
0,
0
] |
Title: A Bootstrap Lasso + Partial Ridge Method to Construct Confidence Intervals for Parameters in High-dimensional Sparse Linear Models,
Abstract: For high-dimensional sparse linear models, how to construct confidence
intervals for coefficients remains a difficult question. The main reason is the
complicated limiting distributions of common estimators such as the Lasso.
Several confidence interval construction methods have been developed, and
Bootstrap Lasso+OLS is notable for its simple technicality, good
interpretability, and comparable performance with other more complicated
methods. However, Bootstrap Lasso+OLS depends on the beta-min assumption, a
theoretic criterion that is often violated in practice. In this paper, we
introduce a new method called Bootstrap Lasso+Partial Ridge (LPR) to relax this
assumption. LPR is a two-stage estimator: first using Lasso to select features
and subsequently using Partial Ridge to refit the coefficients. Simulation
results show that Bootstrap LPR outperforms Bootstrap Lasso+OLS when there
exist small but non-zero coefficients, a common situation violating the
beta-min assumption. For such coefficients, compared to Bootstrap Lasso+OLS,
confidence intervals constructed by Bootstrap LPR have on average 50% larger
coverage probabilities. Bootstrap LPR also has on average 35% shorter
confidence interval lengths than the de-sparsified Lasso methods, regardless of
whether linear models are misspecified. Additionally, we provide theoretical
guarantees of Bootstrap LPR under appropriate conditions and implement it in
the R package "HDCI." | [
0,
0,
0,
1,
0,
0
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.