text
stringlengths 57
2.88k
| labels
sequencelengths 6
6
|
---|---|
Title: On central leaves of Hodge-type Shimura varieties with parahoric level structure,
Abstract: Kisin and Pappas constructed integral models of Hodge-type Shimura varieties
with parahoric level structure at $p>2$, such that the formal neighbourhood of
a mod~$p$ point can be interpreted as a deformation space of $p$-divisible
group with some Tate cycles (generalising Faltings' construction). In this
paper, we study the central leaf and the closed Newton stratum in the formal
neighbourhoods of mod~$p$ points of Kisin-Pappas integral models with parahoric
level structure; namely, we obtain the dimension of central leaves and the
almost product structure of Newton strata. In the case of hyperspecial level
strucure (i.e., in the good reduction case), our main results were already
obtained by Hamacher, and the result of this paper holds for ramified groups as
well. | [
0,
0,
1,
0,
0,
0
] |
Title: The sequential loss of allelic diversity,
Abstract: This paper gives a new flavor of what Peter Jagers and his co-authors call
`the path to extinction'. In a neutral population with constant size $N$, we
assume that each individual at time $0$ carries a distinct type, or allele. We
consider the joint dynamics of these $N$ alleles, for example the dynamics of
their respective frequencies and more plainly the nonincreasing process
counting the number of alleles remaining by time $t$. We call this process the
extinction process. We show that in the Moran model, the extinction process is
distributed as the process counting (in backward time) the number of common
ancestors to the whole population, also known as the block counting process of
the $N$-Kingman coalescent. Stimulated by this result, we investigate: (1)
whether it extends to an identity between the frequencies of blocks in the
Kingman coalescent and the frequencies of alleles in the extinction process,
both evaluated at jump times; (2) whether it extends to the general case of
$\Lambda$-Fleming-Viot processes. | [
0,
0,
0,
0,
1,
0
] |
Title: A Parsimonious Dynamical Model for Structural Learning in the Human Brain,
Abstract: The human brain is capable of diverse feats of intelligence. A particularly
salient example is the ability to deduce structure from time-varying auditory
and visual stimuli, enabling humans to master the rules of language and to
build rich expectations of their physical environment. The broad relevance of
this ability for human cognition motivates the need for a first-principles
model explicating putative mechanisms. Here we propose a general framework for
structural learning in the brain, composed of an evolving, high-dimensional
dynamical system driven by external stimuli or internal processes. We
operationalize the scenario in which humans learn the rules that generate a
sequence of stimuli, rather than the exemplar stimuli themselves. We model
external stimuli as seemingly disordered chaotic time series generated by
complex dynamical systems; the underlying structure being deduced is then that
of the corresponding chaotic attractor. This approach allows us to demonstrate
and theoretically explain the emergence of five distinct phenomena reminiscent
of cognitive functions: (i) learning the structure of a chaotic system purely
from time series, (ii) generating new streams of stimuli from a chaotic system,
(iii) switching stream generation among multiple learned chaotic systems,
either spontaneously or in response to external perturbations, (iv) inferring
missing data from sparse observations of the chaotic system, and (v)
deciphering superimposed input from different chaotic systems. Numerically, we
show that these phenomena emerge naturally from a recurrent neural network of
Erdos-Renyi topology in which the synaptic strengths adapt in a Hebbian-like
manner. Broadly, our work blends chaotic theory and artificial neural networks
to answer the long standing question of how neural systems can learn the
structure underlying temporal sequences of stimuli. | [
0,
0,
0,
0,
1,
0
] |
Title: Estimation of mean residual life,
Abstract: Yang (1978) considered an empirical estimate of the mean residual life
function on a fixed finite interval. She proved it to be strongly uniformly
consistent and (when appropriately standardized) weakly convergent to a
Gaussian process. These results are extended to the whole half line, and the
variance of the the limiting process is studied. Also, nonparametric
simultaneous confidence bands for the mean residual life function are obtained
by transforming the limiting process to Brownian motion. | [
0,
0,
1,
1,
0,
0
] |
Title: The Effects of Protostellar Disk Turbulence on CO Emission Lines: A Comparison Study of Disks with Constant CO Abundance vs. Chemically Evolving Disks,
Abstract: Turbulence is the leading candidate for angular momentum transport in
protoplanetary disks and therefore influences disk lifetimes and planet
formation timescales. However, the turbulent properties of protoplanetary disks
are poorly constrained observationally. Recent studies have found turbulent
speeds smaller than what fully-developed MRI would produce (Flaherty et al.
2015, 2017). However, existing studies assumed a constant CO/H2 ratio of 0.0001
in locations where CO is not frozen-out or photo-dissociated. Our previous
studies of evolving disk chemistry indicate that CO is depleted by
incorporation into complex organic molecules well inside the freeze-out radius
of CO. We consider the effects of this chemical depletion on measurements of
turbulence. Simon et al. (2015) suggested that the ratio of the peak line flux
to the flux at line center of the CO J=3-2 transition is a reasonable
diagnostic of turbulence, so we focus on that metric, while adding some
analysis of the more complex effects on spatial distribution. We simulate the
emission lines of CO based on chemical evolution models presented in Yu et al.
(2016), and find that the peak-to-trough ratio changes as a function of time as
CO is destroyed. Specifically, a CO-depleted disk with high turbulent velocity
mimics the peak-to-trough ratios of a non-CO-depleted disk with lower turbulent
velocity. We suggest that disk observers and modelers take into account the
possibility of CO depletion when using line peak-to-trough ratios to constrain
the degree of turbulence in disks. Assuming that CO/H2 = 0.0001 at all disk
radii can lead to underestimates of turbulent speeds in the disk by at least
0.2 km/s. | [
0,
1,
0,
0,
0,
0
] |
Title: Autonomous Urban Localization and Navigation with Limited Information,
Abstract: Urban environments offer a challenging scenario for autonomous driving.
Globally localizing information, such as a GPS signal, can be unreliable due to
signal shadowing and multipath errors. Detailed a priori maps of the
environment with sufficient information for autonomous navigation typically
require driving the area multiple times to collect large amounts of data,
substantial post-processing on that data to obtain the map, and then
maintaining updates on the map as the environment changes. This paper addresses
the issue of autonomous driving in an urban environment by investigating
algorithms and an architecture to enable fully functional autonomous driving
with limited information. An algorithm to autonomously navigate urban roadways
with little to no reliance on an a priori map or GPS is developed. Localization
is performed with an extended Kalman filter with odometry, compass, and sparse
landmark measurement updates. Navigation is accomplished by a compass-based
navigation control law. Key results from Monte Carlo studies show success rates
of urban navigation under different environmental conditions. Experiments
validate the simulated results and demonstrate that, for given test conditions,
an expected range can be found for a given success rate. | [
1,
0,
0,
0,
0,
0
] |
Title: Iterative Refinement for $\ell_p$-norm Regression,
Abstract: We give improved algorithms for the $\ell_{p}$-regression problem, $\min_{x}
\|x\|_{p}$ such that $A x=b,$ for all $p \in (1,2) \cup (2,\infty).$ Our
algorithms obtain a high accuracy solution in $\tilde{O}_{p}(m^{\frac{|p-2|}{2p
+ |p-2|}}) \le \tilde{O}_{p}(m^{\frac{1}{3}})$ iterations, where each iteration
requires solving an $m \times m$ linear system, $m$ being the dimension of the
ambient space.
By maintaining an approximate inverse of the linear systems that we solve in
each iteration, we give algorithms for solving $\ell_{p}$-regression to $1 /
\text{poly}(n)$ accuracy that run in time $\tilde{O}_p(m^{\max\{\omega,
7/3\}}),$ where $\omega$ is the matrix multiplication constant. For the current
best value of $\omega > 2.37$, we can thus solve $\ell_{p}$ regression as fast
as $\ell_{2}$ regression, for all constant $p$ bounded away from $1.$
Our algorithms can be combined with fast graph Laplacian linear equation
solvers to give minimum $\ell_{p}$-norm flow / voltage solutions to $1 /
\text{poly}(n)$ accuracy on an undirected graph with $m$ edges in
$\tilde{O}_{p}(m^{1 + \frac{|p-2|}{2p + |p-2|}}) \le
\tilde{O}_{p}(m^{\frac{4}{3}})$ time.
For sparse graphs and for matrices with similar dimensions, our iteration
counts and running times improve on the $p$-norm regression algorithm by
[Bubeck-Cohen-Lee-Li STOC`18] and general-purpose convex optimization
algorithms. At the core of our algorithms is an iterative refinement scheme for
$\ell_{p}$-norms, using the smoothed $\ell_{p}$-norms introduced in the work of
Bubeck et al. Given an initial solution, we construct a problem that seeks to
minimize a quadratically-smoothed $\ell_{p}$ norm over a subspace, such that a
crude solution to this problem allows us to improve the initial solution by a
constant factor, leading to algorithms with fast convergence. | [
1,
0,
0,
1,
0,
0
] |
Title: Minimal free resolution of the associated graded ring of certain monomial curves,
Abstract: In this article, we give the explicit minimal free resolution of the
associated graded ring of certain affine monomial curves in affine 4-space
based on the standard basis theory. As a result, we give the minimal graded
free resolution and compute the Hilbert function of the tangent cone of these
families. | [
0,
0,
1,
0,
0,
0
] |
Title: Forbidden triads and Creative Success in Jazz: The Miles Davis Factor,
Abstract: This article argues for the importance of forbidden triads - open triads with
high-weight edges - in predicting success in creative fields. Forbidden triads
had been treated as a residual category beyond closed and open triads, yet I
argue that these structures provide opportunities to combine socially evolved
styles in new ways. Using data on the entire history of recorded jazz from 1896
to 2010, I show that observed collaborations have tolerated the openness of
high weight triads more than expected, observed jazz sessions had more
forbidden triads than expected, and the density of forbidden triads contributed
to the success of recording sessions, measured by the number of record releases
of session material. The article also shows that the sessions of Miles Davis
had received an especially high boost from forbidden triads. | [
0,
1,
0,
1,
0,
0
] |
Title: Towards fully automated protein structure elucidation with NMR spectroscopy,
Abstract: Nuclear magnetic resonance (NMR) spectroscopy is one of the leading
techniques for protein studies. The method features a number of properties,
allowing to explain macromolecular interactions mechanistically and resolve
structures with atomic resolution. However, due to laborious data analysis, a
full potential of NMR spectroscopy remains unexploited. Here we present an
approach aiming at automation of two major bottlenecks in the analysis
pipeline, namely, peak picking and chemical shift assignment. Our approach
combines deep learning, non-parametric models and combinatorial optimization,
and is able to detect signals of interest in a multidimensional NMR data with
high accuracy and match them with atoms in medium-length protein sequences,
which is a preliminary step to solve protein spatial structure. | [
0,
0,
0,
1,
0,
0
] |
Title: On families of fibred knots with equal Seifert forms,
Abstract: For every genus $g\geq 2$, we construct an infinite family of strongly
quasipositive fibred knots having the same Seifert form as the torus knot
$T(2,2g+1)$. In particular, their signatures and four-genera are maximal and
their homological monodromies (hence their Alexander module structures) agree.
On the other hand, the geometric stretching factors are pairwise distinct and
the knots are pairwise not ribbon concordant. | [
0,
0,
1,
0,
0,
0
] |
Title: A Comparison of Spatial-based Targeted Disease Containment Strategies using Mobile Phone Data,
Abstract: Epidemic outbreaks are an important healthcare challenge, especially in
developing countries where they represent one of the major causes of mortality.
Approaches that can rapidly target subpopulations for surveillance and control
are critical for enhancing containment processes during epidemics.
Using a real-world dataset from Ivory Coast, this work presents an attempt to
unveil the socio-geographical heterogeneity of disease transmission dynamics.
By employing a spatially explicit meta-population epidemic model derived from
mobile phone Call Detail Records (CDRs), we investigate how the differences in
mobility patterns may affect the course of a realistic infectious disease
outbreak. We consider different existing measures of the spatial dimension of
human mobility and interactions, and we analyse their relevance in identifying
the highest risk sub-population of individuals, as the best candidates for
isolation countermeasures. The approaches presented in this paper provide
further evidence that mobile phone data can be effectively exploited to
facilitate our understanding of individuals' spatial behaviour and its
relationship with the risk of infectious diseases' contagion. In particular, we
show that CDRs-based indicators of individuals' spatial activities and
interactions hold promise for gaining insight of contagion heterogeneity and
thus for developing containment strategies to support decision-making during
country-level pandemics. | [
1,
1,
0,
0,
0,
0
] |
Title: Attosecond Streaking in the Water Window: A New Regime of Attosecond Pulse Characterization,
Abstract: We report on the first streaking measurement of water-window attosecond
pulses generated via high harmonic generation, driven by sub-2-cycle,
CEP-stable, 1850 nm laser pulses. Both the central photon energy and the energy
bandwidth far exceed what has been demonstrated thus far, warranting the
investigation of the attosecond streaking technique for the soft X-ray regime
and the limits of the FROGCRAB retrieval algorithm under such conditions. We
also discuss the problem of attochirp compensation and issues regarding much
lower photo-ionization cross sections compared with the XUV in addition to the
fact that several shells of target gases are accessed simultaneously. Based on
our investigation, we caution that the vastly different conditions in the soft
X-ray regime warrant a diligent examination of the fidelity of the measurement
and the retrieval procedure. | [
0,
1,
0,
0,
0,
0
] |
Title: Interplay of Fluorescence and Phosphorescence in Organic Biluminescent Emitters,
Abstract: Biluminescent organic emitters show simultaneous fluorescence and
phosphorescence at room temperature. So far, the optimization of the room
temperature phosphorescence (RTP) in these materials has drawn the attention of
research. However, the continuous wave operation of these emitters will
consequently turn them into systems with vastly imbalanced singlet and triplet
populations, which is due to the respective excited state lifetimes. This study
reports on the exciton dynamics of the biluminophore NPB
(N,N-di(1-naphthyl)-N,N-diphenyl-(1,1-biphenyl)-4,4-diamine). In the extreme
case, the singlet and triplet exciton lifetimes stretch from 3 ns to 300 ms,
respectively. Through sample engineering and oxygen quenching experiments, the
triplet exciton density can be controlled over several orders of magnitude
allowing to studying exciton interactions between singlet and triplet
manifolds. The results show, that singlet-triplet annihilation reduces the
overall biluminescence efficiency already at moderate excitation levels.
Additionally, the presented system represents an illustrative role model to
study excitonic effects in organic materials. | [
0,
1,
0,
0,
0,
0
] |
Title: Planning Hybrid Driving-Stepping Locomotion on Multiple Levels of Abstraction,
Abstract: Navigating in search and rescue environments is challenging, since a variety
of terrains has to be considered. Hybrid driving-stepping locomotion, as
provided by our robot Momaro, is a promising approach. Similar to other
locomotion methods, it incorporates many degrees of freedom---offering high
flexibility but making planning computationally expensive for larger
environments.
We propose a navigation planning method, which unifies different levels of
representation in a single planner. In the vicinity of the robot, it provides
plans with a fine resolution and a high robot state dimensionality. With
increasing distance from the robot, plans become coarser and the robot state
dimensionality decreases. We compensate this loss of information by enriching
coarser representations with additional semantics. Experiments show that the
proposed planner provides plans for large, challenging scenarios in feasible
time. | [
1,
0,
0,
0,
0,
0
] |
Title: Blackbody Radiation in Classical Physics: A Historical Perspective,
Abstract: We point out that current textbooks of modern physics are a century
out-of-date in their treatment of blackbody radiation within classical physics.
Relativistic classical electrodynamics including classical electromagnetic
zero-point radiation gives the Planck spectrum with zero-point radiation as the
blackbody radiation spectrum. In contrast, nonrelativistic mechanics cannot
support the idea of zero-point energy; therefore if nonrelativistic classical
statistical mechanics or nonrelativistic mechanical scatterers are invoked for
radiation equilibrium, one arrives at only the low-frequency Rayleigh-Jeans
part of the spectrum which involves no zero-point energy, and does not include
the high-frequency part of the spectrum involving relativistically-invariant
classical zero-point radiation. Here we first discuss the correct understanding
of blackbody radiation within relativistic classical physics, and then we
review the historical treatment. Finally, we point out how the presence of
Lorentz-invariant classical zero-point radiation and the use of relativistic
particle interactions transform the previous historical arguments so as now to
give the Planck spectrum including classical zero-point radiation. Within
relativistic classical electromagnetic theory, Planck's constant h appears as
the scale of source-free zero-point radiation. | [
0,
1,
0,
0,
0,
0
] |
Title: The effect of stellar and AGN feedback on the low redshift Lyman-$α$ forest in the Sherwood simulation suite,
Abstract: We study the effect of different feedback prescriptions on the properties of
the low redshift ($z\leq1.6$) Ly$\alpha$ forest using a selection of
hydrodynamical simulations drawn from the Sherwood simulation suite. The
simulations incorporate stellar feedback, AGN feedback and a simplified scheme
for efficiently modelling the low column density Ly$\alpha$ forest. We confirm
a discrepancy remains between Cosmic Origins Spectrograph (COS) observations of
the Ly$\alpha$ forest column density distribution function (CDDF) at $z \simeq
0.1$ for high column density systems ($N_{\rm HI}>10^{14}\rm\,cm^{-2}$), as
well as Ly$\alpha$ velocity widths that are too narrow compared to the COS
data. Stellar or AGN feedback -- as currently implemented in our simulations --
have only a small effect on the CDDF and velocity width distribution. We
conclude that resolving the discrepancy between the COS data and simulations
requires an increase in the temperature of overdense gas with $\Delta=4$--$40$,
either through additional He$\,\rm \scriptstyle II\ $ photo-heating at $z>2$ or
fine-tuned feedback that ejects overdense gas into the IGM at just the right
temperature for it to still contribute significantly to the Ly$\alpha$ forest.
Alternatively a larger, currently unresolved turbulent component to the line
width could resolve the discrepancy. | [
0,
1,
0,
0,
0,
0
] |
Title: Exploring a search for long-duration transient gravitational waves associated with magnetar bursts,
Abstract: Soft gamma repeaters and anomalous X-ray pulsars are thought to be magnetars,
neutron stars with strong magnetic fields of order $\mathord{\sim}
10^{13}$--$10^{15} \, \mathrm{gauss}$. These objects emit intermittent bursts
of hard X-rays and soft gamma rays. Quasiperiodic oscillations in the X-ray
tails of giant flares imply the existence of neutron star oscillation modes
which could emit gravitational waves powered by the magnetar's magnetic energy
reservoir. We describe a method to search for transient gravitational-wave
signals associated with magnetar bursts with durations of 10s to 1000s of
seconds. The sensitivity of this method is estimated by adding simulated
waveforms to data from the sixth science run of Laser Interferometer
Gravitational-wave Observatory (LIGO). We find a search sensitivity in terms of
the root sum square strain amplitude of $h_{\mathrm{rss}} = 1.3 \times 10^{-21}
\, \mathrm{Hz}^{-1/2}$ for a half sine-Gaussian waveform with a central
frequency $f_0 = 150 \, \mathrm{Hz}$ and a characteristic time $\tau = 400 \,
\mathrm{s}$. This corresponds to a gravitational wave energy of
$E_{\mathrm{GW}} = 4.3 \times 10^{46} \, \mathrm{erg}$, the same order of
magnitude as the 2004 giant flare which had an estimated electromagnetic energy
of $E_{\mathrm{EM}} = \mathord{\sim} 1.7 \times 10^{46} (d/ 8.7 \,
\mathrm{kpc})^2 \, \mathrm{erg}$, where $d$ is the distance to SGR 1806-20. We
present an extrapolation of these results to Advanced LIGO, estimating a
sensitivity to a gravitational wave energy of $E_{\mathrm{GW}} = 3.2 \times
10^{43} \, \mathrm{erg}$ for a magnetar at a distance of $1.6 \, \mathrm{kpc}$.
These results suggest this search method can probe significantly below the
energy budgets for magnetar burst emission mechanisms such as crust cracking
and hydrodynamic deformation. | [
0,
1,
0,
0,
0,
0
] |
Title: Coupling between a charge density wave and magnetism in an Heusler material,
Abstract: The Prototypical magnetic memory shape alloy Ni$_2$MnGa undergoes various
phase transitions as a function of temperature, pressure, and doping. In the
low-temperature phases below 260 K, an incommensurate structural modulation
occurs along the [110] direction which is thought to arise from softening of a
phonon mode. It is not at present clear how this phenomenon is related, if at
all, to the magnetic memory effect. Here we report time-resolved measurements
which track both the structural and magnetic components of the phase transition
from the modulated cubic phase as it is brought into the high-symmetry phase.
The results suggest that the photoinduced demagnetization modifies the Fermi
surface in regions that couple strongly to the periodicity of the structural
modulation through the nesting vector. The amplitude of the periodic lattice
distortion, however, appears to be less affected by the demagnetizaton. | [
0,
1,
0,
0,
0,
0
] |
Title: Best-Effort FPGA Programming: A Few Steps Can Go a Long Way,
Abstract: FPGA-based heterogeneous architectures provide programmers with the ability
to customize their hardware accelerators for flexible acceleration of many
workloads. Nonetheless, such advantages come at the cost of sacrificing
programmability. FPGA vendors and researchers attempt to improve the
programmability through high-level synthesis (HLS) technologies that can
directly generate hardware circuits from high-level language descriptions.
However, reading through recent publications on FPGA designs using HLS, one
often gets the impression that FPGA programming is still hard in that it leaves
programmers to explore a very large design space with many possible
combinations of HLS optimization strategies.
In this paper we make two important observations and contributions. First, we
demonstrate a rather surprising result: FPGA programming can be made easy by
following a simple best-effort guideline of five refinement steps using HLS. We
show that for a broad class of accelerator benchmarks from MachSuite, the
proposed best-effort guideline improves the FPGA accelerator performance by
42-29,030x. Compared to the baseline CPU performance, the FPGA accelerator
performance is improved from an average 292.5x slowdown to an average 34.4x
speedup. Moreover, we show that the refinement steps in the best-effort
guideline, consisting of explicit data caching, customized pipelining,
processing element duplication, computation/communication overlapping and
scratchpad reorganization, correspond well to the best practice guidelines for
multicore CPU programming. Although our best-effort guideline may not always
lead to the optimal solution, it substantially simplifies the FPGA programming
effort, and will greatly support the wide adoption of FPGA-based acceleration
by the software programming community. | [
1,
0,
0,
0,
0,
0
] |
Title: Distributional Adversarial Networks,
Abstract: We propose a framework for adversarial training that relies on a sample
rather than a single sample point as the fundamental unit of discrimination.
Inspired by discrepancy measures and two-sample tests between probability
distributions, we propose two such distributional adversaries that operate and
predict on samples, and show how they can be easily implemented on top of
existing models. Various experimental results show that generators trained with
our distributional adversaries are much more stable and are remarkably less
prone to mode collapse than traditional models trained with pointwise
prediction discriminators. The application of our framework to domain
adaptation also results in considerable improvement over recent
state-of-the-art. | [
1,
0,
0,
0,
0,
0
] |
Title: Achieving non-discrimination in prediction,
Abstract: Discrimination-aware classification is receiving an increasing attention in
data science fields. The pre-process methods for constructing a
discrimination-free classifier first remove discrimination from the training
data, and then learn the classifier from the cleaned data. However, they lack a
theoretical guarantee for the potential discrimination when the classifier is
deployed for prediction. In this paper, we fill this gap by mathematically
bounding the probability of the discrimination in prediction being within a
given interval in terms of the training data and classifier. We adopt the
causal model for modeling the data generation mechanism, and formally defining
discrimination in population, in a dataset, and in prediction. We obtain two
important theoretical results: (1) the discrimination in prediction can still
exist even if the discrimination in the training data is completely removed;
and (2) not all pre-process methods can ensure non-discrimination in prediction
even though they can achieve non-discrimination in the modified training data.
Based on the results, we develop a two-phase framework for constructing a
discrimination-free classifier with a theoretical guarantee. The experiments
demonstrate the theoretical results and show the effectiveness of our two-phase
framework. | [
1,
0,
0,
1,
0,
0
] |
Title: Undersampled dynamic X-ray tomography with dimension reduction Kalman filter,
Abstract: In this paper, we consider prior-based dimension reduction Kalman filter for
undersampled dynamic X-ray tomography. With this method, the X-ray
reconstructions are parameterized by a low-dimensional basis. Thus, the
proposed method is a) computationally very light; and b) extremely robust as
all the computations can be done explicitly. With real and simulated
measurement data, we show that the method provides accurate reconstructions
even with very limited number of angular directions. | [
0,
0,
0,
1,
0,
0
] |
Title: Extended Sammon Projection and Wavelet Kernel Extreme Learning Machine for Gait-Based Legitimate User Identification on Smartphones,
Abstract: Smartphones have ubiquitously integrated into our home and work environments,
however, users normally rely on explicit but inefficient identification
processes in a controlled environment. Therefore, when a device is stolen, a
thief can have access to the owner's personal information and services against
the stored password/s. As a result of this potential scenario, this work
demonstrates the possibilities of legitimate user identification in a
semi-controlled environment through the built-in smartphones motion dynamics
captured by two different sensors. This is a two-fold process: sub-activity
recognition followed by user/impostor identification. Prior to the
identification; Extended Sammon Projection (ESP) method is used to reduce the
redundancy among the features. To validate the proposed system, we first
collected data from four users walking with their device freely placed in one
of their pants pockets. Through extensive experimentation, we demonstrate that
together time and frequency domain features optimized by ESP to train the
wavelet kernel based extreme learning machine classifier is an effective system
to identify the legitimate user or an impostor with \(97\%\) accuracy. | [
1,
0,
0,
0,
0,
0
] |
Title: On the free path length distribution for linear motion in an n-dimensional box,
Abstract: We consider the distribution of free path lengths, or the distance between
consecutive bounces of random particles, in an n-dimensional rectangular box.
If each particle travels a distance R, then, as R tends to infinity the free
path lengths coincides with the distribution of the length of the intersection
of a random line with the box (for a natural ensemble of random lines) and we
give an explicit formula (piecewise real analytic) for the probability density
function in dimension two and three.
In dimension two we also consider a closely related model where each particle
is allowed to bounce N times, as N tends to infinity, and give an explicit
(again piecewise real analytic) formula for its probability density function.
Further, in both models we can recover the side lengths of the box from the
location of the discontinuities of the probability density functions. | [
0,
0,
1,
0,
0,
0
] |
Title: Spin-Orbit Misalignments of Three Jovian Planets via Doppler Tomography,
Abstract: We present measurements of the spin-orbit misalignments of the hot Jupiters
HAT-P-41 b and WASP-79 b, and the aligned warm Jupiter Kepler-448 b. We
obtained these measurements with Doppler tomography, where we spectroscopically
resolve the line profile perturbation during the transit due to the
Rossiter-McLaughlin effect. We analyze time series spectra obtained during
portions of five transits of HAT-P-41 b, and find a value of the spin-orbit
misalignment of $\lambda = -22.1_{-6.0}^{+0.8 \circ}$. We reanalyze the radial
velocity Rossiter-McLaughlin data on WASP-79 b obtained by Addison et al.
(2013) using Doppler tomographic methodology. We measure
$\lambda=-99.1_{-3.9}^{+4.1\circ}$, consistent with but more precise than the
value found by Addison et al. (2013). For Kepler-448 b we perform a joint fit
to the Kepler light curve, Doppler tomographic data, and a radial velocity
dataset from Lillo-Box et al. (2015). We find an approximately aligned orbit
($\lambda=-7.1^{+4.2 \circ}_{-2.8}$), in modest disagreement with the value
found by Bourrier et al. (2015). Through analysis of the Kepler light curve we
measure a stellar rotation period of $P_{\mathrm{rot}}=1.27 \pm 0.11$ days, and
use this to argue that the full three-dimensional spin-orbit misalignment is
small, $\psi\sim0^{\circ}$. | [
0,
1,
0,
0,
0,
0
] |
Title: Nauticle: a general-purpose particle-based simulation tool,
Abstract: Nauticle is a general-purpose simulation tool for the flexible and highly
configurable application of particle-based methods of either discrete or
continuum phenomena. It is presented that Nauticle has three distinct layers
for users and developers, then the top two layers are discussed in detail. The
paper introduces the Symbolic Form Language (SFL) of Nauticle, which
facilitates the formulation of user-defined numerical models at the top level
in text-based configuration files and provides simple application examples of
use. On the other hand, at the intermediate level, it is shown that the SFL can
be intuitively extended with new particle methods without tedious recoding or
even the knowledge of the bottom level. Finally, the efficiency of the code is
also tested through a performance benchmark. | [
1,
1,
0,
0,
0,
0
] |
Title: Microwave SQUID Multiplexer demonstration for Cosmic Microwave Background Imagers,
Abstract: Key performance characteristics are demonstrated for the microwave SQUID
multiplexer ($\mu$MUX) coupled to transition edge sensor (TES) bolometers that
have been optimized for cosmic microwave background (CMB) observations. In a
64-channel demonstration, we show that the $\mu$MUX produces a white, input
referred current noise level of 29~pA$/\sqrt{\mathrm{Hz}}$ at -77~dB microwave
probe tone power, which is well below expected fundamental detector and photon
noise sources for a ground-based CMB-optimized bolometer. Operated with
negligible photon loading, we measure 98~pA$/\sqrt{\mathrm{Hz}}$ in the
TES-coupled channels biased at 65% of the sensor normal resistance. This noise
level is consistent with that predicted from bolometer thermal fluctuation
(i.e., phonon) noise. Furthermore, the power spectral density exhibits a white
spectrum at low frequencies ($\sim$~100~mHz), which enables CMB mapping on
large angular scales that constrain the physics of inflation. Additionally, we
report cross-talk measurements that indicate a level below 0.3%, which is less
than the level of cross-talk from multiplexed readout systems in deployed CMB
imagers. These measurements demonstrate the $\mu$MUX as a viable readout
technique for future CMB imaging instruments. | [
0,
1,
0,
0,
0,
0
] |
Title: Preserving Differential Privacy Between Features in Distributed Estimation,
Abstract: Privacy is crucial in many applications of machine learning. Legal, ethical
and societal issues restrict the sharing of sensitive data making it difficult
to learn from datasets that are partitioned between many parties. One important
instance of such a distributed setting arises when information about each
record in the dataset is held by different data owners (the design matrix is
"vertically-partitioned").
In this setting few approaches exist for private data sharing for the
purposes of statistical estimation and the classical setup of differential
privacy with a "trusted curator" preparing the data does not apply. We work
with the notion of $(\epsilon,\delta)$-distributed differential privacy which
extends single-party differential privacy to the distributed,
vertically-partitioned case. We propose PriDE, a scalable framework for
distributed estimation where each party communicates perturbed random
projections of their locally held features ensuring
$(\epsilon,\delta)$-distributed differential privacy is preserved. For
$\ell_2$-penalized supervised learning problems PriDE has bounded estimation
error compared with the optimal estimates obtained without privacy constraints
in the non-distributed setting. We confirm this empirically on real world and
synthetic datasets. | [
1,
0,
0,
1,
0,
0
] |
Title: Parallel implementation of a vehicle rail dynamical model for multi-core systems,
Abstract: This research presents a model of a complex dynamic object running on a
multi-core system. Discretization and numerical integration for multibody
models of vehicle rail elements in the vertical longitudinal plane fluctuations
is considered. The implemented model and solution of the motion differential
equations allow estimating the basic processes occurring in the system with
various external influences. Hence the developed programming model can be used
for performing analysis and comparing new vehicle designs.
Keywords-dynamic model; multi-core system; SMP system; rolling stock. | [
1,
0,
0,
0,
0,
0
] |
Title: A Python Calculator for Supernova Remnant Evolution,
Abstract: A freely available Python code for modelling SNR evolution has been created.
This software is intended for two purposes: to understand SNR evolution; and to
use in modelling observations of SNR for obtaining good estimates of SNR
properties. It includes all phases for the standard path of evolution for
spherically symmetric SNRs. In addition, alternate evolutionary models are
available, including evolution in a cloudy ISM, the fractional energy loss
model, and evolution in a hot low-density ISM. The graphical interface takes in
various parameters and produces outputs such as shock radius and velocity vs.
time, SNR surface brightness profile and spectrum. Some interesting properties
of SNR evolution are demonstrated using the program. | [
0,
1,
0,
0,
0,
0
] |
Title: Über die Präzision interprozeduraler Analysen,
Abstract: In this work, we examine two approaches to interprocedural data-flow analysis
of Sharir and Pnueli in terms of precision: the functional and the call-string
approach. In doing so, not only the theoretical best, but all solutions are
regarded which occur when using abstract interpretation or widening
additionally. It turns out that the solutions of both approaches coincide. This
property is preserved when using abstract interpretation; in the case of
widening, a comparison of the results is not always possible. | [
1,
0,
0,
0,
0,
0
] |
Title: Cancellation theorem for Grothendieck-Witt-correspondences and Witt-correspondences,
Abstract: The cancellation theorem for Grothendieck-Witt-correspondences and
Witt-correspondences between smooth varieties over an infinite prefect field
$k$, $char k \neq 2$, is proved, the isomorphism
$$Hom_{\mathbf{DM}^\mathrm{GW}_\mathrm{eff}}(A^\bullet,B^\bullet) \simeq
Hom_{\mathbf{DM}^\mathrm{GW}_\mathrm{eff}}(A^\bullet(1),B^\bullet(1)),$$ for
$A^\bullet,B^\bullet\in \mathbf{DM}^\mathrm{GW}_\mathrm{eff}(k)$ in the
category of effective Grothendieck-Witt-motives constructed in
\cite{AD_DMGWeff} is obtained (and similarly for Witt-motives).
This implies that the canonical functor $\Sigma_{\mathbb G_m^{\wedge
1}}^\infty\colon \mathbf{DM}^\mathrm{GW}_\mathrm{eff}(k)\to
\mathbf{DM}^\mathrm{GW}(k)$ is fully faithful, where
$\mathbf{DM}^\mathrm{GW}(k)$ is the category of non-effective GW-motives
(defined by stabilization of $\mathbf{DM}^\mathrm{GW}_\mathrm{eff}(k)$ along
$\mathbb G_m^{\wedge 1}$) and yields the main property of motives of smooth
varieties in the category $\mathbf{DM}^\mathrm{GW}(k)$: $$
Hom_{\mathbf{DM}^\mathrm{GW}(k)}(M^{GW}(X), \Sigma_{\mathbb G_m^{\wedge
1}}^\infty\mathcal F[i]) \simeq H^i_{Nis}(X,\mathcal F) ,$$ for any smooth
variety $X$ and homotopy invariant sheave with GW-transfers $\mathcal F$ (and
similarly for $\mathbf{DM}^\mathrm{W}(k)$). | [
0,
0,
1,
0,
0,
0
] |
Title: Jumping across biomedical contexts using compressive data fusion,
Abstract: Motivation: The rapid growth of diverse biological data allows us to consider
interactions between a variety of objects, such as genes, chemicals, molecular
signatures, diseases, pathways and environmental exposures. Often, any pair of
objects--such as a gene and a disease--can be related in different ways, for
example, directly via gene-disease associations or indirectly via functional
annotations, chemicals and pathways. Different ways of relating these objects
carry different semantic meanings. However, traditional methods disregard these
semantics and thus cannot fully exploit their value in data modeling.
Results: We present Medusa, an approach to detect size-k modules of objects
that, taken together, appear most significant to another set of objects. Medusa
operates on large-scale collections of heterogeneous data sets and explicitly
distinguishes between diverse data semantics. It advances research along two
dimensions: it builds on collective matrix factorization to derive different
semantics, and it formulates the growing of the modules as a submodular
optimization program. Medusa is flexible in choosing or combining semantic
meanings and provides theoretical guarantees about detection quality. In a
systematic study on 310 complex diseases, we show the effectiveness of Medusa
in associating genes with diseases and detecting disease modules. We
demonstrate that in predicting gene-disease associations Medusa compares
favorably to methods that ignore diverse semantic meanings. We find that the
utility of different semantics depends on disease categories and that, overall,
Medusa recovers disease modules more accurately when combining different
semantics. | [
1,
0,
0,
1,
0,
0
] |
Title: Memories of a Theoretical Physicist,
Abstract: While I was dealing with a brain injury and finding it difficult to work, two
friends (Derek Westen, a friend of the KITP, and Steve Shenker, with whom I was
recently collaborating), suggested that a new direction might be good. Steve in
particular regarded me as a good writer and suggested that I try that. I
quickly took to Steve's suggestion. Having only two bodies of knowledge, myself
and physics, I decided to write an autobiography about my development as a
theoretical physicist.
This is not written for any particular audience, but just to give myself a
goal. It will probably have too much physics for a nontechnical reader, and too
little for a physicist, but perhaps there with be different things for each.
Parts may be tedious. But it is somewhat unique, I think, a blow-by-blow
history of where I started and where I got to.
Probably the target audience is theoretical physicists, especially young
ones, who may enjoy comparing my struggles with their own. Some disclaimers:
This is based on my own memories, jogged by the arXiv and Inspire. There will
surely be errors and omissions. And note the title: this is about my memories,
which will be different for other people. Also, it would not be possible for me
to mention all the authors whose work might intersect mine, so this should not
be treated as a reference work. | [
0,
1,
0,
0,
0,
0
] |
Title: Best-Choice Edge Grafting for Efficient Structure Learning of Markov Random Fields,
Abstract: Incremental methods for structure learning of pairwise Markov random fields
(MRFs), such as grafting, improve scalability by avoiding inference over the
entire feature space in each optimization step. Instead, inference is performed
over an incrementally grown active set of features. In this paper, we address
key computational bottlenecks that current incremental techniques still suffer
by introducing best-choice edge grafting, an incremental, structured method
that activates edges as groups of features in a streaming setting. The method
uses a reservoir of edges that satisfy an activation condition, approximating
the search for the optimal edge to activate. It also reorganizes the search
space using search-history and structure heuristics. Experiments show a
significant speedup for structure learning and a controllable trade-off between
the speed and quality of learning. | [
1,
0,
0,
1,
0,
0
] |
Title: Cavity-enhanced transport of charge,
Abstract: We theoretically investigate charge transport through electronic bands of a
mesoscopic one-dimensional system, where inter-band transitions are coupled to
a confined cavity mode, initially prepared close to its vacuum. This coupling
leads to light-matter hybridization where the dressed fermionic bands interact
via absorption and emission of dressed cavity-photons. Using a self-consistent
non-equilibrium Green's function method, we compute electronic transmissions
and cavity photon spectra and demonstrate how light-matter coupling can lead to
an enhancement of charge conductivity in the steady-state. We find that
depending on cavity loss rate, electronic bandwidth, and coupling strength, the
dynamics involves either an individual or a collective response of Bloch
states, and explain how this affects the current enhancement. We show that the
charge conductivity enhancement can reach orders of magnitudes under
experimentally relevant conditions. | [
0,
1,
0,
0,
0,
0
] |
Title: Entropy Production Rate is Maximized in Non-Contractile Actomyosin,
Abstract: The actin cytoskeleton is an active semi-flexible polymer network whose
non-equilibrium properties coordinate both stable and contractile behaviors to
maintain or change cell shape. While myosin motors drive the actin cytoskeleton
out-of-equilibrium, the role of myosin-driven active stresses in the
accumulation and dissipation of mechanical energy is unclear. To investigate
this, we synthesize an actomyosin material in vitro whose active stress content
can tune the network from stable to contractile. Each increment in activity
determines a characteristic spectrum of actin filament fluctuations which is
used to calculate the total mechanical work and the production of entropy in
the material. We find that the balance of work and entropy does not increase
monotonically and, surprisingly, the entropy production rate is maximized in
the non-contractile, stable state. Our study provides evidence that the origins
of system entropy production and activity-dependent dissipation arise from
disorder in the molecular interactions between actin and myosin | [
0,
0,
0,
0,
1,
0
] |
Title: Intrinsic resolving power of XUV diffraction gratings measured with Fizeau interferometry,
Abstract: We introduce a method for using Fizeau interferometry to measure the
intrinsic resolving power of a diffraction grating. This method is more
accurate than traditional techniques based on a long-trace profiler (LTP),
since it is sensitive to long-distance phase errors not revealed by a d-spacing
map. We demonstrate 50,400 resolving power for a mechanically ruled XUV grating
from Inprentus, Inc. | [
0,
1,
0,
0,
0,
0
] |
Title: Random matrix approach for primal-dual portfolio optimization problems,
Abstract: In this paper, we revisit the portfolio optimization problems of the
minimization/maximization of investment risk under constraints of budget and
investment concentration (primal problem) and the maximization/minimization of
investment concentration under constraints of budget and investment risk (dual
problem) for the case that the variances of the return rates of the assets are
identical. We analyze both optimization problems by using the Lagrange
multiplier method and the random matrix approach. Thereafter, we compare the
results obtained from our proposed approach with the results obtained in
previous work. Moreover, we use numerical experiments to validate the results
obtained from the replica approach and the random matrix approach as methods
for analyzing both the primal and dual portfolio optimization problems. | [
1,
1,
0,
0,
0,
0
] |
Title: Insight into the modeling of seismic waves for detection of underground cavities,
Abstract: Motivated by the need to detect an underground cavity within the procedure of
an On-Site-Inspection (OSI), of the Comprehensive Nuclear Test Ban Treaty
Organization, the aim of this paper is to present results on the comparison of
our numerical simulations with an analytic solution. The accurate numerical
modeling can facilitate the development of proper analysis techniques to detect
the remnants of an underground nuclear test. The larger goal is to help set a
rigorous scientific base of OSI and to contribute to bringing the Treaty into
force. For our 3D numerical simulations, we use the discontinuous Galerkin
Spectral Element Code SPEED jointly developed at MOX (The Laboratory for
Modeling and Scientific Computing, Department of Mathematics) and at DICA
(Department of Civil and Environmental Engineering) of the Politecnico di
Milano. | [
0,
1,
0,
0,
0,
0
] |
Title: How to centralize and normalize quandle extensions,
Abstract: We show that quandle coverings in the sense of Eisermann form a (regular
epi)-reflective subcategory of the category of surjective quandle
homomorphisms, both by using arguments coming from categorical Galois theory
and by constructing concretely a centralization congruence. Moreover, we show
that a similar result holds for normal quandle extensions. | [
0,
0,
1,
0,
0,
0
] |
Title: Geometric Fluctuation Theorem,
Abstract: We derive an extended fluctuation theorem for a geometric pumping in a
spin-boson system under a periodic control of environmental temperatures by
using a Markovian quantum master equation. We perform the Monte-Carlo
simulation and obtain the current distribution, the average current and the
fluctuation. Using the extended fluctuation theorem we try to explain the
results of our simulation. The fluctuation theorem leads to the fluctuation
dissipation relations but the absence of the conventional reciprocal relation. | [
0,
1,
0,
0,
0,
0
] |
Title: Unsupervised Domain Adaptation Based on Source-guided Discrepancy,
Abstract: Unsupervised domain adaptation is the problem setting where data generating
distributions in the source and target domains are different, and labels in the
target domain are unavailable. One important question in unsupervised domain
adaptation is how to measure the difference between the source and target
domains. A previously proposed discrepancy that does not use the source domain
labels requires high computational cost to estimate and may lead to a loose
generalization error bound in the target domain. To mitigate these problems, we
propose a novel discrepancy called source-guided discrepancy (S-disc), which
exploits labels in the source domain. As a consequence, S-disc can be computed
efficiently with a finite sample convergence guarantee. In addition, we show
that S-disc can provide a tighter generalization error bound than the one based
on an existing discrepancy. Finally, we report experimental results that
demonstrate the advantages of S-disc over the existing discrepancies. | [
0,
0,
0,
1,
0,
0
] |
Title: On structured surfaces with defects: geometry, strain incompatibility, internal stress, and natural shapes,
Abstract: Given a distribution of defects on a structured surface, such as those
represented by 2-dimensional crystalline materials, liquid crystalline
surfaces, and thin sandwiched shells, what is the resulting stress field and
the deformed shape? Motivated by this concern, we first classify, and quantify,
the translational, rotational, and metrical defects allowable over a broad
class of structured surfaces. With an appropriate notion of strain, the defect
densities are then shown to appear as sources of strain incompatibility. The
strain incompatibility relations, with appropriate kinematical assumptions on
the decomposition of strain into elastic and plastic parts, and the stress
equilibrium relations, with a suitable choice of material response, provide the
necessary equations for determining both the internal stress field and the
deformed shape. We demonstrate this by applying our theory to Kirchhoff-Love
shells with a kinematics which allows for small in-surface strains but
moderately large rotations. | [
0,
1,
1,
0,
0,
0
] |
Title: Fatiguing STDP: Learning from Spike-Timing Codes in the Presence of Rate Codes,
Abstract: Spiking neural networks (SNNs) could play a key role in unsupervised machine
learning applications, by virtue of strengths related to learning from the fine
temporal structure of event-based signals. However, some spike-timing-related
strengths of SNNs are hindered by the sensitivity of spike-timing-dependent
plasticity (STDP) rules to input spike rates, as fine temporal correlations may
be obstructed by coarser correlations between firing rates. In this article, we
propose a spike-timing-dependent learning rule that allows a neuron to learn
from the temporally-coded information despite the presence of rate codes. Our
long-term plasticity rule makes use of short-term synaptic fatigue dynamics. We
show analytically that, in contrast to conventional STDP rules, our fatiguing
STDP (FSTDP) helps learn the temporal code, and we derive the necessary
conditions to optimize the learning process. We showcase the effectiveness of
FSTDP in learning spike-timing correlations among processes of different rates
in synthetic data. Finally, we use FSTDP to detect correlations in real-world
weather data from the United States in an experimental realization of the
algorithm that uses a neuromorphic hardware platform comprising phase-change
memristive devices. Taken together, our analyses and demonstrations suggest
that FSTDP paves the way for the exploitation of the spike-based strengths of
SNNs in real-world applications. | [
1,
0,
0,
1,
0,
0
] |
Title: Coresets for Vector Summarization with Applications to Network Graphs,
Abstract: We provide a deterministic data summarization algorithm that approximates the
mean $\bar{p}=\frac{1}{n}\sum_{p\in P} p$ of a set $P$ of $n$ vectors in
$\REAL^d$, by a weighted mean $\tilde{p}$ of a \emph{subset} of $O(1/\eps)$
vectors, i.e., independent of both $n$ and $d$. We prove that the squared
Euclidean distance between $\bar{p}$ and $\tilde{p}$ is at most $\eps$
multiplied by the variance of $P$. We use this algorithm to maintain an
approximated sum of vectors from an unbounded stream, using memory that is
independent of $d$, and logarithmic in the $n$ vectors seen so far. Our main
application is to extract and represent in a compact way friend groups and
activity summaries of users from underlying data exchanges. For example, in the
case of mobile networks, we can use GPS traces to identify meetings, in the
case of social networks, we can use information exchange to identify friend
groups. Our algorithm provably identifies the {\it Heavy Hitter} entries in a
proximity (adjacency) matrix. The Heavy Hitters can be used to extract and
represent in a compact way friend groups and activity summaries of users from
underlying data exchanges. We evaluate the algorithm on several large data
sets. | [
1,
0,
0,
0,
0,
0
] |
Title: Two-dimensional matter-wave solitons and vortices in competing cubic-quintic nonlinear lattices,
Abstract: The nonlinear lattice---a new and nonlinear class of periodic
potentials---was recently introduced to generate various nonlinear localized
modes. Several attempts failed to stabilize two-dimensional (2D) solitons
against their intrinsic critical collapse in Kerr media. Here, we provide a
possibility for supporting 2D matter-wave solitons and vortices in an extended
setting---the cubic and quintic model---by introducing another nonlinear
lattice whose period is controllable and can be different from its cubic
counterpart, to its quintic nonlinearity, therefore making a fully `nonlinear
quasi-crystal'.
A variational approximation based on Gaussian ansatz is developed for the
fundamental solitons and in particular, their stability exactly follows the
inverted \textit{Vakhitov-Kolokolov} stability criterion, whereas the vortex
solitons are only studied by means of numerical methods. Stability regions for
two types of localized mode---the fundamental and vortex solitons---are
provided. A noteworthy feature of the localized solutions is that the vortex
solitons are stable only when the period of the quintic nonlinear lattice is
the same as the cubic one or when the quintic nonlinearity is constant, while
the stable fundamental solitons can be created under looser conditions. Our
physical setting (cubic-quintic model) is in the framework of the
Gross-Pitaevskii equation (GPE) or nonlinear Schrödinger equation, the
predicted localized modes thus may be implemented in Bose-Einstein condensates
and nonlinear optical media with tunable cubic and quintic nonlinearities. | [
0,
1,
0,
0,
0,
0
] |
Title: NimbRo-OP2X: Adult-sized Open-source 3D Printed Humanoid Robot,
Abstract: Humanoid robotics research depends on capable robot platforms, but recently
developed advanced platforms are often not available to other research groups,
expensive, dangerous to operate, or closed-source. The lack of available
platforms forces researchers to work with smaller robots, which have less
strict dynamic constraints or with simulations, which lack many real-world
effects. We developed NimbRo-OP2X to address this need. At a height of 135 cm
our robot is large enough to interact in a human environment. Its low weight of
only 19 kg makes the operation of the robot safe and easy, as no special
operational equipment is necessary. Our robot is equipped with a fast onboard
computer and a GPU to accelerate parallel computations. We extend our already
open-source software by a deep-learning based vision system and gait parameter
optimisation. The NimbRo-OP2X was evaluated during RoboCup 2018 in Montréal,
Canada, where it won all possible awards in the Humanoid AdultSize class. | [
1,
0,
0,
0,
0,
0
] |
Title: Approximate Steepest Coordinate Descent,
Abstract: We propose a new selection rule for the coordinate selection in coordinate
descent methods for huge-scale optimization. The efficiency of this novel
scheme is provably better than the efficiency of uniformly random selection,
and can reach the efficiency of steepest coordinate descent (SCD), enabling an
acceleration of a factor of up to $n$, the number of coordinates. In many
practical applications, our scheme can be implemented at no extra cost and
computational efficiency very close to the faster uniform selection. Numerical
experiments with Lasso and Ridge regression show promising improvements, in
line with our theoretical guarantees. | [
1,
0,
1,
0,
0,
0
] |
Title: Learning Latent Representations for Speech Generation and Transformation,
Abstract: An ability to model a generative process and learn a latent representation
for speech in an unsupervised fashion will be crucial to process vast
quantities of unlabelled speech data. Recently, deep probabilistic generative
models such as Variational Autoencoders (VAEs) have achieved tremendous success
in modeling natural images. In this paper, we apply a convolutional VAE to
model the generative process of natural speech. We derive latent space
arithmetic operations to disentangle learned latent representations. We
demonstrate the capability of our model to modify the phonetic content or the
speaker identity for speech segments using the derived operations, without the
need for parallel supervisory data. | [
1,
0,
0,
1,
0,
0
] |
Title: On the generalized nonlinear Camassa-Holm equation,
Abstract: In this paper, a generalized nonlinear Camassa-Holm equation with time- and
space-dependent coefficients is considered. We show that the control of the
higher order dispersive term is possible by using an adequate weight function
to define the energy. The existence and uniqueness of solutions are obtained
via a Picard iterative method. | [
0,
0,
1,
0,
0,
0
] |
Title: Early MFCC And HPCP Fusion for Robust Cover Song Identification,
Abstract: While most schemes for automatic cover song identification have focused on
note-based features such as HPCP and chord profiles, a few recent papers
surprisingly showed that local self-similarities of MFCC-based features also
have classification power for this task. Since MFCC and HPCP capture
complementary information, we design an unsupervised algorithm that combines
normalized, beat-synchronous blocks of these features using cross-similarity
fusion before attempting to locally align a pair of songs. As an added bonus,
our scheme naturally incorporates structural information in each song to fill
in alignment gaps where both feature sets fail. We show a striking jump in
performance over MFCC and HPCP alone, achieving a state of the art mean
reciprocal rank of 0.87 on the Covers80 dataset. We also introduce a new
medium-sized hand designed benchmark dataset called "Covers 1000," which
consists of 395 cliques of cover songs for a total of 1000 songs, and we show
that our algorithm achieves an MRR of 0.9 on this dataset for the first
correctly identified song in a clique. We provide the precomputed HPCP and MFCC
features, as well as beat intervals, for all songs in the Covers 1000 dataset
for use in further research. | [
1,
0,
0,
0,
0,
0
] |
Title: P-Governance Technology: Using Big Data for Political Party Management,
Abstract: Information and Communication Technology (ICT) has been playing a pivotal
role since the last decade in developing countries that brings citizen services
to the doorsteps and connecting people. With this aspiration ICT has introduced
several technologies of citizen services towards all categories of people. The
purpose of this study is to examine the Governance technology perspectives for
political party, emphasizing on the basic critical steps through which it could
be operationalized. We call it P-Governance. P-Governance shows technologies to
ensure governance, management, interaction communication in a political party
by improving decision making processes using big data. P-Governance challenges
the competence perspective to apply itself more assiduously to
operationalization, including the need to choose and give definition to one or
more units of analysis (of which the routine is a promising candidate). This
paper is to focus on research challenges posed by competence to which
P-Governance can and should respond include different strategy issues faced by
particular sections. Both the qualitative as well as quantitative research
approaches were conducted. The standard of citizen services, choice &
consultation, courtesy & consultation, entrance & information, and value for
money have found the positive relation with citizen's satisfaction. This study
results how can be technology make important roles on political movements in
developing countries using big data. | [
1,
0,
0,
0,
0,
0
] |
Title: Analogues of the $p^n$th Hilbert symbol in characteristic $p$ (updated),
Abstract: The $p$th degree Hilbert symbol $(\cdot,\cdot )_p:K^\times/K^{\times p}\times
K^\times/K^{\times p}\to{}_p{\rm Br}(K)$ from characteristic $\neq p$ has two
analogues in characteristic $p$, $$[\cdot,\cdot )_p:K/\wp (K)\times
K^\times/K^{\times p}\to{}_p{\rm Br}(K),$$ where $\wp$ is the Artin-Schreier
map $x\mapsto x^p-x$, and $$((\cdot,\cdot ))_p:K/K^p\times K/K^p\to{}_p{\rm
Br}(K).$$
The symbol $[\cdot,\cdot )_p$ generalizes to an analogue of $(\cdot,\cdot
)_{p^n}$ via the Witt vectors, $$[\cdot,\cdot )_{p^n}:W_n(K)/\wp (W_n(K))\times
K^\times/K^{\times p^n}\to{}_{p^n}{\rm Br}(K).$$
Here $W_n(K)$ is the truncation of length $n$ of the ring of $p$-typical Witt
wectors, i.e. $W_{\{1,p,\ldots,p^{n-1}\}}(K)$.
In this paper we construct similar generalizations for $((\cdot,\cdot ))_p$.
Our construction involves Witt vectors and Weyl algebras. In the process we
obtain a new kind of Weyl algebras in characteristic $p$, with many interesting
properties.
The symbols we introduce, $((\cdot,\cdot ))_{p^n}$ and, more generally,
$((\cdot,\cdot ))_{p^m,p^n}$, which here are defined in terms of central simple
algebras, coincide with the homonymous symbols we introduced in
[arXiv:1711.00980] in terms of the symbols $[\cdot,\cdot )_{p^n}$. This will be
proved in a future paper. In the present paper we only introduce the symbols
and we prove that they have the same properties with the symbols from
[arXiv:1711.00980]. These properies are enough to obtain the representation
theorem for ${}_{p^n}{\rm Br}(K)$ from [arXiv:1711.00980], Theorem 4.10. | [
0,
0,
1,
0,
0,
0
] |
Title: GAMER-2: a GPU-accelerated adaptive mesh refinement code -- accuracy, performance, and scalability,
Abstract: We present GAMER-2, a GPU-accelerated adaptive mesh refinement (AMR) code for
astrophysics. It provides a rich set of features, including adaptive
time-stepping, several hydrodynamic schemes, magnetohydrodynamics,
self-gravity, particles, star formation, chemistry and radiative processes with
GRACKLE, data analysis with yt, and memory pool for efficient object
allocation. GAMER-2 is fully bitwise reproducible. For the performance
optimization, it adopts hybrid OpenMP/MPI/GPU parallelization and utilizes
overlapping CPU computation, GPU computation, and CPU-GPU communication. Load
balancing is achieved using a Hilbert space-filling curve on a level-by-level
basis without the need to duplicate the entire AMR hierarchy on each MPI
process. To provide convincing demonstrations of the accuracy and performance
of GAMER-2, we directly compare with Enzo on isolated disk galaxy simulations
and with FLASH on galaxy cluster merger simulations. We show that the physical
results obtained by different codes are in very good agreement, and GAMER-2
outperforms Enzo and FLASH by nearly one and two orders of magnitude,
respectively, on the Blue Waters supercomputers using $1-256$ nodes. More
importantly, GAMER-2 exhibits similar or even better parallel scalability
compared to the other two codes. We also demonstrate good weak and strong
scaling using up to 4096 GPUs and 65,536 CPU cores, and achieve a uniform
resolution as high as $10{,}240^3$ cells. Furthermore, GAMER-2 can be adopted
as an AMR+GPUs framework and has been extensively used for the wave dark matter
($\psi$DM) simulations. GAMER-2 is open source (available at
this https URL) and new contributions are welcome. | [
0,
1,
0,
0,
0,
0
] |
Title: Distributed sub-optimal resource allocation over weight-balanced graph via singular perturbation,
Abstract: In this paper, we consider distributed optimization design for resource
allocation problems over weight-balanced graphs. With the help of singular
perturbation analysis, we propose a simple sub-optimal continuous-time
optimization algorithm. Moreover, we prove the existence and uniqueness of the
algorithm equilibrium, and then show the convergence with an exponential rate.
Finally, we verify the sub-optimality of the algorithm, which can approach the
optimal solution as an adjustable parameter tends to zero. | [
0,
0,
1,
0,
0,
0
] |
Title: Decentralized P2P Energy Trading under Network Constraints in a Low-Voltage Network,
Abstract: The increasing uptake of distributed energy resources (DERs) in distribution
systems and the rapid advance of technology have established new scenarios in
the operation of low-voltage networks. In particular, recent trends in
cryptocurrencies and blockchain have led to a proliferation of peer-to-peer
(P2P) energy trading schemes, which allow the exchange of energy between the
neighbors without any intervention of a conventional intermediary in the
transactions. Nevertheless, far too little attention has been paid to the
technical constraints of the network under this scenario. A major challenge to
implementing P2P energy trading is that of ensuring that network constraints
are not violated during the energy exchange. This paper proposes a methodology
based on sensitivity analysis to assess the impact of P2P transactions on the
network and to guarantee an exchange of energy that does not violate network
constraints. The proposed method is tested on a typical UK low-voltage network.
The results show that our method ensures that energy is exchanged between users
under the P2P scheme without violating the network constraints, and that users
can still capture the economic benefits of the P2P architecture. | [
1,
0,
0,
0,
0,
0
] |
Title: Fictitious GAN: Training GANs with Historical Models,
Abstract: Generative adversarial networks (GANs) are powerful tools for learning
generative models. In practice, the training may suffer from lack of
convergence. GANs are commonly viewed as a two-player zero-sum game between two
neural networks. Here, we leverage this game theoretic view to study the
convergence behavior of the training process. Inspired by the fictitious play
learning process, a novel training method, referred to as Fictitious GAN, is
introduced. Fictitious GAN trains the deep neural networks using a mixture of
historical models. Specifically, the discriminator (resp. generator) is updated
according to the best-response to the mixture outputs from a sequence of
previously trained generators (resp. discriminators). It is shown that
Fictitious GAN can effectively resolve some convergence issues that cannot be
resolved by the standard training approach. It is proved that asymptotically
the average of the generator outputs has the same distribution as the data
samples. | [
0,
0,
0,
1,
0,
0
] |
Title: The Wavefunction of the Collapsing Bose-Einstein Condensate,
Abstract: Bose-Einstein condensates with tunable interatomic interactions have been
studied intensely in recent experiments. The investigation of the collapse of a
condensate following a sudden change in the nature of the interaction from
repulsive to attractive has led to the observation of a remnant condensate that
did not undergo further collapse. We suggest that this high-density remnant is
in fact the absolute minimum of the energy, if the attractive atomic
interactions are nonlocal, and is therefore inherently stable. We show that a
variational trial function consisting of a superposition of two distinct
gaussians is an accurate representation of the wavefunction of the ground state
of the conventional local Gross-Pitaevskii field equation for an attractive
condensate and gives correctly the points of emergence of instability. We then
use such a superposition of two gaussians as a variational trial function in
order to calculate the minima of the energy when it includes a nonlocal
interaction term. We use experimental data in order to study the long range of
the nonlocal interaction, showing that they agree very well with a
dimensionally derived expression for this range. | [
0,
1,
0,
0,
0,
0
] |
Title: Not-So-Random Features,
Abstract: We propose a principled method for kernel learning, which relies on a
Fourier-analytic characterization of translation-invariant or
rotation-invariant kernels. Our method produces a sequence of feature maps,
iteratively refining the SVM margin. We provide rigorous guarantees for
optimality and generalization, interpreting our algorithm as online
equilibrium-finding dynamics in a certain two-player min-max game. Evaluations
on synthetic and real-world datasets demonstrate scalability and consistent
improvements over related random features-based methods. | [
1,
0,
0,
1,
0,
0
] |
Title: CLEAR: Coverage-based Limiting-cell Experiment Analysis for RNA-seq,
Abstract: Direct cDNA preamplification protocols developed for single-cell RNA-seq
(scRNA-seq) have enabled transcriptome profiling of rare cells without having
to pool multiple samples or to perform RNA extraction. We term this approach
limiting-cell RNA-seq (lcRNA-seq). Unlike scRNA-seq, which focuses on
'cell-atlasing', lcRNA-seq focuses on identifying differentially expressed
genes (DEGs) between experimental groups. This requires accounting for systems
noise which can obscure biological differences. We present CLEAR, a workflow
that identifies robust transcripts in lcRNA-seq data for between-group
comparisons. To develop CLEAR, we compared DEGs from RNA extracted from
FACS-derived CD5+ and CD5- cells from a single chronic lymphocytic leukemia
patient diluted to input RNA levels of 10-, 100- and 1,000pg. Data quality at
ultralow input levels are known to be noisy. When using CLEAR transcripts vs.
using all available transcripts, downstream analyses reveal more shared DEGs,
improved Principal Component Analysis separation of cell type, and increased
similarity between results across different input RNA amounts. CLEAR was
applied to two publicly available ultralow input RNA-seq data and an in-house
murine neural cell lcRNA-seq dataset. CLEAR provides a novel way to visualize
the public datasets while validates cell phenotype markers for astrocytes,
neural stem and progenitor cells. | [
0,
0,
0,
0,
1,
0
] |
Title: Reducing Storage of Global Wind Ensembles with Stochastic Generators,
Abstract: Wind has the potential to make a significant contribution to future energy
resources. Locating the sources of this renewable energy on a global scale is
however extremely challenging, given the difficulty to store very large data
sets generated by modern computer models. We propose a statistical model that
aims at reproducing the data-generating mechanism of an ensemble of runs via a
Stochastic Generator (SG) of global annual wind data. We introduce an
evolutionary spectrum approach with spatially varying parameters based on
large-scale geographical descriptors such as altitude to better account for
different regimes across the Earth's orography. We consider a multi-step
conditional likelihood approach to estimate the parameters that explicitly
accounts for nonstationary features while also balancing memory storage and
distributed computation. We apply the proposed model to more than 18 million
points of yearly global wind speed. The proposed SG requires orders of
magnitude less storage for generating surrogate ensemble members from wind than
does creating additional wind fields from the climate model, even if an
effective lossy data compression algorithm is applied to the simulation output. | [
0,
0,
0,
1,
0,
0
] |
Title: Continual One-Shot Learning of Hidden Spike-Patterns with Neural Network Simulation Expansion and STDP Convergence Predictions,
Abstract: This paper presents a constructive algorithm that achieves successful
one-shot learning of hidden spike-patterns in a competitive detection task. It
has previously been shown (Masquelier et al., 2008) that spike-timing-dependent
plasticity (STDP) and lateral inhibition can result in neurons competitively
tuned to repeating spike-patterns concealed in high rates of overall
presynaptic activity. One-shot construction of neurons with synapse weights
calculated as estimates of converged STDP outcomes results in immediate
selective detection of hidden spike-patterns. The capability of continual
learning is demonstrated through the successful one-shot detection of new sets
of spike-patterns introduced after long intervals in the simulation time.
Simulation expansion (Lightheart et al., 2013) has been proposed as an approach
to the development of constructive algorithms that are compatible with
simulations of biological neural networks. A simulation of a biological neural
network may have orders of magnitude fewer neurons and connections than the
related biological neural systems; therefore, simulated neural networks can be
assumed to be a subset of a larger neural system. The constructive algorithm is
developed using simulation expansion concepts to perform an operation
equivalent to the exchange of neurons between the simulation and the larger
hypothetical neural system. The dynamic selection of neurons to simulate within
a larger neural system (hypothetical or stored in memory) may be a starting
point for a wide range of developments and applications in machine learning and
the simulation of biology. | [
0,
0,
0,
1,
0,
0
] |
Title: Interpretable High-Dimensional Inference Via Score Projection with an Application in Neuroimaging,
Abstract: In the fields of neuroimaging and genetics, a key goal is testing the
association of a single outcome with a very high-dimensional imaging or genetic
variable. Often, summary measures of the high-dimensional variable are created
to sequentially test and localize the association with the outcome. In some
cases, the results for summary measures are significant, but subsequent tests
used to localize differences are underpowered and do not identify regions
associated with the outcome. Here, we propose a generalization of Rao's score
test based on projecting the score statistic onto a linear subspace of a
high-dimensional parameter space. In addition, we provide methods to localize
signal in the high-dimensional space by projecting the scores to the subspace
where the score test was performed. This allows for inference in the
high-dimensional space to be performed on the same degrees of freedom as the
score test, effectively reducing the number of comparisons. Simulation results
demonstrate the test has competitive power relative to others commonly used. We
illustrate the method by analyzing a subset of the Alzheimer's Disease
Neuroimaging Initiative dataset. Results suggest cortical thinning of the
frontal and temporal lobes may be a useful biological marker of Alzheimer's
risk. | [
0,
0,
1,
1,
0,
0
] |
Title: Drawing Big Graphs using Spectral Sparsification,
Abstract: Spectral sparsification is a general technique developed by Spielman et al.
to reduce the number of edges in a graph while retaining its structural
properties. We investigate the use of spectral sparsification to produce good
visual representations of big graphs. We evaluate spectral sparsification
approaches on real-world and synthetic graphs. We show that spectral
sparsifiers are more effective than random edge sampling. Our results lead to
guidelines for using spectral sparsification in big graph visualization. | [
1,
0,
0,
0,
0,
0
] |
Title: Finding polynomial loop invariants for probabilistic programs,
Abstract: Quantitative loop invariants are an essential element in the verification of
probabilistic programs. Recently, multivariate Lagrange interpolation has been
applied to synthesizing polynomial invariants. In this paper, we propose an
alternative approach. First, we fix a polynomial template as a candidate of a
loop invariant. Using Stengle's Positivstellensatz and a transformation to a
sum-of-squares problem, we find sufficient conditions on the coefficients.
Then, we solve a semidefinite programming feasibility problem to synthesize the
loop invariants. If the semidefinite program is unfeasible, we backtrack after
increasing the degree of the template. Our approach is semi-complete in the
sense that it will always lead us to a feasible solution if one exists and
numerical errors are small. Experimental results show the efficiency of our
approach. | [
1,
0,
0,
0,
0,
0
] |
Title: Multifrequency Excitation and Detection Scheme in Apertureless Scattering Near Field Scanning Optical Microscopy,
Abstract: We theoretically and experimentally demonstrate a multifrequency excitation
and detection scheme in apertureless near field optical microscopy, that
exceeds current state of the art sensitivity and background suppression. By
exciting the AFM tip at its two first flexural modes, and demodulating the
detected signal at the harmonics of their sum, we extract a near field signal
with a twofold improved sensitivity and deep sub-wavelength resolution,
reaching $\lambda/230$. Furthermore, the method offers rich control over
experimental degrees of freedom, expanding the parameter space for achieving
complete optical background suppression. This approach breaks the ground for
non-interferometric complete phase and amplitude retrieval of the near field
signal, and is suitable for any multimodal excitation and higher harmonic
demodulation. | [
0,
1,
0,
0,
0,
0
] |
Title: Global entropy solutions to the compressible Euler equations in the isentropic nozzle flow for large data: Application of the modified Godunov scheme and the generalized invariant regions,
Abstract: We study the motion of isentropic gas in nozzles. This is a major subject in
fluid dynamics. In fact, the nozzle is utilized to increase the thrust of
rocket engines. Moreover, the nozzle flow is closely related to astrophysics.
These phenomena are governed by the compressible Euler equation, which is one
of crucial equations in inhomogeneous conservation laws.
In this paper, we consider its unsteady flow and devote to proving the global
existence and stability of solutions to the Cauchy problem for the general
nozzle. The theorem has been proved in (Tsuge in Arch. Ration. Mech. Anal.
209:365-400 (2013)). However, this result is limited to small data. Our aim in
the present paper is to remove this restriction, that is, we consider large
data. Although the subject is important in Mathematics, Physics and
engineering, it remained open for a long time. The problem seems to lie in a
bounded estimate of approximate solutions, because we have only method to
investigate the behavior with respect to the time variable. To solve this, we
first introduce a generalized invariant region. Compared with the existing
ones, its upper and lower bounds are extended constants to functions of the
space variable. However, we cannot apply the new invariant region to the
traditional difference method. Therefore, we invent the modified Godunov
scheme. The approximate solutions consist of some functions corresponding to
the upper and lower bounds of the invariant regions. These methods enable us to
investigate the behavior of approximate solutions with respect to the space
variable. The ideas are also applicable to other nonlinear problems involving
similar difficulties. | [
0,
0,
1,
0,
0,
0
] |
Title: Gravitational instabilities in a protosolar-like disc II: continuum emission and mass estimates,
Abstract: Gravitational instabilities (GIs) are most likely a fundamental process
during the early stages of protoplanetary disc formation. Recently, there have
been detections of spiral features in young, embedded objects that appear
consistent with GI-driven structure. It is crucial to perform hydrodynamic and
radiative transfer simulations of gravitationally unstable discs in order to
assess the validity of GIs in such objects, and constrain optimal targets for
future observations. We utilise the radiative transfer code LIME to produce
continuum emission maps of a $0.17\,\mathrm{M}_{\odot}$ self-gravitating
protosolar-like disc. We note the limitations of using LIME as is and explore
methods to improve upon the default gridding. We use CASA to produce synthetic
observations of 270 continuum emission maps generated across different
frequencies, inclinations and dust opacities. We find that the spiral structure
of our protosolar-like disc model is distinguishable across the majority of our
parameter space after 1 hour of observation, and is especially prominent at
230$\,$GHz due to the favourable combination of angular resolution and
sensitivity. Disc mass derived from the observations is sensitive to the
assumed dust opacities and temperatures, and therefore can be underestimated by
a factor of at least 30 at 850$\,$GHz and 2.5 at 90$\,$GHz. As a result, this
effect could retrospectively validate GIs in discs previously thought not
massive enough to be gravitationally unstable, which could have a significant
impact on the understanding of the formation and evolution of protoplanetary
discs. | [
0,
1,
0,
0,
0,
0
] |
Title: Information Processing by Networks of Quantum Decision Makers,
Abstract: We suggest a model of a multi-agent society of decision makers taking
decisions being based on two criteria, one is the utility of the prospects and
the other is the attractiveness of the considered prospects. The model is the
generalization of quantum decision theory, developed earlier for single
decision makers realizing one-step decisions, in two principal aspects. First,
several decision makers are considered simultaneously, who interact with each
other through information exchange. Second, a multistep procedure is treated,
when the agents exchange information many times. Several decision makers
exchanging information and forming their judgement, using quantum rules, form a
kind of a quantum information network, where collective decisions develop in
time as a result of information exchange. In addition to characterizing
collective decisions that arise in human societies, such networks can describe
dynamical processes occurring in artificial quantum intelligence composed of
several parts or in a cluster of quantum computers. The practical usage of the
theory is illustrated on the dynamic disjunction effect for which three
quantitative predictions are made: (i) the probabilistic behavior of decision
makers at the initial stage of the process is described; (ii) the decrease of
the difference between the initial prospect probabilities and the related
utility factors is proved; (iii) the existence of a common consensus after
multiple exchange of information is predicted. The predicted numerical values
are in very good agreement with empirical data. | [
1,
0,
0,
0,
0,
0
] |
Title: From Strings to Sets,
Abstract: A complete proof is given of relative interpretability of Adjunctive Set
Theory with Extensionality in an elementary concatenation theory. | [
0,
0,
1,
0,
0,
0
] |
Title: Evolutionary game of coalition building under external pressure,
Abstract: We study the fragmentation-coagulation (or merging and splitting)
evolutionary control model as introduced recently by one of the authors, where
$N$ small players can form coalitions to resist to the pressure exerted by the
principal. It is a Markov chain in continuous time and the players have a
common reward to optimize. We study the behavior as $N$ grows and show that the
problem converges to a (one player) deterministic optimization problem in
continuous time, in the infinite dimensional state space. | [
0,
0,
1,
0,
0,
0
] |
Title: Graphene-based electron transport layers in perovskite solar cells: a step-up for an efficient carrier collection,
Abstract: The electron transport layer (ETL) plays a fundamental role in perovskite
solar cells. Recently, graphene-based ETLs have been proved to be good
candidate for scalable fabrication processes and to achieve higher carrier
injection with respect to most commonly used ETLs. In this work we
experimentally study the effects of different graphene-based ETLs in sensitized
MAPI solar cells. By means of time-integrated and picosecond time-resolved
photoluminescence techniques, the carrier recombination dynamics in MAPI films
embedded in different ETLs is investigated. Using graphene doped mesoporous
TiO2 (G+mTiO2) with the addition of a lithium-neutralized graphene oxide
(GO-Li) interlayer as ETL, we find that the carrier collection efficiency is
increased by about a factor two with respect to standard mTiO2. Taking
advantage of the absorption coefficient dispersion, we probe the MAPI layer
morphology, along the thickness, finding that the MAPI embedded in the ETL
composed by G+mTiO2 plus GO-Li brings to a very good crystalline quality of the
MAPI layer with a trap density about one order of magnitude lower than that
found with the other ETLs. In addition, this ETL freezes MAPI at the tetragonal
phase, regardless of the temperature. Graphene-based ETLs can open the way to
significant improvement of perovskite solar cells. | [
0,
1,
0,
0,
0,
0
] |
Title: Learning and Transferring IDs Representation in E-commerce,
Abstract: Many machine intelligence techniques are developed in E-commerce and one of
the most essential components is the representation of IDs, including user ID,
item ID, product ID, store ID, brand ID, category ID etc. The classical
encoding based methods (like one-hot encoding) are inefficient in that it
suffers sparsity problems due to its high dimension, and it cannot reflect the
relationships among IDs, either homogeneous or heterogeneous ones. In this
paper, we propose an embedding based framework to learn and transfer the
representation of IDs. As the implicit feedbacks of users, a tremendous amount
of item ID sequences can be easily collected from the interactive sessions. By
jointly using these informative sequences and the structural connections among
IDs, all types of IDs can be embedded into one low-dimensional semantic space.
Subsequently, the learned representations are utilized and transferred in four
scenarios: (i) measuring the similarity between items, (ii) transferring from
seen items to unseen items, (iii) transferring across different domains, (iv)
transferring across different tasks. We deploy and evaluate the proposed
approach in Hema App and the results validate its effectiveness. | [
1,
0,
0,
1,
0,
0
] |
Title: Differentiable Supervector Extraction for Encoding Speaker and Phrase Information in Text Dependent Speaker Verification,
Abstract: In this paper, we propose a new differentiable neural network alignment
mechanism for text-dependent speaker verification which uses alignment models
to produce a supervector representation of an utterance. Unlike previous works
with similar approaches, we do not extract the embedding of an utterance from
the mean reduction of the temporal dimension. Our system replaces the mean by a
phrase alignment model to keep the temporal structure of each phrase which is
relevant in this application since the phonetic information is part of the
identity in the verification task. Moreover, we can apply a convolutional
neural network as front-end, and thanks to the alignment process being
differentiable, we can train the whole network to produce a supervector for
each utterance which will be discriminative with respect to the speaker and the
phrase simultaneously. As we show, this choice has the advantage that the
supervector encodes the phrase and speaker information providing good
performance in text-dependent speaker verification tasks. In this work, the
process of verification is performed using a basic similarity metric, due to
simplicity, compared to other more elaborate models that are commonly used. The
new model using alignment to produce supervectors was tested on the
RSR2015-Part I database for text-dependent speaker verification, providing
competitive results compared to similar size networks using the mean to extract
embeddings. | [
1,
0,
0,
0,
0,
0
] |
Title: Partial-wave Coulomb t-matrices for like-charged particles at ground-state energy,
Abstract: We study a special case at which the analytical solution of the
Lippmann-Schwinger integral equation for the partial wave two-body Coulomb
transition matrix for likely charged particles at negative energy is possible.
With the use of the Fock's method of the stereographic projection of the
momentum space onto the four-dimensional unit sphere, the analytical
expressions for s-, p- and d-wave partial Coulomb transition matrices for
repulsively interacting particles at bound-state energy have been derived. | [
0,
1,
0,
0,
0,
0
] |
Title: The Cross-section of a Spherical Double Cone,
Abstract: We show that the poset of $SL(n)$-orbit closures in the product of two
partial flag varieties is a lattice if the action of $SL(n)$ is spherical. | [
0,
0,
1,
0,
0,
0
] |
Title: SESA: Supervised Explicit Semantic Analysis,
Abstract: In recent years supervised representation learning has provided state of the
art or close to the state of the art results in semantic analysis tasks
including ranking and information retrieval. The core idea is to learn how to
embed items into a latent space such that they optimize a supervised objective
in that latent space. The dimensions of the latent space have no clear
semantics, and this reduces the interpretability of the system. For example, in
personalization models, it is hard to explain why a particular item is ranked
high for a given user profile. We propose a novel model of representation
learning called Supervised Explicit Semantic Analysis (SESA) that is trained in
a supervised fashion to embed items to a set of dimensions with explicit
semantics. The model learns to compare two objects by representing them in this
explicit space, where each dimension corresponds to a concept from a knowledge
base. This work extends Explicit Semantic Analysis (ESA) with a supervised
model for ranking problems. We apply this model to the task of Job-Profile
relevance in LinkedIn in which a set of skills defines our explicit dimensions
of the space. Every profile and job are encoded to this set of skills their
similarity is calculated in this space. We use RNNs to embed text input into
this space. In addition to interpretability, our model makes use of the
web-scale collaborative skills data that is provided by users for each LinkedIn
profile. Our model provides state of the art result while it remains
interpretable. | [
1,
0,
0,
0,
0,
0
] |
Title: A Network of Networks Approach to Interconnected Power Grids,
Abstract: We present two different approaches to model power grids as interconnected
networks of networks. Both models are derived from a model for spatially
embedded mono-layer networks and are generalised to handle an arbitrary number
of network layers. The two approaches are distinguished by their use case. The
static glue stick construction model yields a multi-layer network from a
predefined layer interconnection scheme, i.e. different layers are attached
with transformer edges. It is especially suited to construct multi-layer power
grids with a specified number of nodes in and transformers between layers. We
contrast it with a genuine growth model which we label interconnected layer
growth model. | [
0,
1,
0,
0,
0,
0
] |
Title: Marked points on translation surfaces,
Abstract: We show that all GL(2,R) equivariant point markings over orbit closures of
translation surfaces arise from branched covering constructions and periodic
points, completely classify such point markings over strata of quadratic
differentials, and give applications to the finite blocking problem. | [
0,
0,
1,
0,
0,
0
] |
Title: Multitask Learning for Fundamental Frequency Estimation in Music,
Abstract: Fundamental frequency (f0) estimation from polyphonic music includes the
tasks of multiple-f0, melody, vocal, and bass line estimation. Historically
these problems have been approached separately, and only recently, using
learning-based approaches. We present a multitask deep learning architecture
that jointly estimates outputs for various tasks including multiple-f0, melody,
vocal and bass line estimation, and is trained using a large,
semi-automatically annotated dataset. We show that the multitask model
outperforms its single-task counterparts, and explore the effect of various
design decisions in our approach, and show that it performs better or at least
competitively when compared against strong baseline methods. | [
1,
0,
0,
1,
0,
0
] |
Title: Planning with Multiple Biases,
Abstract: Recent work has considered theoretical models for the behavior of agents with
specific behavioral biases: rather than making decisions that optimize a given
payoff function, the agent behaves inefficiently because its decisions suffer
from an underlying bias. These approaches have generally considered an agent
who experiences a single behavioral bias, studying the effect of this bias on
the outcome.
In general, however, decision-making can and will be affected by multiple
biases operating at the same time. How do multiple biases interact to produce
the overall outcome? Here we consider decisions in the presence of a pair of
biases exhibiting an intuitively natural interaction: present bias -- the
tendency to value costs incurred in the present too highly -- and sunk-cost
bias -- the tendency to incorporate costs experienced in the past into one's
plans for the future.
We propose a theoretical model for planning with this pair of biases, and we
show how certain natural behavioral phenomena can arise in our model only when
agents exhibit both biases. As part of our model we differentiate between
agents that are aware of their biases (sophisticated) and agents that are
unaware of them (naive). Interestingly, we show that the interaction between
the two biases is quite complex: in some cases, they mitigate each other's
effects while in other cases they might amplify each other. We obtain a number
of further results as well, including the fact that the planning problem in our
model for an agent experiencing and aware of both biases is computationally
hard in general, though tractable under more relaxed assumptions. | [
1,
1,
0,
0,
0,
0
] |
Title: Zhu reduction for Jacobi $n$-point functions and applications,
Abstract: We establish precise Zhu reduction formulas for Jacobi $n$-point functions
which show the absence of any possible poles arising in these formulas. We then
exploit this to produce results concerning the structure of strongly regular
vertex operator algebras, and also to motivate new differential operators
acting on Jacobi forms. Finally, we apply the reduction formulas to the Fermion
model in order to create polynomials of quasi-Jacobi forms which are Jacobi
forms. | [
0,
0,
1,
0,
0,
0
] |
Title: Achieving rental harmony with a secretive roommate,
Abstract: Given the subjective preferences of n roommates in an n-bedroom apartment,
one can use Sperner's lemma to find a division of the rent such that each
roommate is content with a distinct room. At the given price distribution, no
roommate has a strictly stronger preference for a different room. We give a new
elementary proof that the subjective preferences of only n-1 of the roommates
actually suffice to achieve this envy-free rent division. Our proof, in
particular, yields an algorithm to find such a fair division of rent. The
techniques also give generalizations of Sperner's lemma including a new proof
of a conjecture of the third author. | [
0,
0,
1,
0,
0,
0
] |
Title: A Channel-Based Perspective on Conjugate Priors,
Abstract: A desired closure property in Bayesian probability is that an updated
posterior distribution be in the same class of distributions --- say Gaussians
--- as the prior distribution. When the updating takes place via a statistical
model, one calls the class of prior distributions the `conjugate priors' of the
model. This paper gives (1) an abstract formulation of this notion of conjugate
prior, using channels, in a graphical language, (2) a simple abstract proof
that such conjugate priors yield Bayesian inversions, and (3) a logical
description of conjugate priors that highlights the required closure of the
priors under updating. The theory is illustrated with several standard
examples, also covering multiple updating. | [
1,
0,
0,
0,
0,
0
] |
Title: A FEL Based on a Superlattice,
Abstract: The motion and photon emission of electrons in a superlattice may be
described as in an undulator. Therefore, there is a close analogy between
ballistic electrons in a superlattice and electrons in a free electron laser
(FEL). Touching upon this analogy the intensity of photon emission in the IR
region and the gain are calculated. It is shown that the amplification can be
significant, reaching tens of percent. | [
0,
1,
0,
0,
0,
0
] |
Title: Least Squares Polynomial Chaos Expansion: A Review of Sampling Strategies,
Abstract: As non-institutive polynomial chaos expansion (PCE) techniques have gained
growing popularity among researchers, we here provide a comprehensive review of
major sampling strategies for the least squares based PCE. Traditional sampling
methods, such as Monte Carlo, Latin hypercube, quasi-Monte Carlo, optimal
design of experiments (ODE), Gaussian quadratures, as well as more recent
techniques, such as coherence-optimal and randomized quadratures are discussed.
We also propose a hybrid sampling method, dubbed alphabetic-coherence-optimal,
that employs the so-called alphabetic optimality criteria used in the context
of ODE in conjunction with coherence-optimal samples. A comparison between the
empirical performance of the selected sampling methods applied to three
numerical examples, including high-order PCE's, high-dimensional problems, and
low oversampling ratios, is presented to provide a road map for practitioners
seeking the most suitable sampling technique for a problem at hand. We observed
that the alphabetic-coherence-optimal technique outperforms other sampling
methods, specially when high-order ODE are employed and/or the oversampling
ratio is low. | [
0,
0,
0,
1,
0,
0
] |
Title: Efficient Dense Labeling of Human Activity Sequences from Wearables using Fully Convolutional Networks,
Abstract: Recognizing human activities in a sequence is a challenging area of research
in ubiquitous computing. Most approaches use a fixed size sliding window over
consecutive samples to extract features---either handcrafted or learned
features---and predict a single label for all samples in the window. Two key
problems emanate from this approach: i) the samples in one window may not
always share the same label. Consequently, using one label for all samples
within a window inevitably lead to loss of information; ii) the testing phase
is constrained by the window size selected during training while the best
window size is difficult to tune in practice. We propose an efficient algorithm
that can predict the label of each sample, which we call dense labeling, in a
sequence of human activities of arbitrary length using a fully convolutional
network. In particular, our approach overcomes the problems posed by the
sliding window step. Additionally, our algorithm learns both the features and
classifier automatically. We release a new daily activity dataset based on a
wearable sensor with hospitalized patients. We conduct extensive experiments
and demonstrate that our proposed approach is able to outperform the
state-of-the-arts in terms of classification and label misalignment measures on
three challenging datasets: Opportunity, Hand Gesture, and our new dataset. | [
1,
0,
0,
0,
0,
0
] |
Title: Open Vocabulary Scene Parsing,
Abstract: Recognizing arbitrary objects in the wild has been a challenging problem due
to the limitations of existing classification models and datasets. In this
paper, we propose a new task that aims at parsing scenes with a large and open
vocabulary, and several evaluation metrics are explored for this problem. Our
proposed approach to this problem is a joint image pixel and word concept
embeddings framework, where word concepts are connected by semantic relations.
We validate the open vocabulary prediction ability of our framework on ADE20K
dataset which covers a wide variety of scenes and objects. We further explore
the trained joint embedding space to show its interpretability. | [
1,
0,
0,
0,
0,
0
] |
Title: NetSpam: a Network-based Spam Detection Framework for Reviews in Online Social Media,
Abstract: Nowadays, a big part of people rely on available content in social media in
their decisions (e.g. reviews and feedback on a topic or product). The
possibility that anybody can leave a review provide a golden opportunity for
spammers to write spam reviews about products and services for different
interests. Identifying these spammers and the spam content is a hot topic of
research and although a considerable number of studies have been done recently
toward this end, but so far the methodologies put forth still barely detect
spam reviews, and none of them show the importance of each extracted feature
type. In this study, we propose a novel framework, named NetSpam, which
utilizes spam features for modeling review datasets as heterogeneous
information networks to map spam detection procedure into a classification
problem in such networks. Using the importance of spam features help us to
obtain better results in terms of different metrics experimented on real-world
review datasets from Yelp and Amazon websites. The results show that NetSpam
outperforms the existing methods and among four categories of features;
including review-behavioral, user-behavioral, reviewlinguistic,
user-linguistic, the first type of features performs better than the other
categories. | [
1,
1,
0,
0,
0,
0
] |
Title: Proving Non-Deterministic Computations in Agda,
Abstract: We investigate proving properties of Curry programs using Agda. First, we
address the functional correctness of Curry functions that, apart from some
syntactic and semantic differences, are in the intersection of the two
languages. Second, we use Agda to model non-deterministic functions with two
distinct and competitive approaches incorporating the non-determinism. The
first approach eliminates non-determinism by considering the set of all
non-deterministic values produced by an application. The second approach
encodes every non-deterministic choice that the application could perform. We
consider our initial experiment a success. Although proving properties of
programs is a notoriously difficult task, the functional logic paradigm does
not seem to add any significant layer of difficulty or complexity to the task. | [
1,
0,
0,
0,
0,
0
] |
Title: Contextual Explanation Networks,
Abstract: Modern learning algorithms excel at producing accurate but complex models of
the data. However, deploying such models in the real-world requires extra care:
we must ensure their reliability, robustness, and absence of undesired biases.
This motivates the development of models that are equally accurate but can be
also easily inspected and assessed beyond their predictive performance. To this
end, we introduce contextual explanation networks (CENs)---a class of
architectures that learn to predict by generating and utilizing intermediate,
simplified probabilistic models. Specifically, CENs generate parameters for
intermediate graphical models which are further used for prediction and play
the role of explanations. Contrary to the existing post-hoc model-explanation
tools, CENs learn to predict and to explain jointly. Our approach offers two
major advantages: (i) for each prediction, valid, instance-specific
explanations are generated with no computational overhead and (ii) prediction
via explanation acts as a regularizer and boosts performance in low-resource
settings. We analyze the proposed framework theoretically and experimentally.
Our results on image and text classification and survival analysis tasks
demonstrate that CENs are not only competitive with the state-of-the-art
methods but also offer additional insights behind each prediction, that are
valuable for decision support. We also show that while post-hoc methods may
produce misleading explanations in certain cases, CENs are always consistent
and allow to detect such cases systematically. | [
1,
0,
0,
1,
0,
0
] |
Title: Inference for Differential Equation Models using Relaxation via Dynamical Systems,
Abstract: Statistical regression models whose mean functions are represented by
ordinary differential equations (ODEs) can be used to describe phenomenons
dynamical in nature, which are abundant in areas such as biology, climatology
and genetics. The estimation of parameters of ODE based models is essential for
understanding its dynamics, but the lack of an analytical solution of the ODE
makes the parameter estimation challenging. The aim of this paper is to propose
a general and fast framework of statistical inference for ODE based models by
relaxation of the underlying ODE system. Relaxation is achieved by a properly
chosen numerical procedure, such as the Runge-Kutta, and by introducing
additive Gaussian noises with small variances. Consequently, filtering methods
can be applied to obtain the posterior distribution of the parameters in the
Bayesian framework. The main advantage of the proposed method is computation
speed. In a simulation study, the proposed method was at least 14 times faster
than the other methods. Theoretical results which guarantee the convergence of
the posterior of the approximated dynamical system to the posterior of true
model are presented. Explicit expressions are given that relate the order and
the mesh size of the Runge-Kutta procedure to the rate of convergence of the
approximated posterior as a function of sample size. | [
0,
0,
0,
1,
0,
0
] |
Title: Imaging the Schwarzschild-radius-scale Structure of M87 with the Event Horizon Telescope using Sparse Modeling,
Abstract: We propose a new imaging technique for radio and optical/infrared
interferometry. The proposed technique reconstructs the image from the
visibility amplitude and closure phase, which are standard data products of
short-millimeter very long baseline interferometers such as the Event Horizon
Telescope (EHT) and optical/infrared interferometers, by utilizing two
regularization functions: the $\ell_1$-norm and total variation (TV) of the
brightness distribution. In the proposed method, optimal regularization
parameters, which represent the sparseness and effective spatial resolution of
the image, are derived from data themselves using cross validation (CV). As an
application of this technique, we present simulated observations of M87 with
the EHT based on four physically motivated models. We confirm that $\ell_1$+TV
regularization can achieve an optimal resolution of $\sim 20-30$% of the
diffraction limit $\lambda/D_{\rm max}$, which is the nominal spatial
resolution of a radio interferometer. With the proposed technique, the EHT can
robustly and reasonably achieve super-resolution sufficient to clearly resolve
the black hole shadow. These results make it promising for the EHT to provide
an unprecedented view of the event-horizon-scale structure in the vicinity of
the super-massive black hole in M87 and also the Galactic center Sgr A*. | [
0,
1,
0,
0,
0,
0
] |
Title: The process of purely event-driven programs,
Abstract: Using process algebra, this paper describes the formalisation of the
process/semantics behind the purely event-driven programming language. | [
1,
0,
0,
0,
0,
0
] |
Title: There's more to the multimedia effect than meets the eye: is seeing pictures believing?,
Abstract: Textbooks in applied mathematics often use graphs to explain the meaning of
formulae, even though their benefit is still not fully explored. To test
processes underlying this assumed multimedia effect we collected performance
scores, eye movements, and think-aloud protocols from students solving problems
in vector calculus with and without graphs. Results showed no overall
multimedia effect, but instead an effect to confirm statements that were
accompanied by graphs, irrespective of whether these statements were true or
false. Eye movement and verbal data shed light on this surprising finding.
Students looked proportionally less at the text and the problem statement when
a graph was present. Moreover, they experienced more mental effort with the
graph, as indicated by more silent pauses in thinking aloud. Hence, students
actively processed the graphs. This, however, was not sufficient. Further
analysis revealed that the more students looked at the statement, the better
they performed. Thus, in the multimedia condition the graph drew students'
attention and cognitive capacities away from focusing on the statement. A good
alternative strategy in the multimedia condition was to frequently look between
graph and problem statement, and thus to integrate their information. In
conclusion, graphs influence where students look and what they process, and may
even mislead them into believing accompanying information. Thus, teachers and
textbook designers should be very critical on when to use graphs and carefully
consider how the graphs are integrated with other parts of the problem. | [
0,
1,
1,
0,
0,
0
] |
Title: Methods of Enumerating Two Vertex Maps of Arbitrary Genus,
Abstract: This paper provides an alternate proof to parts of the Goulden-Slofstra
formula for enumerating two vertex maps by genus, which is an extension of the
famous Harer-Zagier formula that computes the Euler characteristic of the
moduli space of curves. This paper also shows a further simplification to the
Goulden-Slofstra formula. Portions of this alternate proof will be used in a
subsequent paper, where it forms a basis for a more general result that applies
for a certain class of maps with an arbitrary number of vertices. | [
0,
0,
1,
0,
0,
0
] |
Title: Light emission by accelerated electric, toroidal and anapole dipolar sources,
Abstract: Emission of electromagnetic radiation by accelerated particles with electric,
toroidal and anapole dipole moments is analyzed. It is shown that ellipticity
of the emitted light can be used to differentiate between electric and toroidal
dipole sources, and that anapoles, elementary neutral non-radiating
configurations, which consist of electric and toroidal dipoles, can emit light
under uniform acceleration. The existence of non-radiating configurations in
electrodynamics implies that it is impossible to fully determine the internal
makeup of the emitter given only the distribution of the emitted light. Here we
demonstrate that there is a loop-hole in this `inverse source problem'. Our
results imply that there may be a whole range of new phenomena to be discovered
by studying the electromagnetic response of matter under acceleration. | [
0,
1,
0,
0,
0,
0
] |
Title: Transição de fase no sistema de Hénon-Heiles (Phase transition in the Henon-Heiles system),
Abstract: The Henon-Heiles system was originally proposed to describe the dynamical
behavior of galaxies, but this system has been widely applied in dynamical
systems by exhibit great details in phase space. This work presents the
formalism to describe Henon-Heiles system and a qualitative approach of
dynamics behavior. The growth of chaotic region in phase space is observed by
Poincare Surface of Section when the total energy increases. Island of
regularity remain around stable points and relevants phenomena appear, such as
sticky. | [
0,
1,
0,
0,
0,
0
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.