text
stringlengths 6
128k
|
---|
Given two polygonal curves, there are many ways to define a notion of
similarity between them. One popular measure is the Fr\'echet distance which
has many desirable properties but is notoriously expensive to calculate,
especially for non-trivial metrics. In 1994, Eiter and Mannila introduced the
discrete Fr\'echet distance which is much easier to implement and approximates
the continuous Fr\'echet distance with a quadratic runtime overhead. However,
this algorithm relies on recursions and is not well suited for modern hardware.
To that end, we introduce the Fast Fr\'echet Distance algorithm, a
recursion-free algorithm that calculates the discrete Fr\'echet distance with a
linear memory overhead and that can utilize modern hardware more effectively.
We showcase an implementation with only four lines of code and present
benchmarks of our algorithm running fast on modern CPUs and GPGPUs.
|
We prove a nonexistence theorem for product type manifolds. In particular we
show that the 4-manifold $\Sigma_g\times\Sigma_h$ does not admit any locally
conformally flat metric arising from discrete and faithful representations for
$g\geq 2$ and $h\geq 1$
|
Though objectives of trusted routing and virtual private networks (VPN) data
transfer methods are to guarantee data transfer securely to from senders to
receivers over public networks like Internet yet there are paramount
differences between the two methods. This paper analyses their differences.
|
We prove that stable-like non-local Dirichlet forms converge to local
Dirichlet form in the sense of Mosco on metric measure spaces. We prove that
subordinated Dirichlet forms converge to the original Dirichlet form in the
sense of Mosco on metric measure spaces.
|
The FlexRay bus is a modern standard used in the automotive industry.It
offers deterministic message transmission with zero jitter while using
time-triggered scheduling in the static segment. When several vehicle variants
(i.e. different models and their versions) share the same signal, the car
manufacturers require to schedule such signal at the same time in all vehicle
variants. This requirement simplifies the signal traceability and diagnostics
in different vehicle variants using the same platform and simplifies reuse of
components and tools.
In this paper, we propose a first fit based heuristic algorithm which creates
the schedules for several vehicle variants at once, while transmitting a given
signal at the same time in all the schedules. The scheduling algorithm also
takes the time constraints as release dates and deadlines into account.
Finally, different algorithm versions are compared on benchmark sets and low
computational time demands are validated on large instances.
|
In the 3SUM-Indexing problem the goal is to preprocess two lists of elements
from $U$, $A=(a_1,a_2,\ldots,a_n)$ and $B=(b_1,b_2,...,b_n)$, such that given
an element $c\in U$ one can quickly determine whether there exists a pair
$(a,b)\in A \times B$ where $a+b=c$. Goldstein et al.~[WADS'2017] conjectured
that there is no algorithm for 3SUM-Indexing which uses $n^{2-\Omega(1)}$ space
and $n^{1-\Omega(1)}$ query time.
We show that the conjecture is false by reducing the 3SUM-Indexing problem to
the problem of inverting functions, and then applying an algorithm of Fiat and
Naor [SICOMP'1999] for inverting functions.
|
Saliency methods generating visual explanatory maps representing the
importance of image pixels for model classification is a popular technique for
explaining neural network decisions. Hierarchical dynamic masks (HDM), a novel
explanatory maps generation method, is proposed in this paper to enhance the
granularity and comprehensiveness of saliency maps. First, we suggest the
dynamic masks (DM), which enables multiple small-sized benchmark mask vectors
to roughly learn the critical information in the image through an optimization
method. Then the benchmark mask vectors guide the learning of large-sized
auxiliary mask vectors so that their superimposed mask can accurately learn
fine-grained pixel importance information and reduce the sensitivity to
adversarial perturbations. In addition, we construct the HDM by concatenating
DM modules. These DM modules are used to find and fuse the regions of interest
in the remaining neural network classification decisions in the mask image in a
learning-based way. Since HDM forces DM to perform importance analysis in
different areas, it makes the fused saliency map more comprehensive. The
proposed method outperformed previous approaches significantly in terms of
recognition and localization capabilities when tested on natural and medical
datasets.
|
Stepped wedge cluster-randomized trial (CRTs) designs randomize clusters of
individuals to intervention sequences, ensuring that every cluster eventually
transitions from a control period to receive the intervention under study by
the end of the study period. The analysis of stepped wedge CRTs is usually more
complex than parallel-arm CRTs due to potential secular trends that result in
changing intra-cluster and period-cluster correlations over time. A further
challenge in the analysis of closed-cohort stepped wedge CRTs, which follow
groups of individuals enrolled in each period longitudinally, is the occurrence
of dropout. This is particularly problematic in studies of individuals at high
risk for mortality, which causes non-ignorable missing outcomes. If not
appropriately addressed, missing outcomes from death will erode statistical
power, at best, and bias treatment effect estimates, at worst. Joint
longitudinal-survival models can accommodate informative dropout and
missingness patterns in longitudinal studies. Specifically, within this
framework one directly models the dropout process via a time-to-event submodel
together with the longitudinal outcome of interest. The two submodels are then
linked using a variety of possible association structures. This work extends
linear mixed-effects models by jointly modeling the dropout process to
accommodate informative missing outcome data in closed-cohort stepped wedge
CRTs. We focus on constant intervention and general time-on-treatment effect
parametrizations for the longitudinal submodel and study the performance of the
proposed methodology using Monte Carlo simulation under several data-generating
scenarios. We illustrate the joint modeling methodology in practice by
reanalyzing the `Frail Older Adults: Care in Transition' (ACT) trial, a stepped
wedge CRT of a multifaceted geriatric care model versus usual care in the
Netherlands.
|
Deep Reinforcement Learning (Deep RL) has had incredible achievements on high
dimensional problems, yet its learning process remains unstable even on the
simplest tasks. Deep RL uses neural networks as function approximators. These
neural models are largely inspired by developments in the (un)supervised
machine learning community. Compared to these learning frameworks, one of the
major difficulties of RL is the absence of i.i.d. data. One way to cope with
this difficulty is to control the rate of change of the policy at every
iteration. In this work, we challenge the common practices of the
(un)supervised learning community of using a fixed neural architecture, by
having a neural model that grows in size at each policy update. This allows a
closed form entropy regularized policy update, which leads to a better control
of the rate of change of the policy at each iteration and help cope with the
non i.i.d. nature of RL. Initial experiments on classical RL benchmarks show
promising results with remarkable convergence on some RL tasks when compared to
other deep RL baselines, while exhibiting limitations on others.
|
We investigate properties of the conserved charge in general relativity,
recently proposed by one of the present authors with his collaborators, in the
inflation era, the matter dominated era and the radiation dominated era of the
expanding Universe. We show that the conserved charge in the inflation era
becomes the Bekenstein-Hawking entropy for de Sitter space, and it becomes the
matter entropy and the radiation entropy in the matter and radiation dominated
eras, respectively, while the charge itself is always conserved. These
properties are qualitatively confirmed by a numerical analysis of a model with
a scalar field and radiations. Results in this paper provide more evidences on
the interpretation that the conserved charge in general relativity corresponds
to entropy.
|
Nash-Williams proved that every graph has a well-balanced orientation. A key
ingredient in his proof is admissible odd-vertex pairings. We show that for two
slightly different definitions of admissible odd-vertex pairings, deciding
whether a given odd-vertex pairing is admissible is co-NP-complete. This
resolves a question of Frank. We also show that deciding whether a given graph
has an orientation that satisfies arbitrary local arc-connectivity requirements
is NP-complete.
|
We carry out an analytical and numerical study of the motion of an isolated
vortex in thermal equilibrium, the vortex being defined as the point
singularity of a complex scalar field $\psi(\r,t)$ obeying a nonlinear
stochastic Schr\"odinger equation. Because hydrodynamic fluctuations are
included in this description, the dynamical picture of the vortex emerges as
that of both a massive particle in contact with a heat bath, and as a passive
scalar advected to a background random flow. We show that the vortex does not
execute a simple random walk and that the probability distribution of vortex
flights has non-Gaussian (exponential) tails.
|
A coordinate transformation is found which diagonalizes the axisymmetric
pp-waves. Its effect upon concrete solutions, including impulsive and shock
waves, is discussed.
|
We investigate orbital resonances expected to arise when a system of two
planets, with masses in the range 1-4 Earth masses, undergoes convergent
migration while embedded in a section of gaseous disc where the flow is
laminar. We consider surface densities corresponding to 0.5-4 times that
expected for a minimum mass solar nebula at 5.2 AU. Using hydrodynamic
simulations we find that when the configuration is such that convergent
migration occurs the planets can become locked in a first order
commensurability for which the period ratio is (p+1)/p with p being an integer
and migrate together maintaining it for many orbits. Relatively rapid
convergent migration as tends to occur for disparate masses, results in
commensurabilities with p larger than 2. However, in these cases the dynamics
is found to have a stochastic character. When the convergent migration is
slower, such as occurs in the equal mass case, lower p commensurabilities such
as 3:2 are attained which show much greater stability. In one already known
example of a system with nearly equal masses in the several Earth mass range
(planets around pulsar PSR B1257+12) the two largest planets are intriguingly
close to a 3:2 commensurability. A very similar behaviour is obtained when the
systems are modeled using an N body code with simple prescriptions for the disc
planet interaction. Using that, we found that an 8:7 resonance established in a
hydrodynamic simulation run for 10-100 thousand orbits could be maintained for
more than million orbits. Resonant capture leads to a rise in eccentricities
that can be predicted using a simple analytic model constructed in this paper.
We find that the system with the 8:7 commensurability is fully consistent with
this prediction.
|
Multi-label classification deals with the problem where each instance is
associated with multiple class labels. Because evaluation in multi-label
classification is more complicated than single-label setting, a number of
performance measures have been proposed. It is noticed that an algorithm
usually performs differently on different measures. Therefore, it is important
to understand which algorithms perform well on which measure(s) and why. In
this paper, we propose a unified margin view to revisit eleven performance
measures in multi-label classification. In particular, we define label-wise
margin and instance-wise margin, and prove that through maximizing these
margins, different corresponding performance measures will be optimized. Based
on the defined margins, a max-margin approach called LIMO is designed and
empirical results verify our theoretical findings.
|
We investigate soliton collisions a one-parameter family of scalar field
theories in 1+1 dimensions which was first discussed by Christ and Lee. The
models have a sextic potential with three local minima, and for suitably small
values of the parameter its kinks have an internal structure in the form of two
weakly-bound subkinks. We show that for these values of the parameter kink
collisions are best understood as an independent sequence of collisions of
these subkinks, and that a static mode analysis is not enough to explain
resonant structures emerging in this model. We also emphasise the role of
radiation and oscillon formation in the collision process.
|
We collect a number of striking recent results in a study of dimers on
infinite regular bipartite lattices and also on regular bipartite graphs. We
clearly separate rigorously proven results from conjectures. A primary goal is
to show people: here is a field which is ripe for further interesting research.
We separate four classes of endeavor, of which we here extract two items to
whet one's appetite. Primo,for hyper-rectangular lattices of every dimension
the first 20 virial coefficients are positive. (One has no understanding of
this yet!) Secondo, all regular bipartite graphs with less than $14$ vertices
satisfy graph positivity, defined below. (Here there is some understanding.)
|
This paper considers an unsignalized intersection used by two traffic
streams. A stream of cars is using a primary road, and has priority over the
other stream. Cars belonging to the latter stream cross the primary road if the
gaps between two subsequent cars on the primary road are larger than their
critical headways. A question that naturally arises relates to the capacity of
the secondary road: given the arrival pattern of cars on the primary road, what
is the maximum arrival rate of low-priority cars that can be sustained? This
paper addresses this issue by considering a compact model that sheds light on
the dynamics of the considered unsignalized intersection. The model, which is
of a queueing-theoretic nature, reveals interesting insights into the impact of
the user behavior on stability.
The contributions of this paper are threefold. First, we obtain new results
for the aforementioned model that includes driver impatience. Secondly, we
reveal some surprising aspects that have remained unobserved in the existing
literature so far, many of which are caused by the fact that the capacity of
the minor road cannot be expressed in terms of the \emph{mean} gap size;
instead more detailed characteristics of the critical headway distribution play
a crucial role. The third contribution is the introduction of a new form of
bunching on the main road, called Markov platooning. The tractability of this
model allows us to study the impact of various platoon formations on the main
road on the capacity of the minor road.
|
In this paper, we consider an online distributed composite optimization
problem over a time-varying multi-agent network that consists of multiple
interacting nodes, where the objective function of each node consists of two
parts: a loss function that changes over time and a regularization function.
This problem naturally arises in many real-world applications ranging from
wireless sensor networks to signal processing. We propose a class of online
distributed optimization algorithms that are based on approximate mirror
descent, which utilize the Bregman divergence as distance-measuring function
that includes the Euclidean distances as a special case. We consider two
standard information feedback models when designing the algorithms, that is,
full-information feedback and bandit feedback. For the full-information
feedback model, the first algorithm attains an average regularized regret of
order $\mathcal{O}(1/\sqrt{T})$ with the total number of rounds $T$. The second
algorithm, which only requires the information of the values of the loss
function at two predicted points instead of the gradient information, achieves
the same average regularized regret as that of the first algorithm. Simulation
results of a distributed online regularized linear regression problem are
provided to illustrate the performance of the proposed algorithms.
|
A study of the relation between the electrostatic charge density at a point
on a conducting surface and the curvature of the surface (at that point) is
presented. Two major scientific literature on this topic are reviewed and the
apparent discrepancy between them is resolved. Hence, a step is taken towards
obtaining a general analytic formula for relating the charge density with
surface curvature of conductors. The merit of this formula and its limitations
are discussed.
|
The charge density wave phase transition of 1T-TiSe2 is studied by
angle-resolved photoemission over a wide temperature range. An important
chemical potential shift which strongly evolves with temperature is evidenced.
In the framework of the exciton condensate phase, the detailed temperature
dependence of the associated order parameter is extracted. Having a
mean-field-like behaviour at low temperature, it exhibits a non-zero value
above the transition, interpreted as the signature of strong excitonic
fluctuations, reminiscent of the pseudo-gap phase of high temperature
superconductors. Integrated intensity around the Fermi level is found to
display a trend similar to the measured resistivity and is discussed within the
model.
|
Coherent superposition is a key feature of quantum mechanics that underlies
the advantage of quantum technologies over their classical counterparts.
Recently, coherence has been recast as a resource theory in an attempt to
identify and quantify it in an operationally well-defined manner. Here we study
how the coherence present in a state can be used to implement a quantum channel
via incoherent operations and, in turn, to assess its degree of coherence. We
introduce the robustness of coherence of a quantum channel---which reduces to
the homonymous measure for states when computed on constant-output
channels---and prove that: i) it quantifies the minimal rank of a maximally
coherent state required to implement the channel; ii) its logarithm quantifies
the amortized cost of implementing the channel provided some coherence is
recovered at the output; iii) its logarithm also quantifies the zero-error
asymptotic cost of implementation of many independent copies of a channel. We
also consider the generalized problem of imperfect implementation with
arbitrary resource states. Using the robustness of coherence, we find that in
general a quantum channel can be implemented without employing a maximally
coherent resource state. In fact, we prove that \textit{every} pure coherent
state in dimension larger than $2$, however weakly so, turns out to be a
valuable resource to implement \textit{some} coherent unitary channel. We
illustrate our findings for the case of single-qubit unitary channels.
|
We characterize a semiconductor external cavity diode laser whose optical
feedback is provided by a guided mode resonance filter (GMRF). We focus on the
spectral properties. The wavelength of operation falls in the telecom range
(1506 nm). The GMRF acting both as a wavelength intracavity filter and feedback
mirror allows for a compact laser design. The single-mode operation is verified
in a wide range of driving currents. We finely tune the cavity length to adjust
the frequency by 14 GHz without mode-hops in agreement with the expected
free-spectral range of the resonator $\sim$20 GHz. The compactness of the
cavity allows fast frequency sweeps when modulating the current (90 MHz/mA at
100 kHz, the modulation bandwidth). The frequency noise (366 kHz white-noise
contribution) is also analysed to evaluate the potential of our design for
high-resolution applications.
|
We revisit the Lichnerowicz-York method, and an alternative method of York,
in order to obtain some conformally covariant systems. This type of
parameterization is certainly more natural for non constant mean curvature
initial data.
|
We review origins and main properties of the most important bracket
operations appearing canonically in differential geometry and mathematical
physics in the classical, as well as the supergeometric setting. The review is
supplemented by a few new concepts and examples.
|
We address the stability of superfluid currents in a system of interacting
lattice bosons. We consider various Gutzwiller trial states for the quantum
phase model which provides a good approximation for the Bose-Hubbard model in
the limit of large interactions and boson populations. We thoroughly analyze
the current-carrying stationary states of the dynamics ensuing from a Gaussian
ansatz, and derive analytical results for the critical lines signaling their
modulational and energetic instability, as well as the maximum of the carried
current. We show that these analytical results are in good qualitative
agreement with those obtained numerically in previous works on the Bose-Hubbard
model, and in the present work for the quantum phase model.
|
Isochronous mass spectrometry (IMS) in storage rings is a successful
technique for accurate mass measurements of short-lived nuclides with relative
precision of about $10^{-5}-10^{-7}$. Instabilities of the magnetic fields in
storage rings are one of the major contributions limiting the achievable mass
resolving power, which is directly related to the precision of the obtained
mass values. A new data analysis method is proposed allowing one to minimise
the effect of such instabilities. The masses of the previously measured at the
CSRe $^{41}$Ti, $^{43}$V, $^{47}$Mn, $^{49}$Fe, $^{53}$Ni and $^{55}$Cu
nuclides were re-determined with this method. An improvement of the mass
precision by a factor of $\sim 1.7$ has been achieved for $^{41}$Ti and
$^{43}$V. The method can be applied to any isochronous mass experiment
irrespective of the accelerator facility. Furthermore, the method can be used
as an on-line tool for checking the isochronous conditions of the storage ring.
|
In the framework of a Skyrme-Hartree-Fock approach combined with BCS method,
the role of the tensor force on the pseudospin energy splitting for tin isotope
chain is investigated. The tensor force turns out to obviously affect the
pseudospin energy splitting of the spin-unsaturated nuclei. Since the tensor
force shifts the single-particle levels, it modifies the single-particle level
density and the shell correction energy thereof. The influence of the tensor
interaction on shell correction energy is considerable according to our
analysis taking a magic nucleus $^{132}$Sn as well as a superheavy nucleus
$^{298}114$ as examples. This modification of the shell correction energy due
to the tensor component affects the stability of the superheavy nuclei.
|
We calculated the effects of spin-orbit interaction (SOI) on the energy
bands, ballistic conductance and the electron-diffusion thermoelectric power of
a nanowire by varying the temperature, electron density and width of the wire.
The potential barriers at the edges of the wire are assumed to be very high. A
consequence of the boundary conditions used in this model is determined by the
energy band structure, resulting in wider plateaus when the electron density is
increased due to larger energy-level separation as the higher subbands are
occupied by electrons. The nonlinear dependence of the transverse confinement
on position with respect to the well center excludes the "pole-like feature" in
the conductance which is obtained when a harmonic potential is employed for
confinement. At low temperature, the electron diffusion thermoelectric power
increases linearly with T but deviates from the linear behavior for large
values of T.
|
Suppose L is any finite algebraic extension of either the ordinary rational
numbers or the p-adic rational numbers. Also let g_1,...,g_k be polynomials in
n variables, with coefficients in L, such that the total number of monomial
terms appearing in at least one g_i is exactly m. We prove that the maximum
number of isolated roots of G:=(g_1,...,g_k) in L^n is finite and depends
solely on (m,n,L), i.e., is independent of the degrees of the g_i. We thus
obtain an arithmetic analogue of Khovanski's Theorem on Fewnomials, extending
earlier work of Denef, Van den Dries, Lipshitz, and Lenstra.
|
We analyze the cooling of a mechanical resonator coupled to an ensemble of
interacting two-level systems via an open quantum systems approach. Using an
exact analytical result, we find optimal cooling occurs when the phonon mode is
critically coupled ($\gamma \sim g$) to the two-level system ensemble. Typical
systems operate in sub-optimal cooling regimes due to the intrinsic parameter
mismatch ($\gamma \gg g$) between the dissipative decay rate $\gamma$ and the
coupling factor $g$. To overcome this obstacle, we show that carefully
engineering the coupling parameters through the strain profile of the
mechanical resonator allows phonon cooling to proceed through the dark
(subradiant) entangled states of an \emph{interacting} ensemble, thereby
resulting in optimal phonon cooling. Our results provide a new avenue for
ground-state cooling and should be accessible for experimental demonstrations.
|
This paper addresses the problem of sparse recovery with graph constraints in
the sense that we can take additive measurements over nodes only if they induce
a connected subgraph. We provide explicit measurement constructions for several
special graphs. A general measurement construction algorithm is also proposed
and evaluated. For any given graph $G$ with $n$ nodes, we derive order optimal
upper bounds of the minimum number of measurements needed to recover any
$k$-sparse vector over $G$ ($M^G_{k,n}$). Our study suggests that $M^G_{k,n}$
may serve as a graph connectivity metric.
|
The first fully-documented study into the quantitative impact of errors in
operational spreadsheets identified an interesting anomaly. One of the five
participating organisations involved in the study contributed a set of five
spreadsheets of such quality that they set the organisation apart in a
statistical sense. This virtuoso performance gave rise to a simple sampling
test - The Clean Sheet Test - which can be used to objectively evaluate if an
organisation is in control of the spreadsheets it is using in important
processes such as financial reporting.
|
We demonstrate a nonlinear metamaterial in microwave frequency regime with
hysteresis effect and bistable states, which can be utilized as a remotely
controllable micro second switching device. A varactor loaded split-ring
resonator (SRR) design which exhibits power and frequency dependent broadband
tunability of the resonance frequency for an external control signal is used.
More importantly, the SRR shows bistability with distinct transmission levels.
The transition between bi-states is controlled by impulses of an external pump
signal. Furthermore, we experimentally demonstrate that transition rate is in
the order of microseconds by using a varactor loaded double split-ring
resonator (DSRR) design composed of two concentric rings.
|
The increase in distributed energy resources and flexible electricity
consumers has turned TSO-DSO coordination strategies into a challenging
problem. Existing decomposition/decentralized methods apply divide-and-conquer
strategies to trim down the computational burden of this complex problem, but
rely on access to proprietary information or fail-safe real-time communication
infrastructures. To overcome these drawbacks, we propose in this paper a
TSO-DSO coordination strategy that only needs a series of observations of the
nodal price and the power intake at the substations connecting the transmission
and distribution networks. Using this information, we learn the price response
of active distribution networks (DN) using a decreasing step-wise function that
can also adapt to some contextual information. The learning task can be carried
out in a computationally efficient manner and the curve it produces can be
interpreted as a market bid, thus averting the need to revise the current
operational procedures for the transmission network. Inaccuracies derived from
the learning task may lead to suboptimal decisions. However, results from a
realistic case study show that the proposed methodology yields operating
decisions very close to those obtained by a fully centralized coordination of
transmission and distribution.
|
We study the eigenvalue problem for the $p$-Laplacian on K\"ahler manifolds.
Our first result is a lower bound for the first nonzero eigenvalue of the
$p$-Laplacian on compact K\"ahler manifolds in terms of dimension, diameter,
and lower bounds of holomorphic sectional curvature and orthogonal Ricci
curvature for $p\in (1, 2]$. Our second result is a sharp lower bound for the
first Dirichlet eigenvalue of the $p$-Laplacian on compact K\"ahler manifolds
with smooth boundary for $p\in (1, \infty)$. Our results generalize
corresponding results for the Laplace eigenvalues on K\"ahler manifolds proved
in [14].
|
Adaptive indexing is a concept that considers index creation in databases as
a by-product of query processing; as opposed to traditional full index creation
where the indexing effort is performed up front before answering any queries.
Adaptive indexing has received a considerable amount of attention, and several
algorithms have been proposed over the past few years; including a recent
experimental study comparing a large number of existing methods. Until now,
however, most adaptive indexing algorithms have been designed single-threaded,
yet with multi-core systems already well established, the idea of designing
parallel algorithms for adaptive indexing is very natural. In this regard only
one parallel algorithm for adaptive indexing has recently appeared in the
literature: The parallel version of standard cracking. In this paper we
describe three alternative parallel algorithms for adaptive indexing, including
a second variant of a parallel standard cracking algorithm. Additionally, we
describe a hybrid parallel sorting algorithm, and a NUMA-aware method based on
sorting. We then thoroughly compare all these algorithms experimentally; along
a variant of a recently published parallel version of radix sort. Parallel
sorting algorithms serve as a realistic baseline for multi-threaded adaptive
indexing techniques. In total we experimentally compare seven parallel
algorithms. Additionally, we extensively profile all considered algorithms. The
initial set of experiments considered in this paper indicates that our parallel
algorithms significantly improve over previously known ones. Our results
suggest that, although adaptive indexing algorithms are a good design choice in
single-threaded environments, the rules change considerably in the parallel
case. That is, in future highly-parallel environments, sorting algorithms could
be serious alternatives to adaptive indexing.
|
The Cherenkov Telescope Array is expected to lead to the detection of many
new supernova remnants in the TeV and multi-TeV range. In addition to the
individual study of each SNR, the study of these objects as a population can
help constraining the parameters describing the acceleration of particles and
increasing our understanding of the mechanisms involved. We present Monte Carlo
simulations of the population of Galactic SNRs emitting TeV gamma rays. We also
discuss how the simulated population can be confronted with future observations
to provide a novel test for the SNR hypothesis of cosmic ray origins.
|
We consider federated edge learning (FEEL) among mobile devices that harvest
the required energy from their surroundings, and share their updates with the
parameter server (PS) through a shared wireless channel. In particular, we
consider energy harvesting FL with over-the-air (OTA) aggregation, where the
participating devices perform local computations and wireless transmission only
when they have the required energy available, and transmit the local updates
simultaneously over the same channel bandwidth. In order to prevent bias among
heterogeneous devices, we utilize a weighted averaging with respect to their
latest energy arrivals and data cardinalities. We provide a convergence
analysis and carry out numerical experiments with different energy arrival
profiles, which show that even though the proposed scheme is robust against
devices with heterogeneous energy arrivals in error-free scenarios, we observe
a 5-10% performance loss in energy harvesting OTA FL.
|
Relative $t$-designs in the $n$-dimensional hypercube $\mathcal{Q}_n$ are
equivalent to weighted regular $t$-wise balanced designs, which generalize
combinatorial $t$-$(n,k,\lambda)$ designs by allowing multiple block sizes as
well as weights. Partly motivated by the recent study on tight Euclidean
$t$-designs on two concentric spheres, in this paper we discuss tight relative
$t$-designs in $\mathcal{Q}_n$ supported on two shells. We show under a mild
condition that such a relative $t$-design induces the structure of a coherent
configuration with two fibers. Moreover, from this structure we deduce that a
polynomial from the family of the Hahn hypergeometric orthogonal polynomials
must have only integral simple zeros. The Terwilliger algebra is the main tool
to establish these results. By explicitly evaluating the behavior of the zeros
of the Hahn polynomials when they degenerate to the Hermite polynomials under
an appropriate limit process, we prove a theorem which gives a partial evidence
that the non-trivial tight relative $t$-designs in $\mathcal{Q}_n$ supported on
two shells are rare for large $t$.
|
The ages and masses of neutron stars (NSs) are two fundamental threads that
make pulsars accessible to other sub-disciplines of astronomy and physics. A
realistic and accurate determination of these two derived parameters play an
important role in understanding of advanced stages of stellar evolution and the
physics that govern relevant processes. Here I summarize new constraints on the
ages and masses of NSs with an evolutionary perspective. I show that the
observed P-Pdot demographics is more diverse than what is theoretically
predicted for the standard evolutionary channel. In particular, standard
recycling followed by dipole spin-down fails to reproduce the population of
millisecond pulsars with higher magnetic fields (B > 4 x 10^{8} G) at rates
deduced from observations. A proper inclusion of constraints arising from
binary evolution and mass accretion offers a more realistic insight into the
age distribution. By analytically implementing these constraints, I propose a
"modified" spin-down age for millisecond pulsars that gives estimates closer to
the true age. Finally, I independently analyze the peak, skewness and cutoff
values of the underlying mass distribution from a comprehensive list of radio
pulsars for which secure mass measurements are available. The inferred mass
distribution shows clear peaks at 1.35 Msun and 1.50 Msun for NSs in double
neutron star (DNS) and neutron star-white dwarf (NS-WD) systems respectively. I
find a mass cutoff at 2 Msun for NSs with WD companions, which establishes a
firm lower bound for the maximum mass of NSs.
|
In 2009, the BESIII experiment has collected about 225M $\jpsi$ and 106M
$\psip$ samples, both of which are the world largest on-peak charmonium
production. Based on these dataset, BESIII has made great effort on the study
of the charmonium decays, some important of which have been reviewed in this
proceeding. In addition, a searching for new physics through the $CP/P$
violation process is reported.
|
The three-dimensional structures of individual trees are important pieces of
information necessary to understand the effect of trees on urban environments.
In this study, we demonstrate a method for estimating the leaf area density
(LAD) distribution of individual trees using high-resolution airborne LiDAR.
This method improves upon the previously proposed method, which calculates LAD
based on the contact frequency between the laser beams and leaves by tracing
the paths of the laser beams. The proposed method in this study exploits the
last and intermediate pulses in addition to the first and single pulses to
capture the foliage distribution in the inner part of the crown. Each laser
beam is traced from a point derived by the last pulse to the point derived by
the first or intermediate pulse that is recorded immediately before the last
pulse. The laser beam interceptions and intersections can thus be accurately
reproduced while considering the last and intermediate pulses. We verify the
estimation accuracy of the three-dimensional LAD distribution using terrestrial
LiDAR data from a single tree (Z. serrata). The appropriate voxel size for
representing the LAD distribution from the airborne LiDAR is first determined
by comparing the distribution of voxels containing one or more airborne LiDAR
points with that containing one or more terrestrial LiDAR points. The estimated
LAD distribution with a voxel size of 1 m by 1 m by 0.5 m is subsequently
compared to the terrestrial LiDAR-derived LAD distribution. When only the first
and single pulses are used, the LAD is overestimated and underestimated in the
upper and lower part of the crown, respectively. We confirmed that using the
last and intermediate pulses improves the estimation accuracy of the entire
crown area.
|
Quantum Computing promises accelerated simulation of certain classes of
problems, in particular in plasma physics. Given the nascent interest in
applying quantum computing techniques to study plasma systems, a compendium of
the relevant literature would be most useful. As a novel field, new results are
common, and it is important for researchers to stay up-to-date on the latest
developments. With this in mind, the goal of this document is to provide a
regularly up-to-date and thorough list of citations for those developing and
applying these quantum computing approaches to experimental or theoretical work
in plasma physics. As a living document, it will be updated as often as
possible to incorporate the latest developments. References are grouped by
topic, both in itemized format and through the use of tags. We provide
instructions on how to participate, and suggestions are welcome.
|
Covering from photography to depth and spectral estimation, diverse
computational imaging (CI) applications benefit from the versatile modulation
of coded apertures (CAs). The light wave fields as space, time, or spectral can
be modulated to obtain projected encoded information at the sensor that is then
decoded by efficient methods, such as the modern deep learning decoders.
Despite the CA can be fabricated to produce an analog modulation, a binary CA
is mostly preferred since easier calibration, higher speed, and lower storage
are achieved. As the performance of the decoder mainly depends on the structure
of the CA, several works optimize the CA ensembles by customizing regularizers
for a particular application without considering critical physical constraints
of the CAs. This work presents an end-to-end (E2E) deep learning-based
optimization of CAs for CI tasks. The CA design method aims to cover a wide
range of CI problems easily changing the loss function of the deep approach.
The designed loss function includes regularizers to fulfill the widely used
sensing requirements of the CI applications. Mainly, the regularizers can be
selected to optimize the transmittance, the compression ratio, and the
correlation between measurements, while a binary CA solution is encouraged, and
the performance of the CI task is maximized in applications such as
restoration, classification, and semantic segmentation.
|
The mechanisms and properties of synchronization of oscillating ecological
populations attract attention because it is a fairly common phenomenon and
because spatial synchrony may elevate a risk of extinction and may lead to
other environmental impacts. Conditions for stable synchronization in a system
of linearly coupled predator-prey oscillators have been considered in the past.
However, the spatial dispersion coupling may be relatively weak and may not
necessarily lead to a stable, complete synchrony. If the coupling between
oscillators is too weak to induce a stable synchrony, oscillators may be
engaged into intermittent synchrony, when episodes of synchronized dynamics are
interspersed with the episodes of desynchronized dynamics. In the present study
we consider the temporal patterning of this kind of intermittent synchronized
dynamics in a system of two dispersal-coupled Rosenzweig-MacArthur
predator-prey oscillators. We consider the properties of the distributions of
durations of desynchronized intervals and their dependence on the model
parameters. We show that the temporal patterning of synchronous dynamics (an
ecological network phenomenon) may depend on the properties of individual
predator-prey patch (individual oscillator) and may vary independently of the
strength of dispersal. We also show that if the dynamics of predator is slow
relative to the dynamics of the prey (a situation that may promote brief but
large outbreaks), dispersal-coupled predator-prey oscillating populations
exhibit numerous short desynchronizations (as opposed to few long
desynchronizations) and may require weaker dispersal in order to reach strong
synchrony.
|
The Z Cam stars IW And and V513 Cas are unusual in having outbursts following
their standstills in contrast to the usual Z Cam behavior of quiescence
following standstills. In order to gain further understanding of these
little-studied systems, we obtained spectra correlated with photometry from the
AAVSO throughout a 3-4 month interval in 2011. In addition, time-resolved
spectra were obtained in 2012 that provided orbital periods of 3.7 hrs for IW
And and 5.2 hrs for V513 Cas. The photometry of V513 Cas revealed a regular
pattern of standstills and outbursts with little time at quiescence, while IW
And underwent many excursions from quiescence to outburst to short standstills.
The spectra of IW And are similar to normal dwarf novae, with strong Balmer
emission at quiescence and absorption at outburst. In contrast, V513 Cas shows
a much flatter/redder spectrum near outburst with strong HeII emission and
prominent emission cores in the Balmer lines. Part of this continuum difference
may be due to reddening effects. While our attempts to model the outburst and
standstill states of IW And indicate a mass accretion rate near 3E-9 solar
masses per year, we could find no obvious reason why these systems behave
differently following standstill compared to normal Z Cam stars.
|
This paper aims to quantitatively explain rationales of each prediction that
is made by a pre-trained convolutional neural network (CNN). We propose to
learn a decision tree, which clarifies the specific reason for each prediction
made by the CNN at the semantic level. I.e., the decision tree decomposes
feature representations in high conv-layers of the CNN into elementary concepts
of object parts. In this way, the decision tree tells people which object parts
activate which filters for the prediction and how much they contribute to the
prediction score. Such semantic and quantitative explanations for CNN
predictions have specific values beyond the traditional pixel-level analysis of
CNNs. More specifically, our method mines all potential decision modes of the
CNN, where each mode represents a common case of how the CNN uses object parts
for prediction. The decision tree organizes all potential decision modes in a
coarse-to-fine manner to explain CNN predictions at different fine-grained
levels. Experiments have demonstrated the effectiveness of the proposed method.
|
We report Molecular Dynamics (MD) simulations of a generic hydrophobic
nanopore connecting two reservoirs which are initially at different Na+
concentrations, as in a biological cell. The nanopore is impermeable to water
under equilibrium conditions, but the strong electric field caused by the ionic
concentration gradient drives water molecules in. The density and structure of
water in the pore are highly field dependent. In a typical simulation run, we
observe a succession of cation passages through the pore, characterized by
approximately bulk mobility. These ion passages reduce the electric field,
until the pore empties of water and closes to further ion transport, thus
providing a possible mechanism for biological ion channel gating.
|
We study the equational theory of the Weihrauch lattice with multiplication,
meaning the collection of equations between terms built from variables, the
lattice operations $\sqcup$, $\sqcap$, the product $\times$, and the finite
parallelization $(-)^*$ which are true however we substitute Weihrauch degrees
for the variables. We provide a combinatorial description of these in terms of
a reducibility between finite graphs, and moreover, show that deciding which
equations are true in this sense is complete for the third level of the
polynomial hierarchy.
|
The gravitational production of superheavy dark matter, in the
Peebles-Vilenkin quintessential inflation model, is studied in two different
scenarios: When the particles, whose decay products reheat the universe after
the end of the inflationary period, are created gravitationally, and when are
produced via instant preheating. We show that the viability of both scenarios
requires that the mass of the superheavy dark matter to be approximately
between 10^{16} and 10^{17} GeV.
|
Solar flares are some of the most energetic events in the solar system and
can be studied to investigate the physics of plasmas and stellar processes. One
interesting aspect of solar flares is the presence of accelerated (nonthermal)
particles, whose signatures appear in solar flare hard X-ray emissions. Debate
has been ongoing since the early days of the space age as to how these
particles are accelerated, and one way to probe relevant acceleration
mechanisms is by investigating short-timescale (tens of milliseconds)
variations in solar flare hard X-ray flux. The Impulsive Phase Rapid Energetic
Solar Spectrometer (IMPRESS) CubeSat mission aims to measure these fast hard
X-ray variations. In order to produce the best possible science data from this
mission, we characterize the IMPRESS scintillator detectors using Geant4 Monte
Carlo models. We show that the Geant4 Monte Carlo detector model is consistent
with an analytical model. We find that Geant4 simulations of X-ray and optical
interactions explain observed features in experimental data, but do not
completely account for our measured energy resolution. We further show that
nonuniform light collection leads to double-peak behavior at the 662 keV
$^{137}$Cs photopeak and can be corrected in Geant4 models and likely in the
lab.
|
Computing in-memory (CiM) has emerged as an attractive technique to mitigate
the von-Neumann bottleneck. Current digital CiM approaches for in-memory
operands are based on multi-wordline assertion for computing bit-wise Boolean
functions and arithmetic functions such as addition. However, most of these
techniques, due to the many-to-one mapping of input vectors to bitline
voltages, are limited to CiM of commutative functions, leaving out an important
class of computations such as subtraction. In this paper, we propose a CiM
approach, which solves the mapping problem through an asymmetric wordline
biasing scheme, enabling (a) simultaneous single-cycle memory read and CiM of
primitive Boolean functions (b) computation of any Boolean function and (c) CiM
of non-commutative functions such as subtraction and comparison. While the
proposed technique is technology-agnostic, we show its utility for
ferroelectric transistor (FeFET)-based non-volatile memory. Compared to the
standard near-memory methods (which require two full memory accesses per
operation), we show that our method can achieve a full scale two-operand
digital CiM using just one memory access, leading to a 23.2% - 72.6% decrease
in energy-delay product (EDP).
|
The proliferation of social media platforms has fueled the rapid
dissemination of fake news, posing threats to our real-life society. Existing
methods use multimodal data or contextual information to enhance the detection
of fake news by analyzing news content and/or its social context. However,
these methods often overlook essential textual news content (articles) and
heavily rely on sequential modeling and global attention to extract semantic
information. These existing methods fail to handle the complex, subtle twists
in news articles, such as syntax-semantics mismatches and prior biases, leading
to lower performance and potential failure when modalities or social context
are missing. To bridge these significant gaps, we propose a novel multi-hop
syntax aware fake news detection (MSynFD) method, which incorporates
complementary syntax information to deal with subtle twists in fake news.
Specifically, we introduce a syntactical dependency graph and design a
multi-hop subgraph aggregation mechanism to capture multi-hop syntax. It
extends the effect of word perception, leading to effective noise filtering and
adjacent relation enhancement. Subsequently, a sequential relative
position-aware Transformer is designed to capture the sequential information,
together with an elaborate keyword debiasing module to mitigate the prior bias.
Extensive experimental results on two public benchmark datasets verify the
effectiveness and superior performance of our proposed MSynFD over
state-of-the-art detection models.
|
In this paper, we prove a theorem on the rate of convergence for the optimal
cost computed using PS methods. It is a first proved convergence rate in the
literature of PS optimal control. In addition to the high-order convergence
rate, two theorems are proved for the existence and convergence of the
approximate solutions. This paper contains several essential differences from
existing papers on PS optimal control as well as some other direct
computational methods. The proofs do not use necessary conditions of optimal
control. Furthermore, we do not make coercivity type of assumptions. As a
result, the theory does not require the local uniqueness of optimal solutions.
In addition, a restrictive assumption on the cluster points of discrete
solutions made in existing convergence theorems are removed.
|
We obtain existence, uniqueness, and stability results for the modified
1-homogeneous infinity Laplace equation \[ -\Delta_\infty u - \beta |Du| = f,
\] subject to Dirichlet or mixed Dirichlet-Neumann boundary conditions. Our
arguments rely on comparing solutions of the PDE to subsolutions and
supersolutions of a certain finite difference approximation.
|
Pseudo-Goldstone bosons in 4D strongly coupled theories have a dual
description in terms of 5D gauge theories in warped backgrounds. We introduce
systematic methods of computing the pseudo-Goldstone potential for an arbitrary
warp factor in 5D. When applied to electroweak symmetry breaking, our approach
clarifies the relation of physical observables to geometrical quantities in
five dimensions.
|
In this paper we determine the radius of convexity for three kind of
normalized Bessel functions of the first kind. In the mentioned cases the
normalized Bessel functions are starlike-univalent and convex-univalent,
respectively, on the determined disks. The key tools in the proofs of the main
results are some new Mittag-Leffler expansions for quotients of Bessel
functions of the first kind, special properties of the zeros of Bessel
functions of the first kind and their derivative, and the fact that the
smallest positive zeros of some Dini functions are less than the first positive
zero of the Bessel function of the first kind. Moreover, we find the optimal
parameters for which these normalized Bessel functions are convex in the open
unit disk. In addition, we disprove a conjecture of Baricz and Ponnusamy
concerning the convexity of the Bessel function of the first kind.
|
Persistent homology and persistent entropy have recently become useful tools
for patter recognition. In this paper, we find requirements under which
persistent entropy is stable to small perturbations in the input data and scale
invariant. In addition, we describe two new stable summary functions combining
persistent entropy and the Betti curve. Finally, we use the previously defined
summary functions in a material classification task to show their usefulness in
machine learning and pattern recognition.
|
I discuss some theoretical aspects of how to observe leptonic CP violation.
It is divided into two parts, one for CP violation due to Majorana, and the
other more conventional leptonic Kobayashi-Maskawa (KM) phases. In the first
part, I estimate the effect of Majorana phase to observable of neutrinoless
double beta decay experiments by paying a careful attention to the definition
of the atmospheric scale Delta m^2. In the second part, I discuss
Tokai-to-Kamioka-Korea two detector complex which receives neutrino superbeam
from J-PARC as a concrete setting for discovering CP violation due to the KM
phase, as well as resolving mass hierarchy and the theta_{23} octant
degeneracy. A cautionary remark is also given on comparison between various
projects aiming at exploring CP violation and the mass hierarchy.
|
Automated Program Repair (APR) is defined as the process of fixing a
bug/defect in the source code, by an automated tool. APR tools have recently
experienced promising results by leveraging state-of-the-art Neural Language
Processing (NLP) techniques. APR tools such as TFix and CodeXGLUE combine
text-to-text transformers with software-specific techniques are outperforming
alternatives, these days. However, in most APR studies the train and test sets
are chosen from the same set of projects. In reality, however, APR models are
meant to be generalizable to new and different projects. Therefore, there is a
potential threat that reported APR models with high effectiveness perform
poorly when the characteristics of the new project or its bugs are different
than the training set's(Domain Shift).
In this study, we first define and measure the domain shift problem in
automated program repair. Then, we then propose a domain adaptation framework
that can adapt an APR model for a given target project. We conduct an empirical
study with three domain adaptation methods FullFineTuning,
TuningWithLightWeightAdapterLayers, and CurriculumLearning using two
state-of-the-art domain adaptation tools (TFix and CodeXGLUE) and two APR
models on 611 bugs from 19 projects. The results show that our proposed
framework can improve the effectiveness of TFix by 13.05% and CodeXGLUE by
23.4%. Another contribution of this study is the proposal of a data synthesis
method to address the lack of labelled data in APR. We leverage transformers to
create a bug generator model. We use the generated synthetic data to domain
adapt TFix and CodeXGLUE on the projects with no data (Zero-shot learning),
which results in an average improvement of 5.76% and 24.42% for TFix and
CodeXGLUE, respectively.
|
We investigated the representation thoery of an Ariki-Koike algebra whose
Poincare polynomial associated with the "bottom", i.e., the subgroup on which
the symmetric group acts, is non-zero in the base field. We proved that the
module category of such an Ariki-Koike algebra is Morita equivalent to the
module category of a direct sum of tensor products of Hecke algebras associated
with certain symmetric groups. We also generalized this Morita equivalence
theorem to give a Morita equivalenve between a $q$-Schur$^m$ algebra and a
direct sum of tensor products of certain $q$-Schur algebras.
|
Short pulse lasers are used to characterize the nonlinear response of
amplified photodetectors. Two widely used balanced detectors are characterized
in terms of amplitude, area, broadening, and balancing the mismatch of their
impulse response. The dynamic impact of pulses on the detector is also
discussed. It is demonstrated that using photodetectors with short pulses
triggers nonlinearities even when the source average power is well below the
detector continuous power saturation threshold.
|
We investigate the lifetime of macroscopic entanglement under the influence
of decoherence. For GHZ-type superposition states we find that the lifetime
decreases with the size of the system (i.e. the number of independent degrees
of freedom) and the effective number of subsystems that remain entangled
decreases with time. For a class of other states (e.g. cluster states),
however, we show that the lifetime of entanglement is independent of the size
of the system.
|
Electronic structure across the metal-insulator (MI) transition of
electron-doped V1-xWxO2 epitaxial films (x = 0-0.06) grown on alfa-Al2O3
substrates was studied by means of thermopower (S) measurements. Significant
increase of |S|-values accompanied by MI transition was observed, and the
transition temperatures of S (TS) decreased with x in good linear relation with
MI transition temperatures. |S| values of V1-xWxO2 films at T > TS were
constant at low values of 23 microV K-1 independently of x, which reflects a
metallic electronic structure, whereas, those at T < TS almost linearly
decreased with logarithmic W-concentrations. The gradient of -213 microV K-1
agrees well with -kB/e*ln10 (-198 microV K-1), suggesting that V1-xWxO2 films
have insulating electronic structures with a parabolic density of state around
the conduction band bottom.
|
It is shown that the internal stationary state of the Schwarzschild black
hole can be represented by a maximally entangled two-mode squeezed state of
collapsing matter and infalling Hawking radiation. The final boundary condition
at the singularity is then described by the random unitary transformation
acting on the collapsing matter field. The outgoing Hawking radiation is
obtained by the final state projection on the total wave function, which looks
like a quantum teleportation process without the classical information
transmitted. The black hole evaporation process as seen by the observer outside
the black hole is now a unitary process but non-local physics is required to
transmit the information outside the black hole. It is also shown that the
final state projection by the evaporation process is strongly affected by the
quantum state outside the event horizon, which clearly violates the locality
principle.
|
The relative partition function and the relative zeta function of the
perturbation of the Laplace operator by a Coulomb potential plus a point
interaction centered in the origin is discussed. Applications to the study of
the Casimir effect are indicated.
|
We present here a relationship among massive self-dual models for spin-3
particles in $D=2+1$ via the master action procedure. Starting with a first
order model (in derivatives) $S_{SD(1)}$ we have constructed a master action
which interpolates among a sequence of four self-dual models $S_{SD(i)}$ where
$i=1,2,3,4$. By analyzing the particle content of mixing terms, we give
additional arguments that explain why it is apparently impossible to jump from
the fourth order model to a higher order model. We have also analyzed
similarities and differences between the fourth order $K$-term in the spin-2
case and the analogous fourth order term in the spin-3 context.
|
Several aspects of classical and quantum mechanics applied to a class of
strongly chaotic systems are studied. These consist of single particles moving
without external forces on surfaces of constant negative Gaussian curvature
whose corresponding fundamental groups are supplied with an arithmetic
structure. It is shown that the arithmetic features of the considered systems
lead to exceptional properties of the corresponding spectra of lengths of
periodic orbits. The most significant one is an exponential growth of
degeneracies in these length spectra. Furthermore, the arithmetical systems are
distinguished by a structure that appears as a generalization of geometric
symmetries. These pseudosymmetries occur in the quantization of the classical
arithmetic systems as Hecke operators, which form an infinite algebra of
self-adjoint operators commuting with the Hamiltonian. The statistical
properties of quantum energies in the arithmetical have previously been
identified as exceptional. They do not fit into the general scheme of random
matrix theory. It is shown with the help of a simplified model for the spectral
form factor how the spectral statistics in arithmetic quantum chaos can be
understood by the properties of the corresponding classical length spectra. A
decisive is played by the exponentially increasing multiplicities of lengths.
The model developed for the level spacings distribution and for the number
variance is compared to the corresponding quantities obtained from quantum
energies for a specific arithmetical system.
|
We obtain Margulis-type asymptotic estimates for the number of free homotopy
classes of closed geodesics on certain manifolds without conjugate points. Our
results cover all compact surfaces of genus at least 2 without conjugate
points.
|
Production of strange quarks in neutron stars is investigated in this work.
Three cases, one in which the energy and neutrinos produced in the strangeness
production reactions are retained in the reaction region, second in which the
neutrinos are allowed to escape the reaction region but the energy is retained
and the third in which both the energy and neutrinos escape the reaction region
are considered. It is shown that the nonleptonic weak process dominates strange
quark production while semileptonic weak processes, which produce neutrinos,
lead to the cooling if the neutrinos escape the reaction region. It is found
that the time required for the saturation of the strangeness fraction is
between $10^{-7}$ and $10^{-5}$ sec, with the shorter time corresponding to the
first two cases. About 0.2 neutrinos/baryon are emitted during the process in
the first two cases where as the neutrino emission is somewhat suppressed in
the last case. The average energy of the neutrinos produced in all the three
cases is found to be several hundred $MeV$. We also find that a large amount of
energy is released during the strangeness production in the first two cases and
this leads to the heating of the reaction region. Implications of the neutrino
production are investigated.
|
In this dissertation we study the coefficients spaces (SAYD modules) of
Hopf-cyclic cohomology theory over a certain family of bicrossed product Hopf
algebras, and we compute the Hopf-cyclic cohomology of such Hopf algebras with
coefficients. We associate a Hopf algebra, what we call a Lie-Hopf algebra, to
any matched pair of Lie groups, Lie algebras and affine algebraic groups via
the semi-dualization procedure of Majid. We then identify the SAYD modules over
Lie-Hopf algebras with the representations and corepresentations of the total
Lie group, Lie algebra or the affine algebraic group of the matched pair. First
we classify the SAYD modules that correspond only to the representations of a
total Lie group (algebra). We call them induced SAYD modules. We then
generalize this identification, focusing on the matched pair of Lie algebras.
We establish a one-to-one correspondence between the SAYD modules over the
Lie-Hopf algebra associated to a matched pair of Lie algebras and certain SAYD
modules over the total Lie algebra. Once the SAYD modules are associated to the
representations and the corepresentations of Lie algebras, nontrivial examples
can be constructed. This way, we illustrate a highly nontrivial 4-dimensional
SAYD module over the Schwarzian Hopf algebra H_{1S}. In addition, we discuss
the periodic cyclic cohomology of Lie-Hopf algebras with nontrivial SAYD
coefficients. We obtain a general van Est isomorphism identifying the periodic
cyclic cohomology of a Lie-Hopf algebra with the (relative) Lie algebra
cohomology of the corresponding total Lie algebra.
|
Blue noise error patterns are well suited to human perception, and when
applied to stochastic rendering techniques, blue noise masks (blue noise
textures) minimize unwanted low-frequency noise in the final image. Current
methods of applying blue noise masks at each frame independently produce white
noise frequency spectra temporally. This white noise results in slower
integration convergence over time and unstable results when filtered
temporally. Unfortunately, achieving temporally stable blue noise distributions
is non-trivial since 3D blue noise does not exhibit the desired 2D blue noise
properties, and alternative approaches degrade the spatial blue noise
qualities.
We propose novel blue noise patterns that, when animated, produce values at a
pixel that are well distributed over time, converge rapidly for Monte Carlo
integration, and are more stable under TAA, while still retaining spatial blue
noise properties. To do so, we propose an extension to the well-known void and
cluster algorithm that reformulates the underlying energy function to produce
spatiotemporal blue noise masks.
These masks exhibit blue noise frequency spectra in both the spatial and
temporal domains, resulting in visually pleasing error patterns, rapid
convergence speeds, and increased stability when filtered temporally. We
demonstrate these improvements on a variety of applications, including
dithering, stochastic transparency, ambient occlusion, and volumetric
rendering.
By extending spatial blue noise to spatiotemporal blue noise, we overcome the
convergence limitations of prior blue noise works, enabling new applications
for blue noise distributions.
|
The established microalgae growth models are semi-empirical or considerable
fitting coefficients exist currently. Therefore, the ability of the model
prediction is reduced by the numerous fitting coefficients. Furthermore, the
predicted results of the established models are dependent on the size of the
photobioreactor (PBR), light intensity, flow and concentration field. The
growth mechanism of microalgae has not clearly understood in PBR cultivation.
It is difficult to predict the microalgae growth by theoretical methods, owing
to the aforementioned factors. We developed an exploratory bridging microalgae
growth model to predict the microalgae growth rate in PBRs by using the
nondimensional method which is effectively in fluid dynamics and heat transfer.
The analytical solution of the growth rate was obtained for the parallel flow.
The nondimensional growth rate expressed as function of Reynolds number and
Schmidt number, which can be used for arbitrary parallel flow due to the
solution was expressed as nondimensional quantities. The theoretically
predicted growth rate is compared with the experimentally measured microalgae
growth rate on the order of magnitude. The nondimensional method successfully
applied to the microalgae growth problem for the first time. The general
nondimensional solution can unify the numerous experimental data for different
laboratory conditions, and give a direction for the disorder of the microalgae
growth problem. The nondimensional solution may be useful to explain the growth
mechanism of microalgae and design large-scale PBRs for microalgae biofuel
production. The significance of the work is to give a theoretical foundation
and methodology of biological theory of microalgae growth.
|
As Clouds are complex, large-scale, and heterogeneous distributed systems,
management of their resources is a challenging task. They need automated and
integrated intelligent strategies for provisioning of resources to offer
services that are secure, reliable, and cost-efficient. Hence, effective
management of services becomes fundamental in software platforms that
constitute the fabric of computing Clouds. In this direction, this paper
identifies open issues in autonomic resource provisioning and presents
innovative management techniques for supporting SaaS applications hosted on
Clouds. We present a conceptual architecture and early results evidencing the
benefits of autonomic management of Clouds.
|
In a Palatini $f(\mathcal{R})$-model, we define chonodynamical effects due to
the choice of atomic clocks as standard reference clocks and we develop a
formalism able to quantitatively separate them from the usual effective dark
sources one has in extended theories. We apply the formalism to Hubble drift
and briefly discuss the issue about the physical frame. In particular, we argue
that there is no physical frame in the sense one does different things in
different frames and that, in a sense, is the physical characteristic of
extended gravity. As an example, we discuss how Jordan frame may be well suited
to discuss cosmology, though it fails within the solar system.
|
The relations between quantum coherence and quantum interference are
discussed. A general method for generation of quantum coherence through
interference-induced state selection is introduced and then applied to `simple'
atomic systems under two-photon transitions, with applications in quantum
optics and laser cooling.
|
In this paper we develop some new variational principles for the exit time of
non-symmetric diffusions from a domain. As applications, we give some
comparison theorems and monotonicity law between different diffusions.
|
Flow signatures in experimental data from relativistic ion collisions are
usually interpreted as a fingerprint of the presence of a hydrodynamic phase
during the evolution of these systems. In this work, flow signatures arising
from event-by-event viscous hydrodynamics are compared to those arising from
event-by-event non-interacting particle dynamics (free-streaming), both
followed by a late-stage hadronic cascade, in d+Au, 3He+Au at sqrt(s)=200 GeV
and p+Pb collisions at sqrt(s)=5 TeV, respectively. For comparison, also Pb+Pb
collisions at sqrt(s)=2.76 TeV are simulated. It is found that non-hydrodynamic
evolution can give rise to equal or larger radial flow than hydrodynamics with
eta/s=0.08 in all simulated collision systems. In light-on-heavy-ion
collisions, free-streaming gives rise to triangular and quadrupolar flow
comparable to or larger than that from hydrodynamics, but it generally leads to
considerably smaller elliptic flow. As expected, free-streaming leads to
considerably less elliptic, triangular and quadrupolar flow than hydrodynamics
in nucleus-nucleus collisions, such as event-by-event Pb+Pb collisions at
sqrt(s)=2.76 TeV.
|
We briefly review the covariant formulation of the Green-Schwarz superstring
by Berkovits, and describe how a detailed tree-level and one-loop analysis of
this model leads, for the first time, to a derivation of the low-energy
effective action of the heterotic superstring while keeping target-space
supersymmetry manifest. The resulting low-energy theory is old-minimal
supergravity coupled to tensor multiplet. The dilaton is part of the
compensator multiplet.
|
We continue the study initiated in [F. Albiac and P. Wojtaszczyk,
Characterization of $1$-greedy bases, J. Approx. Theory 138 (2006), no. 1,
65-86] of properties related to greedy bases in the case when the constants
involved are sharp, i.e., in the case when they are equal to $1$. Our main goal
here is to provide an example of a Banach space with a basis that satisfies
Property (A) but fails to be $1$-suppression unconditional, thus settling
Problem 4.4 from [F. Albiac and J.L. Ansorena, Characterization of $1$-almost
greedy bases, Rev. Mat. Complut. 30 (2017), no. 1, 13-24]. In particular, our
construction demonstrates that bases with Property (A) need not be $1$-greedy
even with the additional assumption that they are unconditional and symmetric.
We also exhibit a finite-dimensional counterpart of this example and show that,
at least in the finite-dimensional setting, Property (A) does not pass to the
dual. As a by-product of our arguments, we prove that a symmetric basis is
unconditional if and only if it is total, thus generalizing the well-known
result that symmetric Schauder bases are unconditional.
|
We consider the incompressible Euler equations in $R^2$ when the initial
vorticity is bounded, radially symmetric and non-increasing in the radial
direction. Such a radial distribution is stationary, and we show that the
monotonicity produces stability in some weighted norm related to the angular
impulse. For instance, it covers the cases of circular vortex patches and
Gaussian distributions. Our stability does not depend on $L^\infty$-bound or
support size of perturbations. The proof is based on the fact that such a
radial monotone distribution minimizes the impulse of functions having the same
level set measure.
|
In the present paper, we will show that a $(p,q,r)$-pretzel knot has the
representativity 3 if and only if $(p,q,r)$ is either $\pm(-2,3,3)$ or
$\pm(-2,3,5)$. We also show that a large algebraic knot has the
representativity less than or equal to 3.
|
We discuss what is light-cone quantization on a curved spacetime also without
a null Killing vector. Then we consider as an example the light-cone
quantization of a scalar field on a background with a Killing vector and the
connection with the second quantization of the particle in the same background.
It turns out that the proper way to define the light-cone quantization is to
require that the constant light-cone time hypersurface is null or,
equivalently, that the particle Hamiltonian is free of square roots. Moreover,
in order to quantize the scalar theory it is necessary to use not the original
scalar rather a scalar field density, i.e. the Schr\"odinger wave functional
depends on a scalar density and not on the original field. Finally we recover
this result as the second quantization of a particle on the same background,
where it is necessary to add as input the fact that we are dealing with a
scalar density.
|
A $\lambda$-translating soliton with density vector $\vec{v}$ is a surface
$\Sigma$ in Euclidean space ${\mathbb R}^3$ whose mean curvature $H$ satisfies
$2H=2\lambda+\langle N,\vec{v}\rangle$, where $N$ is the Gauss map of $\Sigma$.
In this article we study the shape of a compact $\lambda$-translating soliton
in terms of its boundary. If $\Gamma$ is a given closed curve, we deduce under
what conditions on $\lambda$ there exists a compact $\lambda$-translating
soliton $\Sigma$ with boundary $\Gamma$ and we provide estimates of the surface
area in relation with the height of $\Sigma$. Finally we study the shape of
$\Sigma$ related with the one of $\Gamma$, in particular, we give conditions
that assert that $\Sigma$ inherits the symmetries of its boundary $\Gamma$.
|
Human identification is one of the most common and critical tasks for
condition monitoring, human-machine interaction, and providing assistive
services in smart environments. Recently, human gait has gained new attention
as a biometric for identification to achieve contactless identification from a
distance robust to physical appearances. However, an important aspect of gait
identification through wearables and image-based systems alike is accurate
identification when limited information is available, for example, when only a
fraction of the whole gait cycle or only a part of the subject body is visible.
In this paper, we present a gait identification technique based on temporal and
descriptive statistic parameters of different gait phases as the features and
we investigate the performance of using only single gait phases for the
identification task using a minimum number of sensors. It was shown that it is
possible to achieve high accuracy of over 95.5 percent by monitoring a single
phase of the whole gait cycle through only a single sensor. It was also shown
that the proposed methodology could be used to achieve 100 percent
identification accuracy when the whole gait cycle was monitored through pelvis
and foot sensors combined. The ANN was found to be more robust to fewer data
features compared to SVM and was concluded as the best machine algorithm for
the purpose.
|
Massive multiple-input multiple-output (MIMO) enjoys great advantage in 5G
wireless communication systems owing to its spectrum and energy efficiency.
However, hundreds of antennas require large volumes of pilot overhead to
guarantee reliable channel estimation in FDD massive MIMO system. Compressive
sensing (CS) has been applied for channel estimation by exploiting the inherent
sparse structure of massive MIMO channel but suffer from high complexity. To
overcome this challenge, this paper develops a hybrid channel estimation scheme
by integrating the model-driven CS and data-driven deep unrolling technique.
The proposed scheme consists of a coarse estimation part and a fine correction
part to respectively exploit the inter- and intraframe sparsities of channels
to greatly reduce the pilot overhead. Theoretical result is provided to
indicate the convergence of the fine correction and coarse estimation net.
Simulation results are provided to verify that our scheme can estimate MIMO
channels with low pilot overhead while guaranteeing estimation accuracy with
relatively low complexity.
|
In this paper we consider $\rho$-Einstein solitons of type $M= \left(B^n,
g^{*}\right) \times (F^m,g_F)$, where $\left(B^n,g^{*}\right)$ is conformal to
a pseudo-Euclidean space and invariant under the action of the
pseudo-orthogonal group, and $\left(F^m,g_{F}\right)$ is an Einstein manifold.
We provide all the solutions for the gradient Schouten soliton case. Moreover,
in the Riemannian case, we prove that if
$M= \left(B^n, g^{*}\right) \times (F^m,g_F)$ is a complete gradient Schouten
soliton then $\left(B^{n},g^{*}\right)$ is isometric to $\mathbb{S}^{n-1}\times
\mathbb{R}$ and $F^m$ is a compact Einstein manifold.
|
Cross-contrast image translation is an important task for completing missing
contrasts in clinical diagnosis. However, most existing methods learn separate
translator for each pair of contrasts, which is inefficient due to many
possible contrast pairs in real scenarios. In this work, we propose a unified
Hyper-GAN model for effectively and efficiently translating between different
contrast pairs. Hyper-GAN consists of a pair of hyper-encoder and hyper-decoder
to first map from the source contrast to a common feature space, and then
further map to the target contrast image. To facilitate the translation between
different contrast pairs, contrast-modulators are designed to tune the
hyper-encoder and hyper-decoder adaptive to different contrasts. We also design
a common space loss to enforce that multi-contrast images of a subject share a
common feature space, implicitly modeling the shared underlying anatomical
structures. Experiments on two datasets of IXI and BraTS 2019 show that our
Hyper-GAN achieves state-of-the-art results in both accuracy and efficiency,
e.g., improving more than 1.47 and 1.09 dB in PSNR on two datasets with less
than half the amount of parameters.
|
While teaching a course on integral equations, I noticed that a
straightforward combination of Neumann series and Fourier series for the
resolvent (or the solution) of an integral equation has good approximation
qualities. This short article presents and investigates this combination of
approximating series.
|
Non-relativistic charged open strings coupled with Abelian gauge fields are
quantized in a geometric representation that generalizes the Loop
Representation. The model comprises open-strings interacting through a
Kalb-Ramond field in four dimensions. It is shown that a consistent
geometric-representation can be built using a scheme of ``surfaces and lines of
Faraday'', provided that the coupling constant (the ``charge'' of the string)
is quantized.
|
Nonlinear damping of parallel propagating Alfv\'en waves in high-$\beta$
plasma is considered. Trapping of thermal ions and Coulomb collisions are taken
into account. Saturated damping rate is calculated. Applications are made for
cosmic ray propagation in the Galaxy.
|
An action for a string and a particle with two timelike dimensions is
proposed and analyzed. Due to new gauge symmetries and associated constraints,
the motion of each system in the background of the other is equivalent to
effective motion with a single timelike dimension. The quantum constraints are
consistent only in critical dimensions. For the bosonic system in flat
spacetime the critical dimension is 27 or 28, with signature (25,2) or (26,2),
depending on whether the particle is massive or massless respectively. For the
supersymmetric case the critical dimensions are 11 or 12, with signature (9,2)
or (10,2), under the same circumstances. Generalizations to multiparticles,
strings and p-branes are outlined.
|
Given R groups of numerical variables X1, ... XR, we assume that each group
is the result of one underlying latent variable, and that all latent variables
are bound together through a linear equation system. Moreover, we assume that
some explanatory latent variables may interact pairwise in one or more
equations. We basically consider PLS Path Modelling's algorithm to estimate
both latent variables and the model's coefficients. New "external" estimation
schemes are proposed that draw latent variables towards strong group structures
in a more flexible way. New "internal" estimation schemes are proposed to
enable PLSPM to make good use of variable group complementarity and to deal
with interactions. Application examples are given.
|
We show that $3$-dimensional AdS spacetime can be semiclassically unstable
due to strongly interacting quantum field effects. In our previous paper, we
have pointed out the possibility of such an instability of AdS$_3$ by
inspecting linear perturbations of the (covering space of) static BTZ black
hole with AdS${}_4$ gravity dual in the context of holographic semiclassical
problems. In the present paper, we further study this issue from thermodynamic
viewpoint by constructing asymptotically AdS$_3$ semiclassical solutions and
computing free energies of the solutions. We find two asymptotically AdS${}_3$
solutions to the semiclassical Einstein equations with non-vanishing source
term: the one whose free energy is smaller than that of the BTZ with vanishing
source term and the other whose free energy is smaller than that of the global
AdS$_3$ with no horizon (thus manifestly zero-temperature background). The
instability found in this paper implies the breakdown of the maximal symmetries
of AdS$_3$, and its origin is different from the well-known semiclassical
linear instability since our holographic semiclassical Einstein equations in
$3$-dimensions do not involve higher order derivative terms.
|
Linear Support Vector Machines trained on HOG features are now a de facto
standard across many visual perception tasks. Their popularisation can largely
be attributed to the step-change in performance they brought to pedestrian
detection, and their subsequent successes in deformable parts models. This
paper explores the interactions that make the HOG-SVM symbiosis perform so
well. By connecting the feature extraction and learning processes rather than
treating them as disparate plugins, we show that HOG features can be viewed as
doing two things: (i) inducing capacity in, and (ii) adding prior to a linear
SVM trained on pixels. From this perspective, preserving second-order
statistics and locality of interactions are key to good performance. We
demonstrate surprising accuracy on expression recognition and pedestrian
detection tasks, by assuming only the importance of preserving such local
second-order interactions.
|
We show that the topological complexity of a finitely generated torsion free
hyperbolic group $\pi$ with $\cd\pi=n$ equals $2n$.
|
An arbitrary initial state of an optical or microwave field in a lossy driven
nonlinear cavity can be changed, in the steady-state limit, into a partially
incoherent superposition of only the vacuum and the single-photon states. This
effect is known as single-photon blockade, which is usually analyzed for a
Kerr-type nonlinear cavity parametrically driven by a single-photon process
assuming single-photon loss mechanisms. We study photon blockade engineering
via a squeezed reservoir, i.e., a quantum reservoir, where only two-photon
absorption is allowed. Namely, we analyze a lossy nonlinear cavity
parametrically driven by a two-photon process and allowing two-photon loss
mechanisms, as described by the master equation derived for a two-photon
absorbing reservoir. The nonlinear cavity engineering can be realized by a
linear cavity with a tunable two-level system via the Jaynes-Cummings
interaction in the dispersive limit. We show that by tuning properly the
frequencies of the driving field and the two-level system, the steady state of
the cavity field can be the single-photon Fock state or a partially incoherent
superposition of several Fock states with photon numbers, e.g., (0,2), (1,3),
(0,1,2), or (0,2,4). We observe that an arbitrary initial coherent or
incoherent superposition of Fock states with an even (odd) number of photons
can be changed into a partially incoherent superposition of a few Fock states
of the same photon-number parity. A general solution for an arbitrary initial
state is a weighted mixture of the above two solutions with even and odd photon
numbers, where the weights are given by the probabilities of measuring the even
and odd numbers of photons of the initial cavity field, respectively. Thus, in
contrast to the standard photon blockade, we prove that the steady state in the
engineered photon blockade, can depend on its initial state.
|
The effect of quenched disorder in a many-body system is experimentally
investigated in a controlled fashion. It is done by measuring the phase
synchronization (i.e. mutual coherence) of 400 coupled lasers as a function of
tunable disorder and coupling strengths. The results reveal that correlated
disorder has a non-trivial effect on the decrease of phase synchronization,
which depends on the ratio of the disorder correlation length over the average
number of synchronized lasers. The experimental results are supported by
numerical simulations and analytic derivations.
|
The structure of the $\Delta J = 1$ doublet bands in $^{128}Cs$ is
investigated within the framework of the Interacting Vector Boson Fermion Model
(IVBFM). A new, purely collective interpretation of these bands is given on the
basis of the used boson-fermion dynamical symmetry of the model. The energy
levels of the doublet bands as well as the absolute $B(E2)$ and $B(M1)$
transition probabilities between the states of both yrast and yrare bands are
described quite well. The observed odd-even staggering of both $B(M1)$ and
$B(E2)$ values is reproduced by the introduction of an appropriate interaction
term of quadrupole type, which produces such a staggering effect in the
transition strengths. The calculations show that the appearance of doublet
bands in certain odd-odd nuclei could be a consequence of the realization of a
larger dynamical symmetry based on the non-compact supersymmetry group
$OSp(2\Omega /12, R)$.
|
Subsets and Splits