text
stringlengths 6
128k
|
---|
A proposal to store unsteady energy in potential energy via lifting masses
with a rough quantitative overview. Some applications and methods to harvest
the potential energy are also given. A focus is put on photovoltaically
generated energy.
|
In recent years, image captioning and segmentation have emerged as crucial
tasks in computer vision, with applications ranging from autonomous driving to
content analysis. Although multiple solutions have emerged to help blind and
visually impaired people move around their environment, few are applications
that help them understand and rebuild a scene in their minds through text. Most
built models focus on helping users move and avoid obstacles, restricting the
number of environments blind and visually impaired people can be in.
In this paper, we will propose an approach that helps them understand their
surroundings using image captioning. The particularity of our research is that
we offer them descriptions with positions of regions and objects regarding them
(left, right, front), as well as positional relationships between regions,
while we aim to give them access to theatre plays by applying the solution to
our TS-RGBD dataset.
|
In this paper we investigate the electronic and magnetic properties of
K$_{x}$Fe$_{2-y}$Se$_{2}$ materials at different band fillings utilizing the
multi-orbital Kotliar-Ruckenstein's slave-boson mean field approach. We find
that at three-quarter filling, corresponding to KFe$_{2}$Se$_{2}$, the ground
state is a paramagnetic bad metal. Through band renormalization analysis and
comparison with the angle-resolved photoemission spectra data, we identify that
KFe$_{2}$Se$_{2}$ is also an intermediate correlated system, similar to
iron-pnictide systems. At two-third filling, corresponding to the
Fe$^{2+}$-based systems, the ground state is a striped antiferromagnetic (SAFM)
metal with spin density wave gap partially opened near the Fermi level. In
comparison, at half filling case, corresponding to the Fe$^{3+}$-based
compounds, besides SAFM, a $N\acute{e}el$ antiferromagnetic metallic ground
state without orbital ordering is observed in the intermediate correlation
range, and an orbital selective Mott phase (OSMP) accompanied with an
intermediate-spin to high-spin transition is also found. These results
demonstrate that the band filling and correlation control the electronic state,
Fermi surface topology and magnetism in K$_{x}$Fe$_{2-y}$Se$_{2}$.
|
Transformer-based pre-trained language models such as BERT have achieved
remarkable results in Semantic Sentence Matching. However, existing models
still suffer from insufficient ability to capture subtle differences. Minor
noise like word addition, deletion, and modification of sentences may cause
flipped predictions. To alleviate this problem, we propose a novel Dual
Attention Enhanced BERT (DABERT) to enhance the ability of BERT to capture
fine-grained differences in sentence pairs. DABERT comprises (1) Dual Attention
module, which measures soft word matches by introducing a new dual channel
alignment mechanism to model affinity and difference attention. (2) Adaptive
Fusion module, this module uses attention to learn the aggregation of
difference and affinity features, and generates a vector describing the
matching details of sentence pairs. We conduct extensive experiments on
well-studied semantic matching and robustness test datasets, and the
experimental results show the effectiveness of our proposed method.
|
Motivated by trapping and cooling experiments with non-spherical
nanoparticles, we discuss how their combined rotational and translational
quantum state is affected by the interaction with a gaseous environment. Based
on the quantum master equation in terms of orientation-dependent scattering
amplitudes, we evaluate the localization rate for gas collisions off an
anisotropic van der Waals-type potential and for photon scattering off an
anisotropic dielectric. We also show how pure angular momentum diffusion arises
from these open quantum dynamics in the limit of weak anisotropies.
|
We have found that at least seven hydrogen-deficient carbon (HdC) and R
Coronae Borealis (RCB) stars, have 16O/18O ratios close to and in some cases
less than unity, values that are orders of magnitude lower than measured in
other stars (the Solar value is 500). Greatly enhanced 18O is evident in every
HdC and RCB we have measured that is cool enough to have detectable CO bands.
The three HdC stars measured have 16O/18O < 1, lower values than any of the RCB
stars. These discoveries are important clues in determining the evolutionary
pathways of HdC and RCB stars, for which two models have been proposed: the
double degenerate (white dwarf (WD) merger), and the final helium-shell flash
(FF). No overproduction of 18O is expected in the FF scenario. We have
quantitatively explored the idea that HdC and RCB stars originate in the
mergers of CO- and He-WDs. The merger process is estimated to take only a few
days, with accretion rates of 150 Msun/ yr producing temperatures at the base
of the accreted envelope of 1.2 - 1.9 x 10^8 K. Analysis of a simplified
one-zone calculation shows that nucleosynthesis in the dynamically accreting
material may provide a suitable environment for a significant production of
18O, leading to very low values of 16O/18O, similar to those observed. We also
find qualitative agreement with observed values of 12C/13C and with the CNO
elemental ratios. H-admixture during the accretion process from the small
H-rich C/O WD envelope may play an important role in producing the observed
abundances. Overall our analysis shows that WD mergers may very well be the
progenitors of O18-rich RCB and HdC stars, and that more detailed simulations
and modeling are justified.
|
We propose an integrated photonics device for mapping qubits encoded in the
polarization of a photon onto the spin state of a solid-state defect coupled to
a photonic crystal cavity: a `Polarization-Encoded Photon-to-Spin Interface'
(PEPSI). We perform a theoretical analysis of the state fidelity's dependence
on the device's polarization extinction ratio and atom-cavity cooperativity.
Furthermore, we explore the rate-fidelity trade-off through analytical and
numerical models. In simulation, we show that our design enables efficient,
high fidelity photon-to-spin mapping.
|
Large language models generate fluent texts and can follow natural language
instructions to solve a wide range of tasks without task-specific training.
Nevertheless, it is notoriously difficult to control their generation to
satisfy the various constraints required by different applications. In this
work, we present InstructCTG, a controlled text generation framework that
incorporates different constraints by conditioning on natural language
descriptions and demonstrations of the constraints. In particular, we first
extract the underlying constraints of natural texts through a combination of
off-the-shelf NLP tools and simple heuristics. We then verbalize the
constraints into natural language instructions to form weakly supervised
training data. By prepending natural language descriptions of the constraints
and a few demonstrations, we fine-tune a pre-trained language model to
incorporate various types of constraints. Compared to existing search-based or
score-based methods, InstructCTG is more flexible to different constraint types
and has a much smaller impact on the generation quality and speed because it
does not modify the decoding procedure. Additionally, InstructCTG allows the
model to adapt to new constraints without re-training through the use of
few-shot task generalization and in-context learning abilities of
instruction-tuned language models.
|
The isoscalar and isovector particle densities and the surface tension
coefficients at the average binding energy are used to derive analytical
expressions of the neutron skin thickness and the isovector stiffness of sharp
edged proton-neutron asymmetric nuclei. For most Skyrme forces these quantities
are significantly larger than the well known ones. Using the analytical
isovector surface energy constants in the framework of the hydrodynamical and
the Fermi-liquid droplet models the mean energies and the sum rules of the
isovector giant dipole resonances are in fair agreement with the experimental
data.
|
This is one of the two papers where the optimized perturbation theory was
first formulated. The other paper is published in Theor. Math. Phys. 28,
652--660 (1976). The main idea of the theory is to reorganize the perturbative
sequence by introducing control functions, defined by optimization conditions,
so that the reorganized approximation sequence be convergent. In the present
paper, the theory of perturbations is suggested for statistical systems in the
absence of small interaction parameters. A new form is advanced for
self-consistent conditions defining the optimal parameters for trial Green
functions in iterating nonlinear propagator equations. Superharmonic,
semiharmonic, and pseudoharmonic approximations for a molecular crystal are
considered as examples.
|
We consider the general procedure for proving no-hair theorems for static,
spherically symmetric black holes. We apply this method to the abelian Higgs
model and find a proof of the no-hair conjecture that circumvents the
objections raised against the original proof due to Adler and Pearson.
|
I suggest an effective model between the GUT and the electroweak scale. It
only introduces the two symmetries of $U(1)_{B-L}$ and $U(1)_{D}$ besides the
SM groups. The two symmetries are individually broken at the reheating
temperature of the universe of $10^{12}$ GeV and the scale of $3\sim 4$ TeV.
The model can simultaneously accommodate the tiny neutrino masses, the
matter-antimatter asymmetry and the cold dark matter (CDM). In particular, the
model gives some interesting results and predictions, for instance, the
neutrinos are Dirac nature and their masses are related to the $U(1)_{D}$
breaking, the size of the matter-antimatter asymmetry is closely related to the
mass hierarchy of the quarks and charged leptons, the CDM mass is probably in
the range of $250\sim 350$ GeV. Finally, it is feasible to test the model in
future collider experiments.
|
The emergence of quantum-gravity induced corrective terms for the probability
of emission of a particle from a black hole in the Parikh-Wilczek tunneling
framework is studied. It is shown, in particular, how corrections might arise
from modifications of the surface gravity due to near horizon Planck-scale
effects. Our derivation provides an example of the possible linking between
Planck-scale departures from Lorentz invariance and the appearance of higher
order quantum gravity corrections in the black-hole entropy-area relation.
|
We present the results of the ethynyl (C2H) emission line observations
towards the HII regions S255 and S257 and the molecular cloud between them.
Radial profiles of line brightness, column density, and abundance of C2H are
obtained. We show that the radial profile of the ethynyl abundance is almost
flat towards the HII regions and drops by a factor of two towards the molecular
cloud. At the same time, we find that the abundance of ethynyl is at maximum
towards the point sources in the molecular cloud -- the stars with emission
lines or emitting in X-ray. The line profiles are consistent with the
assumption that both HII regions have front and back neutral walls that move
relative to each other.
|
We characterize the smallest codimension components of the Hodge locus of
smooth degree $d$ hypersurfaces of the projective space $\mathbb{P}^{n+1}$ of
even dimension $n$, passing through the Fermat variety (with $d\neq 3,4,6$).
They correspond to the locus of hypersurfaces containing a linear algebraic
cycle of dimension $\frac{n}{2}$. Furthermore, we prove that among all the
local Hodge loci associated to a non-linear cycle passing through Fermat, the
ones associated to a complete intersection cycle of type $(1,1,\ldots,1,2)$
attain the minimal possible codimension of their Zariski tangent spaces. This
answers a conjecture of Movasati, and generalizes a result of Voisin about the
first gap between the codimension of the components of the Noether-Lefschetz
locus to arbitrary dimension, provided that they contain the Fermat variety.
|
We consider two simple models for the formation of polymers where at the
initial time, each monomer has a certain number of potential links (called arms
in the text) that are consumed when aggregations occur. Loosely speaking, this
imposes restrictions on the number of aggregations. The dynamics of
concentrations are governed by modifications of Smoluchowski's coagulation
equations. Applying classical techniques based on generating functions,
resolution of quasi-linear PDE's, and Lagrange inversion formula, we obtain
explicit solutions to these non-linear systems of ODE's. We also discuss the
asymptotic behavior of the solutions and point at some connexions with certain
known solutions to Smoluchowski's coagulation equations with additive or
multiplicative kernels.
|
The field-dependent equilibrium thermodynamics is derived with two methods:
either by using the potential formalism either by the statistical method.
Therefore, Pontrjagin's extremum principle of control theory is applied to an
extended ensemble average. This approach allows to derive the grand partition
function of thermodynamics as a result of a control problem with the Hamilton
energy. Furthermore, the maximum entropy principle follows and the second law
in a modified form. The derivation can predict second law violations if cycles
with irreversibilities in varying potential fields are included into
consideration. This conclusion is supported indirectly by experimental data
from literature. The upper maximum gain efficiency of a cycle with a known
polymer solution as dielectrics was estimated to less than 1 promille per
cycle. Note added in proof 28th October 2003: Comparing this preprint work with
an analogous ferrofluidic system discrepancies are is found which show that the
concrete model proposed here in section 4 is insufficient to settle the
question. A way to solve the problem is proposed.
|
The first non-trivial case of Hadwiger's conjecture for oriented matroids
reads as follows. If $\mathcal{O}$ is an $M(K_4)$-free oriented matroid, then
$\mathcal{O}$ admits a NZ $3$-coflow, i.e., it is $3$-colourable in the sense
of Hochst\"attler-Ne\v{s}et\v{r}il. The class of gammoids is a class of
$M(K_4)$-free orientable matroids and it is the minimal minor-closed class that
contains all transversal matroids. Towards proving the previous statement for
the class of gammoids, Goddyn, Hochst\"attler, and Neudauer conjectured that
every gammoid has a positive coline (equivalently, a positive double circuit),
which implies that all orientations of gammoids are $3$-colourable. In this
brief note we disprove Goddyn, Hochst\"attler, and Neudauers' conjecture by
exhibiting a large class of bicircular matroids that do not contain positive
double circuits.
|
The calculation of the heating rate of cold atoms in vibrating traps requires
a theory that goes beyond the Kubo linear response formulation. If a strong
"quantum chaos" assumption does not hold, the analysis of transitions shows
similarities with a percolation problem in energy space. We show how the
texture and the sparsity of the perturbation matrix, as determined by the
geometry of the system, dictate the result. An improved sparse random matrix
model is introduced: it captures the essential ingredients of the problem, and
leads to a generalized variable range hopping picture.
|
A semiconductor quantum dot (QD) embedded within an optical microcavity is a
system of fundamental importance within quantum information processing. The
optimization of quantum coherence is crucial in such applications, requiring an
in-depth understanding of the relevant decoherence mechanisms. We provide
herein a critical review of prevalent theoretical treatments of the QD-cavity
system coupled to longitudinal acoustic phonons, comparing predictions against
a recently obtained exact solution. Within this review we consider a range of
temperatures and exciton-cavity coupling strengths. Predictions of the polaron
Nakajima-Zwanzig (NZ) and time-convolutionless (TCL) master equations, as well
as a variation of the former adapted for adiabatic continuous wave excitation
(CWE), are compared against an asymptotically exact solution based upon
Trotter's decomposition (TD) theorem. The NZ and TCL implementations, which
apply a polaron transformation to the Hamiltonian and subsequently treat the
exciton-cavity coupling to second order, do not offer a significant improvement
accuracy relative to the polaron transformation alone. The CWE adaptation
provides a marked improvement, capturing the broadband features of the
absorption spectrum (not present in NZ and TCL implementations). We attribute
this difference to the effect of the Markov approximation, and particularly its
unsuitability in pulsed excitation regime. Even the CWE adaptation, however,
breaks down in the regime of high temperature ($50K$) and strong exciton-cavity
coupling ($g \gtrsim 0.2$ meV). The TD solution is of comparable computational
complexity to the above-mentioned master equation approaches, yet remains
accurate at higher temperatures and across a broad range of exciton-cavity
coupling strengths (at least up to $g=1.5$ meV).
|
These notes are the second part of a common course on Renormalization Theory
given with Professor P. da Veiga at X Jorge Andre Swieca Summer School, Aguas
de Lindoia, Brazil, February 7-12, 1999. I emphasize the rigorous
non-perturbative or constructive aspects of the theory. The usual formalism for
the renormalization group in field theory or statistical mechanics is reviewed,
together with its limits. The constructive formalism is introduced step by
step. Taylor forest formulas allow to perform easily the cluster and Mayer
expansions which are needed for a single step of the renormalization group in
the case of Bosonic theories. The iteration of this single step leads to
further difficulties whose solution is briefly sketched. The second part of the
course is devoted to Fermionic models. These models are easier to treat on the
constructive level, so they are very well suited to beginners in constructive
theory. It is shown how the Taylor forest formulas allow to reorganize
perturbation theory nicely in order to construct the Gross-Neveu2 model without
any need for cluster or Mayer expansions. Finally applications of this
technique to condensed matter and renormalization group around Fermi surface
are briefly reviewed.
|
In this article, we study the blow-up of the damped wave equation in the
\textit{scale-invariant case} and in the presence of two nonlinearities. More
precisely, we consider the following equation: $$u_{tt}-\Delta
u+\frac{\mu}{1+t}u_t=|u_t|^p+|u|^q, \quad \mbox{in}\ \R^N\times[0,\infty), $$
with small initial data.\\ For $\mu < \frac{N(q-1)}{2}$ and $\mu \in (0,
\mu_*)$, where $\mu_*>0$ is depending on the nonlinearties' powers and the
space dimension ($\mu_*$ satisfies $(q-1)\left((N+2\mu_*-1)p-2\right) = 4$), we
prove that the wave equation, in this case, behaves like the one without
dissipation ($\mu =0$). Our result completes the previous studies in the case
where the dissipation is given by $\frac{\mu}{(1+t)^\beta}u_t; \ \beta >1$
(\cite{LT3}), where, contrary to what we obtain in the present work, the effect
of the damping is not significant in the dynamics. Interestingly, in our case,
the influence of the damping term $\frac{\mu}{1+t}u_t$ is important.
|
In higher Landau levels ($N>1$), the ground state of the two-dimensional
electron gas in a strong perpendicular magnetic field evolves from a Wigner
crystal for small filling $\nu $ of the partially filled Landau level, into a
succession of bubble states with increasing number of guiding centers per
bubble as $\nu $ increases, to a modulated stripe state near $\nu =0.5$. In
this work, we compute the frequency-dependent longitudinal conductivity $%
\sigma_{xx}(\omega) $ of the Wigner and bubble crystal states in the presence
of disorder. We apply an elastic theory to the crystal states which is
characterized by a shear and a bulk modulus. We obtain both moduli from the
microscopic time-dependent Hartree-Fock approximation. We then use the replica
and Gaussian variational methods to handle the effects of disorder. Within the
semiclassical approximation we get the dynamical conductivity as well as the
pinning frequency as functions of the Landau level filling factor and compare
our results with recent microwave experiments.
|
The phase structure of the Nambu -- Jona-Lasinio model at zero temperature
and in the presence of baryon- and isospin chemical potentials is investigated.
It is shown that in the chiral limit and for a wide range of model parameters
there exist two different phases with pion condensation. In the first, ordinary
phase, quarks are gapped particles. In the second, gapless pion condensation
phase, there is no energy cost for creating only $u$- or both $u$ and $d$
quarks, and the density of baryons is nonzero.
|
Random walk is one of the basic mechanisms found in many network
applications. We study the epidemic spreading dynamics driven by biased random
walks on complex networks. In our epidemic model, each time infected nodes
constantly spread some infected packets by biased random walks to their
neighbor nodes causing the infection of the susceptible nodes that receive the
packets. An infected node get recovered from infection with a fixed
probability. Simulation and analytical results on model and real-world networks
show that the epidemic spreading becomes intense and wide with the increase of
delivery capacity of infected nodes, average node degree, homogeneity of node
degree distribution. Furthermore, there are corresponding optimal parameters
such that the infected nodes have instantaneously the largest population, and
the epidemic spreading process covers the largest part of a network.
|
Observations on galactic scales seem to be in contradiction with recent high
resolution N-body simulations. This so-called cold dark matter (CDM) crisis has
been addressed in several ways, ranging from a change in fundamental physics by
introducing self-interacting cold dark matter particles to a tuning of complex
astrophysical processes such as global and/or local feedback. All these efforts
attempt to soften density profiles and reduce the abundance of satellites in
simulated galaxy halos. In this paper, we explore a somewhat different approach
which consists of filtering the dark matter power spectrum on small scales,
thereby altering the formation history of low mass objects. The physical
motivation for damping these fluctuations lies in the possibility that the dark
matter particles have a different nature i.e. are warm (WDM) rather than cold.
We show that this leads to some interesting new results in terms of the merger
history and large-scale distribution of low mass halos, as compared to the
standard CDM scenario. However, WDM does not appear to be the ultimate
solution, in the sense that it is not able to fully solve the CDM crisis, even
though one of the main drawbacks, namely the abundance of satellites, can be
remedied. Indeed, the cuspiness of the halo profiles still persists, at all
redshifts, and for all halos and sub-halos that we investigated. Despite the
persistence of the cuspiness problem of DM halos, WDM seems to be still worth
taking seriously, as it alleviates the problems of overabundant sub-structures
in galactic halos and possibly the lack of angular momentum of simulated disk
galaxies. WDM also lessens the need to invoke strong feedback to solve these
problems, and may provide a natural explanation of the clustering properties
and ages of dwarfs.
|
It is shown that wave function renormalization can introduce an important
contribution to the generation of baryon and lepton number asymmetries by heavy
particle decay. These terms, omitted in previous analyses, are of the same
order of magnitude as the standard terms. A complete cancellation of leading
terms can result in some interesting cases.
|
Difference equations, such as a Ricker map, for an increased value of the
parameter, experience instability of the positive equilibrium and transition to
deterministic chaos. To achieve stabilization, various methods can be applied.
Proportional Feedback control suggests a proportional reduction of the state
variable at every $k$th step. First, if $k \neq 1$, a cycle is stabilized
rather than an equilibrium. Second, the equation can incorporate an additive
noise term, describing the variability of the environment, as well as
multiplicative noise corresponding to possible deviations in the control
intensity. The present paper deals with both issues, it justifies a possibility
of getting a stable blurred $k$-cycle. Presented examples include the Ricker
model, as well as equations with unbounded $f$, such as the bobwhite quail
population models. Though the theoretical results justify stabilization for
either multiplicative or additive noise only, numerical simulations illustrate
that a blurred cycle can be stabilized when both multiplicative and additive
noises are involved.
|
The braneworld cosmology, in which our universe is imbedded as a hypersurface
in a higher dimensional bulk, has the peculiar property that the inflationary
consistency relation derived in a four-dimensional cosmology persists. This
consistency condition relates the ratio of tensor and scalar perturbation
amplitudes to the tensor spectral index produced during an epoch of slow-roll
scalar field inflation. We attempt to clarify this surprising degeneracy. Our
argument involves calculating the power spectrum of scalar field fluctuations
around geometries perturbed away from the exact de Sitter case. This
calculation is expected to be valid for perturbations which would not cause a
late-time acceleration of the universe. We use these results to argue that the
emergence of the same consistency relation in the braneworld can be connected
with a specific property, that five-dimensional observables smoothly approach
their four-dimensional counterparts as one takes the brane to infinite tension.
We exhibit an explicit example where this does not occur, and in which a
consistency relation does not persist.
|
Entangled states, like the two-mode squeezed vacuum state, are known to give
quantum advantage in the illumination protocol, a method to detect a weakly
reflecting target submerged in a thermal background. We use non-Gaussian
photon-added and -subtracted states, affected by local Gaussian noise on top of
the omnipresent thermal noise, as probes in the illumination protocol. Based on
the difference between the Chernoff bounds obtained with the coherent state and
the non-Gaussian state having equal signal strengths, whose positive values
denote quantum advantage in illumination, we highlight the hierarchy among
non-Gaussian states, which is compatible with correlations per unit signal
strength, although the Gaussian states offer the best performance.
Interestingly, such hierarchy is different when comparisons are made using the
Chernoff bounds. The entire analysis is performed in the presence of different
imperfect apparatus like faulty twin-beam generator, imperfect photon addition
(subtraction) as well as with noisy non-Gaussian probe states.
|
We introduce an approach to building a custom model from ready-made
self-supervised models via their associating instead of training and
fine-tuning. We demonstrate it with an example of a humanoid robot looking at
the mirror and learning to detect the 3D pose of its own body from the image it
perceives. To build our model, we first obtain features from the visual input
and the postures of the robot's body via models prepared before the robot's
operation. Then, we map their corresponding latent spaces by a sample-efficient
robot's self-exploration at the mirror. In this way, the robot builds the
solicited 3D pose detector, which quality is immediately perfect on the
acquired samples instead of obtaining the quality gradually. The mapping, which
employs associating the pairs of feature vectors, is then implemented in the
same way as the key-value mechanism of the famous transformer models. Finally,
deploying our model for imitation to a simulated robot allows us to study, tune
up, and systematically evaluate its hyperparameters without the involvement of
the human counterpart, advancing our previous research.
|
Optical control of the lateral quantum confinement and number of electrons
confined in nanofabricated GaAs/AlGaAs quantum dots is achieved by illumination
with a weak laser beam that is absorbed in the AlGaAs barrier. Precise tuning
of energy-level structure and electron population is demonstrated by monitoring
the low-lying transitions of the electrons from the lowest quantum-dot energy
shells by resonant inelastic light scattering. These findings open the way to
the manipulation of single electrons in these quantum dots without the need of
external metallic gates.
|
Video gaming streaming services are growing rapidly due to new services such
as passive video streaming, e.g. Twitch.tv, and cloud gaming, e.g. Nvidia
Geforce Now. In contrast to traditional video content, gaming content has
special characteristics such as extremely high motion for some games, special
motion patterns, synthetic content and repetitive content, which makes the
state-of-the-art video and image quality metrics perform weaker for this
special computer generated content. In this paper, we outline our plan to build
a deep learningbased quality metric for video gaming quality assessment. In
addition, we present initial results by training the network based on VMAF
values as a ground truth to give some insights on how to build a metric in
future. The paper describes the method that is used to choose an appropriate
Convolutional Neural Network architecture. Furthermore, we estimate the size of
the required subjective quality dataset which achieves a sufficiently high
performance. The results show that by taking around 5k images for training of
the last six modules of Xception, we can obtain a relatively high performance
metric to assess the quality of distorted video games.
|
There are strong evidences in the literature that quantum non-Markovianity
would hinder the presence of Quantum Darwinism. In this Letter, we study the
relation between quantum Darwinism and approximate quantum Markovianity for
open quantum systems by exploiting the properties of quantum conditional mutual
information. We show that for approximately Markovian quantum processes the
conditional mutual information still has the scaling property for Quantum
Darwinism. Then two general bounds on the backflow of information are obtained,
with which we can show that the presence of Quantum Darwinism restricts the
information backflow and the quantum non-Markovianity must be small.
|
In this paper we show that it is possible to derive the Kerr solution in an
alternative, intuitive way, based on physical reasoning and starting from an
orthogonal metric ansatz having manifest ellipsoidal space-time symmetry
(ellipsoidal symmetry). This is possible because both flat metric in oblate
spheroidal (ellipsoidal) coordinates and Kerr metric in Boyer-Lindquist
coordinates can be rewritten in such a form that the difference between the two
is only in the time-time and radial-radial metric tensor components, just as is
the case with Schwarzschild metric and flat metric in spherical coordinates.
|
An important political and social phenomena discussed in several countries,
like India and Brazil, is the use of WhatsApp to spread false or misleading
content. However, little is known about the information dissemination process
in WhatsApp groups. Attention affects the dissemination of information in
WhatsApp groups, determining what topics or subjects are more attractive to
participants of a group. In this paper, we characterize and analyze how
attention propagates among the participants of a WhatsApp group. An attention
cascade begins when a user asserts a topic in a message to the group, which
could include written text, photos, or links to articles online. Others then
propagate the information by responding to it. We analyzed attention cascades
in more than 1.7 million messages posted in 120 groups over one year. Our
analysis focused on the structural and temporal evolution of attention cascades
as well as on the behavior of users that participate in them. We found specific
characteristics in cascades associated with groups that discuss political
subjects and false information. For instance, we observe that cascades with
false information tend to be deeper, reach more users, and last longer in
political groups than in non-political groups.
|
Photo retouching aims at enhancing the aesthetic visual quality of images
that suffer from photographic defects such as over/under exposure, poor
contrast, inharmonious saturation. Practically, photo retouching can be
accomplished by a series of image processing operations. In this paper, we
investigate some commonly-used retouching operations and mathematically find
that these pixel-independent operations can be approximated or formulated by
multi-layer perceptrons (MLPs). Based on this analysis, we propose an extremely
light-weight framework - Conditional Sequential Retouching Network (CSRNet) -
for efficient global image retouching. CSRNet consists of a base network and a
condition network. The base network acts like an MLP that processes each pixel
independently and the condition network extracts the global features of the
input image to generate a condition vector. To realize retouching operations,
we modulate the intermediate features using Global Feature Modulation (GFM), of
which the parameters are transformed by condition vector. Benefiting from the
utilization of $1\times1$ convolution, CSRNet only contains less than 37k
trainable parameters, which is orders of magnitude smaller than existing
learning-based methods. Extensive experiments show that our method achieves
state-of-the-art performance on the benchmark MIT-Adobe FiveK dataset
quantitively and qualitatively. Code is available at
https://github.com/hejingwenhejingwen/CSRNet.
|
Optical fibres have transformed the way people interact with the world and
now permeate many areas of science. Optical fibres are traditionally thought of
as insensitive to magnetic fields, however many application areas from mining
to biomedicine would benefit from fibre-based remote magnetometry devices. In
this work, we realise such a device by embedding nanoscale magnetic sensors
into tellurite glass fibres. Remote magnetometry is performed on magnetically
active defect centres in nanodiamonds embedded into the glass matrix. Standard
optical magnetometry techniques are applied to initialize and detect local
magnetic field changes with a measured sensitivity of 26 micron Tesla/square
root(Hz). Our approach utilizes straight-forward optical excitation, simple
focusing elements, and low power components. We demonstrate remote magnetometry
by direct reporting of the magnetic ground states of nitrogen-vacancy defect
centres in the optical fibres. In addition, we present and describe
theoretically an all-optical technique that is ideally suited to remote
fibre-based sensing. The implications of our results broaden the applications
of optical fibres, which now have the potential to underpin a new generation of
medical magneto-endoscopes and remote mining sensors.
|
Observations of the dust and gas around embedded stellar clusters reveal some
of the processes involved in their formation and evolution. Large scale mass
infall with rates dM/dt=4e-4 solar masses/year is found to be disrupted on
small scales by protostellar outflows. Observations of the size and velocity
dispersion of clusters suggest that protostellar migration from their
birthplace begins at very early times and is a potentially useful evolutionary
indicator.
|
We reformulate the manifestly T-dual description of the massless sector of
the closed bosonic string, directly from the geometry associated with the (left
and right) affine Lie algebra of the coset space Poincare/Lorentz. This
construction initially doubles not only the (spacetime) coordinates for
translations but also those for Lorentz transformations (and their dual). As a
result, the Lorentz connection couples directly to the string (as does the
vielbein), rather than being introduced ad hoc to the covariant derivative as
previously. This not only reproduces the old definition of T-dual torsion, but
automatically gives a general, covariant definition of T-dual curvature (but
still with some undetermined connections).
|
Limited by fixed step-size and sparsity penalty factor, the conventional
sparsity-aware normalized subband adaptive filtering (NSAF) type algorithms
suffer from trade-off requirements of high filtering accurateness and quicker
convergence behavior for sparse system identification. To deal with this
problem, this paper proposes variable step-size L0-norm constraint NSAF
algorithms (VSS-L0-NSAFs). We first analyze mean-square-deviation (MSD)
statistics behavior of the L0-NSAF innovatively in according to novel weight
recursion form and arrive at corresponding expressions for the cases that
background noise variance is available and unavailable, where correlation
degree of system input is indicated by scaling parameter r. Based on
derivations, we develop an effective variable step-size scheme through
minimizing the upper bounds of the MSD under some reasonable assumptions and
lemma. Furthermore, an effective reset strategy is incorporated into presented
algorithms to tackle with non-stationary situations. Finally, numerical
simulations corroborate that the proposed algorithms achieve better performance
in terms of estimation accurateness and tracking capability in comparison with
existing related algorithms in sparse system identification and adaptive echo
cancellation circumstances.
|
We study the structure of the medium surrounding sites of high-mass star
formation to determine the interrelation between the HII regions and the
environment from which they were formed. The density distribution of the
surroundings is key in determining how the radiation of the newly formed stars
interacts with the surrounds in a way that allows it to be used as a star
formation tracer. We present new Herschel/SPIRE 250, 350 and 500 mum data of
LHA 120-N44 and LHA 120-N63 in the LMC. We construct average spectral energy
distributions (SEDs) for annuli centered on the IR bright part of the star
formation sites. The annuli cover ~10-~100 pc. We use a phenomenological dust
model to fit these SEDs to derive the dust column densities, characterise the
incident radiation field and the abundance of polycyclic aromatic hydrocarbon
molecules. We see a factor 5 decrease in the radiation field energy density as
a function of radial distance around N63. N44 does not show a systematic trend.
We construct a simple geometrical model to derive the 3-D density profile of
the surroundings of these two regions. Herschel/SPIRE data have proven very
efficient in deriving the dust mass distribution. We find that the radiation
field in the two sources behaves very differently. N63 is more or less
spherically symmetric and the average radiation field drops with distance. N44
shows no systematic decrease of the radiation intensity which is probably due
to the inhomogeneity of the surrounding molecular material and to the complex
distribution of several star forming clusters in the region.
|
We show that a sufficient condition for a subset $E$ in the Heisenberg group
(endowed with the Carnot-Carath\'{e}odory metric) to be contained in a
rectifiable curve is that it satisfies a modified analogue of Peter Jones's
geometric lemma. Our estimates improve on those of \cite{FFP}, by replacing the
power $2$ of the Jones-$\beta$-number with any power $r<4$. This complements
(in an open ended way) our work \cite{Li-Schul-beta-leq-length}, where we
showed that such an estimate was necessary, but with $r=4$.
|
We illustrate the detrimental effect, such as overconfident decisions, that
exponential behavior can have in methods like classical LDA and logistic
regression. We then show how polynomiality can remedy the situation. This,
among others, leads purposefully to random-level performance in the tails, away
from the bulk of the training data. A directly related, simple, yet important
technical novelty we subsequently present is softRmax: a reasoned alternative
to the standard softmax function employed in contemporary (deep) neural
networks. It is derived through linking the standard softmax to Gaussian
class-conditional models, as employed in LDA, and replacing those by a
polynomial alternative. We show that two aspects of softRmax, conservativeness
and inherent gradient regularization, lead to robustness against adversarial
attacks without gradient obfuscation.
|
A formalism for spin observables of the reaction
$pd\to ~^3He\eta$ is derived in a model independent way.
The general case with a full set of six independent spin amplitudes is
studied. Furthermore, approximations by five and four spin amplitudes are
investigated in the near threshold region. This region is of great interest to
search for a quasi-bound $^3He-\eta$ state, in particular, by measurement of
energy dependence of relative phases of s- and p-wave amplitudes. Complete
polarization experiments, allowing determination of spin amplitudes, are
analyzed. It is shown that measurement of only analyzing powers and spin
correlation coefficients hardly allows one to separate the s- and p-wave
amplitudes, but additional measurement of polarization transfer coefficients
simplifies this problem. Specific observables, given by products of one s- and
one p-wave amplitudes, are found. Measurement of these observables will provide
new independent information on the $^3He-\eta$ pole position.
|
While deep face recognition (FR) systems have shown amazing performance in
identification and verification, they also arouse privacy concerns for their
excessive surveillance on users, especially for public face images widely
spread on social networks. Recently, some studies adopt adversarial examples to
protect photos from being identified by unauthorized face recognition systems.
However, existing methods of generating adversarial face images suffer from
many limitations, such as awkward visual, white-box setting, weak
transferability, making them difficult to be applied to protect face privacy in
reality. In this paper, we propose adversarial makeup transfer GAN (AMT-GAN), a
novel face protection method aiming at constructing adversarial face images
that preserve stronger black-box transferability and better visual quality
simultaneously. AMT-GAN leverages generative adversarial networks (GAN) to
synthesize adversarial face images with makeup transferred from reference
images. In particular, we introduce a new regularization module along with a
joint training strategy to reconcile the conflicts between the adversarial
noises and the cycle consistence loss in makeup transfer, achieving a desirable
balance between the attack strength and visual changes. Extensive experiments
verify that compared with state of the arts, AMT-GAN can not only preserve a
comfortable visual quality, but also achieve a higher attack success rate over
commercial FR APIs, including Face++, Aliyun, and Microsoft.
|
The unfolding of detector effects in experimental data is critical for
enabling precision measurements in high-energy physics. However, traditional
unfolding methods face challenges in scalability, flexibility, and dependence
on simulations. We introduce a novel unfolding approach using conditional
Denoising Diffusion Probabilistic Models (cDDPM). Our method utilizes the cDDPM
for a non-iterative, flexible posterior sampling approach, which exhibits a
strong inductive bias that allows it to generalize to unseen physics processes
without explicitly assuming the underlying distribution. We test our approach
by training a single cDDPM to perform multidimensional particle-wise unfolding
for a variety of physics processes, including those not seen during training.
Our results highlight the potential of this method as a step towards a
"universal" unfolding tool that reduces dependence on truth-level assumptions.
|
Computing the trajectories of mandibular condyles directly from MRI could
provide a comprehensive examination, allowing for the extraction of both
anatomical and kinematic details. This study aimed to investigate the
feasibility of extracting 3D condylar trajectories from 2D real-time MRI and to
assess their precision.Twenty healthy subjects underwent real-time MRI while
opening and closing their jaws. One axial and two sagittal slices were
segmented using a U-Net-based algorithm. The centers of mass of the resulting
masks were projected onto the coordinate system based on anatomical markers and
temporally adjusted using a common projection. The quality of the computed
trajectories was evaluated using metrics designed to estimate movement
reproducibility, head motion, and slice placement symmetry.The segmentation of
the axial slices demonstrated good-to-excellent quality; however, the
segmentation of the sagittal slices required some fine-tuning. The movement
reproducibility was acceptable for most cases; nevertheless, head motion
displaced the trajectories by 1 mm on average. The difference in the
superior-inferior coordinate of the condyles in the closed jaw position was 1.7
mm on average.Despite limitations in precision, real-time MRI enables the
extraction of condylar trajectories with sufficient accuracy for evaluating
clinically relevant parameters such as condyle displacement, trajectories
aspect, and symmetry.
|
This paper addresses the detection of periodic transients in vibration
signals for detecting faults in rotating machines. For this purpose, we present
a method to estimate periodic-group-sparse signals in noise. The method is
based on the formulation of a convex optimization problem. A fast iterative
algorithm is given for its solution. A simulated signal is formulated to verify
the performance of the proposed approach for periodic feature extraction. The
detection performance of comparative methods is compared with that of the
proposed approach via RMSE values and receiver operating characteristic (ROC)
curves. Finally, the proposed approach is applied to compound faults diagnosis
of motor bearings. The non-stationary vibration data were acquired from a
SpectraQuest's machinery fault simulator. The processed results show the
proposed approach can effectively detect and extract the useful features of
bearing outer race and inner race defect.
|
In this paper we review some of our recent experimental and theoretical
results on transport and thermodynamic properties of heavy-fermion alloys
Ce(1-x)Yb(x)CoIn5. Charge transport measurements under magnetic field and
pressure on these single crystalline alloys revealed that: (i) relatively small
Yb substitution suppresses the field induced quantum critical point, with a
complete suppression for nominal Yb doping x>0.20; (ii) the superconducting
transition temperature Tc and Kondo lattice coherence temperature T* decrease
with x, yet they remain finite over the wide range of Yb concentrations; (iii)
both Tc and T* increase with pressure; (iv) there are two contributions to
resistivity, which show different temperature and pressure dependences,
implying that both heavy and light quasiparticles contribute to inelastic
scattering. We also analyzed theoretically the pressure dependence of both T*
and Tc within the composite pairing theory. In the purely static limit, when we
ignore the lattice dynamics, we find that the composite pairing mechanism
necessarily causes opposite behaviors of T* and Tc with pressure: if T* grows
with pressure, Tc must decrease with pressure and vice versa.
|
Real-time nonlinear Bayesian filtering algorithms are overwhelmed by data
volume, velocity and increasing complexity of computational models. In this
paper, we propose a novel ensemble based nonlinear Bayesian filtering approach
which only requires a small number of simulations and can be applied to
high-dimensional systems in the presence of intractable likelihood functions.
The proposed approach uses linear latent projections to estimate the joint
probability distribution between states, parameters, and observables using a
mixture of Gaussian components generated by the reconstruction error for each
ensemble member. Since it leverages the computational machinery behind linear
latent variable models, it can achieve fast implementations without the need to
compute high-dimensional sample covariance matrices. The performance of the
proposed approach is compared with the performance of ensemble Kalman filter on
a high-dimensional Lorenz nonlinear dynamical system.
|
Robotic systems are becoming pervasive and adopted in increasingly many
domains, such as manufacturing, healthcare, and space exploration. To this end,
engineering software has emerged as a crucial discipline for building
maintainable and reusable robotic systems. Robotics software engineering
research has received increasing attention, fostering autonomy as a fundamental
goal. However, robotics developers are still challenged trying to achieve this
goal given that simulation is not able to deliver solutions to realistically
emulate real-world phenomena. Robots also need to operate in unpredictable and
uncontrollable environments, which require safe and trustworthy self-adaptation
capabilities implemented in software. Typical techniques to address the
challenges are runtime verification, field-based testing, and mitigation
techniques that enable fail-safe solutions. However, there is no clear guidance
to architect ROS-based systems to enable and facilitate runtime verification
and field-based testing. This paper aims to fill in this gap by providing
guidelines that can help developers and QA teams when developing, verifying or
testing their robots in the field. These guidelines are carefully tailored to
address the challenges and requirements of testing robotics systems in
real-world scenarios. We conducted a literature review on studies addressing
runtime verification and field-based testing for robotic systems, mined
ROS-based application repositories, and validated the applicability, clarity,
and usefulness via two questionnaires with 55 answers. We contribute 20
guidelines formulated for researchers and practitioners in robotic software
engineering. Finally, we map our guidelines to open challenges thus far in
runtime verification and field-based testing for ROS-based systems and, we
outline promising research directions in the field.
|
The photon polarization operator in superstrong magnetic fields induces the
dynamical photon "mass" which leads to screening of Coulomb potential at small
distances $z\ll 1/m$, $m$ is the mass of an electron. We demonstrate that this
behaviour is qualitatively different from the case of D=2 QED, where the same
formula for a polarization operator leads to screening at large distances as
well. Because of screening the ground state energy of the hydrogen atom at the
magnetic fields $B \gg m^2/e^3$ has the finite value $E_0 = -me^4/2
\ln^2(1/e^6)$.
|
We report on the thermodynamic and transport properties of the rare-earth
Zintl compound Eu$_5$Sn$_2$As$_6$, which orders as a canted antiferromagnetic
magnetic semiconductor at 10.3~K. The system also displays a complex cascade of
magnetic phases arising from geometric and magnetic exchange frustration, with
high sensitivity to the application and direction of small magnetic fields. At
low temperature, Eu$_5$Sn$_2$As$_6$ exhibits negative colossal
magnetoresistance of up to a factor of $6\times10^3$. This may be a lower bound
as the conductivity appears to be shunted by an unknown conduction channel,
causing the resistivity to saturate. Mechanisms for the low temperature
saturation of resistivity are discussed.
|
For a simple digraph $G$ without directed triangles or digons, let $\beta(G)$
be the size of the smallest subset $X \subseteq E(G)$ such that $G\setminus X$
has no directed cycles, and let $\gamma(G)$ be the number of unordered pairs of
nonadjacent vertices in $G$. In 2008, Chudnovsky, Seymour, and Sullivan showed
that $\beta (G) \le \gamma(G)$, and conjectured that $\beta (G) \le
\gamma(G)/2$. Recently, Dunkum, Hamburger, and P\'or proved that $\beta (G) \le
0.88 \gamma(G)$. In this note, we prove that $\beta (G) \le 0.8616 \gamma(G)$.
|
The stability of nonvolatile thin liquid films and of sessile droplets is
strongly affected by finite size effects. We analyze their stability within the
framework of density functional theory using the sharp kink approximation,
i.e., on the basis of an effective interface Hamiltonian. We show that finite
size effects suppress spinodal dewetting of films because it is driven by a
long-wavelength instability. Therefore nonvolatile films are stable if the
substrate area is too small. Similarly, nonvolatile droplets connected to a
wetting film become unstable if the substrate area is too large. This
instability of a nonvolatile sessile droplet turns out to be equivalent to the
instability of a volatile drop which can attain chemical equilibrium with its
vapor.
|
We report radio observations, made with the Australia Telescope Compact
Array, of the X-ray transient XTE J1701-462. This system has been classified as
a new `Z' source, displaying characteristic patterns of behaviour probably
associated with accretion onto a low magnetic field neutron star at close to
the Eddington limit. The radio counterpart is highly variable, and was detected
in six of sixteen observations over the period 2006 January -- April. The
coupling of radio emission to X-ray state, despite limited sampling, appears to
be similar to that of other `Z' sources, in that there is no radio emission on
the flaring branch. The mean radio and X-ray luminosities are consistent with
the other Z sources for a distance of 5--15 kpc. The radio spectrum is
unusually flat, or even inverted, in contrast to the related sources, Sco X-1
and Cir X-1, which usually display an optically thin radio spectrum. Deep
wide-field observations indicate an extended structure three arcminutes to the
south which is aligned with the X-ray binary. This seems to represent a
significant overdensity of radio sources for the field and so, although a
background source remains a strong possibility, we consider it plausible that
this is a large-scale jet associated with XTE J1701-462.
|
By elaborating on the recent progress made in the area of Feynman integrals,
we apply the intersection theory for twisted de Rham cohomologies to simple
integrals involving orthogonal polynomials, matrix elements of operators in
Quantum Mechanics and Green's functions in Field Theory, showing that the
algebraic identities they obey are related to the decomposition of twisted
cocycles within cohomology groups, and which, therefore, can be derived by
means of intersection numbers. Our investigation suggests an algebraic approach
generically applicable to the study of higher-order moments of probability
distributions, where the dimension of the cohomology groups corresponds to the
number of independent moments; the intersection numbers for twisted cocycles
can be used to derive linear and quadratic relations among them. Our study
offers additional evidence of the intertwinement between physics, geometry, and
statistics.
|
We propose a remarkably simple electronic refrigerator based on the Coulomb
barrier for single-electron tunneling. A fully normal single-electron
transistor is voltage $V$ biased at a gate position such that tunneling through
one of the junctions costs an energy of about $k_BT \ll eV, E_C$, where $T$ is
the temperature and $E_C$ is the transistor charging energy. The tunneling in
the junction with positive energy cost cools both the electrodes attached to
it. Immediate practical realizations of such a refrigerator make use of Andreev
mirrors which suppress heat current while maintaining full electric contact.
|
Non-universal scale transformations of the physical fields are extended to
pure quantum fluids and used to calculate susceptibility, specific heat and the
order parameter along the critical isochore of He3 near its liquid-vapor
critical point. Within the so-called preasymptotic domain, where the Wegner
expansion restricted to the first term of confluent corrections to scaling is
expected valid, the results show agreement with the experimental measurements
and recent predictions, either based on the minimal-substraction
renormalization and the massive renormalization schemes within the
$\Phi\_{d=3}^{4}(n=1)$-model, or based on the crossover parametric equation of
state for Ising-like systems.
|
A digital quantum simulation for the extended Agassi model is proposed using
a quantum platform with eight trapped ions. The extended Agassi model is an
analytically solvable model including both short range pairing and long range
monopole-monopole interactions with applications in nuclear physics and in
other many-body systems. In addition, it owns a rich phase diagram with
different phases and the corresponding phase transition surfaces. The aim of
this work is twofold: on one hand, to propose a quantum simulation of the model
at the present limits of the trapped ions facilities and, on the other hand, to
show how to use a machine learning algorithm on top of the quantum simulation
to accurately determine the phase of the system. Concerning the quantum
simulation, this proposal is scalable with polynomial resources to larger
Agassi systems. Digital quantum simulations of nuclear physics models assisted
by machine learning may enable one to outperform the fastest classical
computers in determining fundamental aspects of nuclear matter.
|
We study the behavior of the Yang-Mills flow for unitary connections on
compact and non-compact oriented surfaces with varying metrics. The flow can be
used to define a one dimensional foliation on the space of SU(2)
representations of a once punctured surface. This foliation universalizes over
Teichm\"uller space and is equivariant with respect to the action of the
mapping class group. It is shown how to extend the foliation as a singular
foliation over the Strebel boundary of Teichm\"uller space, and continuity of
this extension is the main result of the paper.
|
In this paper, we study a few versions of the uncertainty principle for the
short-time Fourier transform on the lattice $\mathbb Z^n \times \mathbb T^n$.
In particular, we establish the uncertainty principle for orthonormal
sequences, Donoho--Stark's uncertainty principle, Benedicks-type uncertainty
principle, Heisenberg-type uncertainty principle and local uncertainty
inequality for this transform on $\mathbb Z^n \times \mathbb T^n$. Also, we
obtain the Heisenberg-type uncertainty inequality using the $k$-entropy of the
short-time Fourier transform on $\mathbb Z^n \times \mathbb T^n$.
|
Assuming only Sobolev regularity of the domain, we prove time-analyticity of
Lagrangian trajectories for solutions of the Euler equations.
|
Developments in IoT applications are playing an important role in our
day-to-day life, starting from business predictions to self driving cars. One
of the area, most influenced by the field of AI and IoT is retail analytics. In
Retail Analytics, Conversion Rates - a metric which is most often used by
retail stores to measure how many people have visited the store and how many
purchases has happened. This retail conversion rate assess the marketing
operations, increasing stock, store outlet and running promotions ..etc. Our
project intends to build a cost-effective people counting system with AI at
Edge, where it calculates Conversion rates using total number of people counted
by the system and number of transactions for the day, which helps in providing
analytical insights for retail store optimization with a very minimum hardware
requirements.
|
Deep neural networks (DNNs) have shown great success in many machine learning
tasks. Their training is challenging since the loss surface of the network
architecture is generally non-convex, or even non-smooth. How and under what
assumptions is guaranteed convergence to a \textit{global} minimum possible? We
propose a reformulation of the minimization problem allowing for a new
recursive algorithmic framework. By using bounded style assumptions, we prove
convergence to an $\varepsilon$-(global) minimum using
$\mathcal{\tilde{O}}(1/\varepsilon^3)$ gradient computations. Our theoretical
foundation motivates further study, implementation, and optimization of the new
algorithmic framework and further investigation of its non-standard bounded
style assumptions. This new direction broadens our understanding of why and
under what circumstances training of a DNN converges to a global minimum.
|
Investigating star formation requires precise knowledge of the properties of
the dense molecular gas. The low metallicity and wide range of star formation
activity of the Large and Small Magellanic Clouds make them prime laboratories
to study how local physical conditions impact the dense gas reservoirs. The aim
of the Dense Gas Survey for the Magellanic Clouds (DeGaS-MC) project is to
expand our knowledge of the relation between dense gas properties and star
formation activity by targeting the LMC and SMC observed in the HCO+(2-1) and
HCN(2-1) transitions. We carried out a pointing survey toward 30 LMC and SMC
molecular clouds using the SEPIA180 instrument installed on the APEX telescope
and a follow-up mapping campaign in 13 star-forming regions. This first paper
provides line characteristic catalogs and integrated line-intensity maps of the
sources. HCO+(2-1) is detected in 20 and HCN(2-1) in 8 of the 29 pointings
observed. The dense gas velocity pattern follows the line-of-sight velocity
field derived from the stellar population. The HCN emission is less extended
than the HCO+ emission. The HCO+(2-1)/HCN(2-1) brightness temperature ratios
range from 1 to 7, which is consistent with the large ratios commonly observed
in low-metallicity environments. A larger number of young stellar objects are
found at high HCO+ intensities and lower HCO+/HCN flux ratios, and thus toward
denser lines of sight. The dense gas luminosities correlate with the star
formation rate traced by the total infrared luminosity over the two orders of
magnitude covered by our observations, although substantial region-to-region
variations are observed.
|
The John-Nirenberg spaces $JN_p$ are generalizations of the space of bounded
mean oscillation $BMO$ with $JN_{\infty}=BMO$. Their vanishing subspaces
$VJN_p$ and $CJN_p$ are defined in similar ways as $VMO$ and $CMO$, which are
subspaces of $BMO$. As our main result, we prove that $VJN_p$ and $CJN_p$
coincide by showing that certain Morrey type integrals of $JN_p$ functions tend
to zero for small and large cubes. We also show that $JN_{p,q}(\mathbb{R}^n) =
L^p(\mathbb{R}^n) / \mathbb{R}$, if $p = q$.
|
Spherical quartz stones of around 1 cm in diameter have been exposed to
anodic arc discharges in a helium atmosphere at 300 Torr. The arc current
flowing between the graphite electrodes was set either in continuous DC mode
(30-150 A) or in pulsed mode at 2 Hz (220 A peak). The ablation rate in each
sample was systematically measured after several seconds of arc plasma
treatment. Optical emission spectroscopy (OES) diagnostics and 2-D fluid
simulations of the arc discharge have shed light on the heat flux transport and
the heating mechanisms of the quartz crystals. A linear correlation is found
between the absorbed power density and the resulting rate of penetration, which
yields a maximal value of 15 cm/h for approximately 150 W/cm2. The linear fit
on the slope provides a specific energy of 40 kJ/cm3. The incident energy flux
onto the sample surface promoted a phase transition from crystalline to glassy
silica, as characterized via Raman spectroscopy. This study points out the
strong potential of arc plasma technology for geothermal drilling applications.
|
Existing techniques for Craig interpolation for the quantifier-free fragment
of the theory of arrays are inefficient for computing sequence and tree
interpolants: the solver needs to run for every partitioning $(A, B)$ of the
interpolation problem to avoid creating $AB$-mixed terms. We present a new
approach using Proof Tree Preserving Interpolation and an array solver based on
Weak Equivalence on Arrays. We give an interpolation algorithm for the lemmas
produced by the array solver. The computed interpolants have worst-case
exponential size for extensionality lemmas and worst-case quadratic size
otherwise. We show that these bounds are strict in the sense that there are
lemmas with no smaller interpolants. We implemented the algorithm and show that
the produced interpolants are useful to prove memory safety for C programs.
|
Dilepton production in intermediate energy nucleus-nucleus collisions as well
as in elementary proton-proton reactions is analysed within the UrQMD transport
model. For C+C collisions at 1 AGeV and 2 AGeV the resulting invariant mass
spectra are compared to recent HADES data. We find that the experimental
spectrum for C+C at 2 AGeV is slightly overestimated by the theoretical
calculations in the region around the vector meson peak, but fairly described
in the low mass region, where the data is satisfactorily saturated by the
Dalitz decay of the $\eta$ meson and the $\Delta$ resonance. At 1 AGeV an
underestimation of the experimental data is found, pointing that at lower
energies the low mass region is not fully saturated by standardly parametrized
$\Delta$ Dalitz decays alone. Furthermore, predictions for dilepton spectra for
$pp$ reactions at 1.25 GeV, 2.2 GeV and 3.5 GeV and Ar+KCl reactions at 1.75
AGeV are presented. The study is complemented by a detailed investigation of
the role of absorption of the parent particles on the corresponding dilepton
yields in the regime which has so far been probed by HADES.
|
Survival prediction is a major concern for cancer management. Deep survival
models based on deep learning have been widely adopted to perform end-to-end
survival prediction from medical images. Recent deep survival models achieved
promising performance by jointly performing tumor segmentation with survival
prediction, where the models were guided to extract tumor-related information
through Multi-Task Learning (MTL). However, these deep survival models have
difficulties in exploring out-of-tumor prognostic information. In addition,
existing deep survival models are unable to effectively leverage multi-modality
images. Empirically-designed fusion strategies were commonly adopted to fuse
multi-modality information via task-specific manually-designed networks, thus
limiting the adaptability to different scenarios. In this study, we propose an
Adaptive Multi-modality Segmentation-to-Survival model (AdaMSS) for survival
prediction from PET/CT images. Instead of adopting MTL, we propose a novel
Segmentation-to-Survival Learning (SSL) strategy, where our AdaMSS is trained
for tumor segmentation and survival prediction sequentially in two stages. This
strategy enables the AdaMSS to focus on tumor regions in the first stage and
gradually expand its focus to include other prognosis-related regions in the
second stage. We also propose a data-driven strategy to fuse multi-modality
information, which realizes adaptive optimization of fusion strategies based on
training data during training. With the SSL and data-driven fusion strategies,
our AdaMSS is designed as an adaptive model that can self-adapt its focus
regions and fusion strategy for different training stages. Extensive
experiments with two large clinical datasets show that our AdaMSS outperforms
state-of-the-art survival prediction methods.
|
Super-entropic black holes possess finite-area but noncompact event horizons
and violate the reverse isoperimetric inequality. It has been conjectured that
such black holes always have negative specific heat at constant volume $C_{V}$
or negative specific heat at constant pressure $C_{P}$ whenever $C_{V}>0$,
making them unstable in extended thermodynamics. In this paper, we test this
instability conjecture on a family of nonlinear electrodynamics black holes,
namely $3$D Einstein-Born-Infeld (EBI) AdS black holes. Our results show that
when nonlinear electrodynamics effects are weak, the instability conjecture is
valid. However, the conjecture can be violated in some parameter region when
nonlinear electrodynamics effects are strong enough. This observation thus
provides a counter example to the instability conjecture, which suggests that
super-entropic black holes could be thermodynamically stable.
|
A Richardson variety $X_\ga^\gc$ in the Orthogonal Grassmannian is defined to
be the intersection of a Schubert variety $X^\gc$ in the Orthogonal
Grassmannian and a opposite Schubert variety $X_\ga$ therein. We give an
explicit description of the initial ideal (with respect to certain conveniently
chosen term order) for the ideal of the tangent cone at any $T$-fixed point of
$X_\ga^\gc$, thus generalizing a result of Raghavan-Upadhyay \cite{Ra-Up2}. Our
proof is based on a generalization of the Robinson-Schensted-Knuth (RSK)
correspondence, which we call the Orthogonal bounded RSK (OBRSK). The OBRSK
correspondence will give a degree-preserving bijection between a set of
monomials defined by the initial ideal of the ideal of the tangent cone (as
mentioned above) and a `standard monomial basis'. A similar work for Richardson
varieties in the ordinary Grassmannian was done by Kreiman in \cite{Kr-bkrs}.
|
We study D2-branes on the K3-fibration P^4_(11222)[8] using matrix
factorizations at the Landau-Ginzburg point and analyze their moduli space and
superpotentials in detail. We find that the open string moduli space consists
of various intersecting branches of different dimensions. Families of D2-branes
wrapping rational curves of degree one intersect with bound state branches. The
influence of non-toric complex structure deformations is investigated in the
Landau-Ginzburg framework, where these deformations arise as bulk moduli from
the twisted sectors.
|
We study the empirical realisation of the memory effect in Yang-Mills theory,
especially in view of the classical vs. quantum nature of the theory. Gauge
invariant analysis of memory in classical U(1) electrodynamics and its
observation by total change of transverse momentum of a charge is reviewed.
Gauge fixing leads to a determination of a gauge transformation at infinity. An
example of Yang-Mills memory then is obtained by reinterpreting known results
on interactions of a quark and a large high energy nucleus in the theory of
Color Glass Condensate. The memory signal is again a kick in transverse
momentum, but it is only obtained in quantum theory after fixing the gauge,
after summing over an ensemble of classical processes.
|
Numerous groups have applied a variety of deep learning techniques to
computer vision problems in highway perception scenarios. In this paper, we
presented a number of empirical evaluations of recent deep learning advances.
Computer vision, combined with deep learning, has the potential to bring about
a relatively inexpensive, robust solution to autonomous driving. To prepare
deep learning for industry uptake and practical applications, neural networks
will require large data sets that represent all possible driving environments
and scenarios. We collect a large data set of highway data and apply deep
learning and computer vision algorithms to problems such as car and lane
detection. We show how existing convolutional neural networks (CNNs) can be
used to perform lane and vehicle detection while running at frame rates
required for a real-time system. Our results lend credence to the hypothesis
that deep learning holds promise for autonomous driving.
|
We prove that, for sufficiently large $n$, every graph of order $n$ with
minimum degree at least $0.852n$ has a fractional edge-decomposition into
triangles. We do this by refining a method used by Dross to establish a bound
of $0.9n$. By a result of Barber, K\"{u}hn, Lo and Osthus, our result implies
that, for each $\epsilon >0$, every graph of sufficiently large order $n$ with
minimum degree at least $(0.852+\epsilon)n$ has a triangle decomposition if and
only if it has all even degrees and number of edges a multiple of three.
|
Quasi-experimental methods have proliferated over the last two decades, as
researchers develop causal inference tools for settings in which randomization
is infeasible. Two popular such methods, difference-in-differences (DID) and
comparative interrupted time series (CITS), compare observations before and
after an intervention in a treated group to an untreated comparison group
observed over the same period. Both methods rely on strong, untestable
counterfactual assumptions. Despite their similarities, the methodological
literature on CITS lacks the mathematical formality of DID. In this paper, we
use the potential outcomes framework to formalize two versions of CITS - a
general version described by Bloom (2005) and a linear version often used in
health services research. We then compare these to two corresponding DID
formulations - one with time fixed effects and one with time fixed effects and
group trends. We also re-analyze three previously published studies using these
methods. We demonstrate that the most general versions of CITS and DID impute
the same counterfactuals and estimate the same treatment effects. The only
difference between these two designs is the language used to describe them and
their popularity in distinct disciplines. We also show that these designs
diverge when one constrains them using linearity (CITS) or parallel trends
(DID). We recommend defaulting to the more flexible versions and provide advice
to practitioners on choosing between the more constrained versions by
considering the data-generating mechanism. We also recommend greater attention
to specifying the outcome model and counterfactuals in papers, allowing for
transparent evaluation of the plausibility of causal assumptions.
|
We describe an electrodynamic mechanism for coherent, quantum mechanical
coupling between spacially separated quantum dots on a microchip. The technique
is based on capacitive interactions between the electron charge and a
superconducting transmission line resonator, and is closely related to atomic
cavity quantum electrodynamics. We investigate several potential applications
of this technique which have varying degrees of complexity. In particular, we
demonstrate that this mechanism allows design and investigation of an on-chip
double-dot microscopic maser. Moreover, the interaction may be extended to
couple spatially separated electron spin states while only virtually populating
fast-decaying superpositions of charge states. This represents an effective,
controllable long-range interaction, which may facilitate implementation of
quantum information processing with electron spin qubits and potentially allow
coupling to other quantum systems such as atomic or superconducting qubits.
|
The Born rule postulates that the probability of measurement in quantum
mechanics is related to the squared modulus of the wave function $\psi$. We
rearrange the equation for energy eigenfunctions to define the energy as the
real part of $\hat{E}\psi/\psi$. For an eigenstate, this definition gives a
constant energy eigenvalue. For a general wave function, the energy fluctuates
in space and time. We consider a particle in a one-dimensional square well
potential in a superposition of two states and average the energy over space
and time. We show that, for most cases, such an energy expectation value
differs by only a few percent from that calculated using the Born rule. This
difference is consistent with experimental tests of the expectation value and
suggests that the Born rule may be an approximation of spacetime averaging.
|
Some recent methods for lossy signal and image compression store only a few
selected pixels and fill in the missing structures by inpainting with a partial
differential equation (PDE). Suitable operators include the Laplacian, the
biharmonic operator, and edge-enhancing anisotropic diffusion (EED). The
quality of such approaches depends substantially on the selection of the data
that is kept. Optimising this data in the domain and codomain gives rise to
challenging mathematical problems that shall be addressed in our work.
In the 1D case, we prove results that provide insights into the difficulty of
this problem, and we give evidence that a splitting into spatial and tonal
(i.e. function value) optimisation does hardly deteriorate the results. In the
2D setting, we present generic algorithms that achieve a high reconstruction
quality even if the specified data is very sparse. To optimise the spatial
data, we use a probabilistic sparsification, followed by a nonlocal pixel
exchange that avoids getting trapped in bad local optima. After this spatial
optimisation we perform a tonal optimisation that modifies the function values
in order to reduce the global reconstruction error. For homogeneous diffusion
inpainting, this comes down to a least squares problem for which we prove that
it has a unique solution. We demonstrate that it can be found efficiently with
a gradient descent approach that is accelerated with fast explicit diffusion
(FED) cycles. Our framework allows to specify the desired density of the
inpainting mask a priori. Moreover, is more generic than other data
optimisation approaches for the sparse inpainting problem, since it can also be
extended to nonlinear inpainting operators such as EED. This is exploited to
achieve reconstructions with state-of-the-art quality.
We also give an extensive literature survey on PDE-based image compression
methods.
|
In this paper we address the relationship between zero temperature Glauber
dynamics and the diffusion-annihilation problem in the free fermion case. We
show that the well-known duality transformation between the two problems can be
formulated as a similarity transformation if one uses appropriate (toroidal)
boundary conditions. This allow us to establish and clarify the precise nature
of the relationship between the two models. In this way we obtain a one-to-one
correspondence between observables and initial states in the two problems. A
random initial state in Glauber dynamics is related to a short range correlated
state in the annihilation problem. In particular the long-time behaviour of the
density in this state is seen to depend on the initial conditions. Hence, we
show that the presence of correlations in the initial state determine the
dependence of the long time behaviour of the density on the initial conditions,
even if such correlations are short-ranged. We also apply a field-theoretical
method to the calculation of multi-time correlation functions in this initial
state.
|
We examine the possibility that very massive stars greatly exceeding the
commonly adopted stellar mass limit of 150 Msun may be present in young star
clusters in the local universe. We identify ten candidate clusters, some of
which may host stars with masses up to 600 Msun formed via runaway collisions.
We estimate the probabilities of these very massive stars being in eclipsing
binaries to be >30%. Although most of these systems cannot be resolved at
present, their transits can be detected at distances of 3 Mpc even under the
contamination of the background cluster light, due to the large associated
luminosities ~10^7 Lsun and mean transit depths of ~10^6 Lsun. Discovery of
very massive eclipsing binaries would flag possible progenitors of
pair-instability supernovae and intermediate-mass black holes.
|
Recent Spitzer Space Telescope observations of several astrophysical
environments such as Planetary Nebulae, Reflection Nebulae, and R Coronae
Borealis stars show the simultaneous presence of mid-infrared features
attributed to neutral fullerene molecules (i.e., C60) and polycyclic aromatic
hydrocarbons (PAHs). If C60 fullerenes and PAHs coexist in fullerene-rich space
environments, then C60 may easily form adducts with a number of different PAH
molecules; at least with catacondensed PAHs. Here we present the laboratory
infrared spectra (~2-25 um) of C60 fullerene and anthracene Dies-Alder mono-
and bis-adducts as produced by sonochemical synthesis. We find that
C60/anthracene Diels-Alder adducts display spectral features strikingly similar
to those from C60 (and C70) fullerenes and other unidentified infrared emission
features. Thus, fullerene-adducts - if formed under astrophysical conditions
and stable/abundant enough - may contribute to the infrared emission features
observed in fullerene-containing circumstellar/interstellar environments.
|
We discuss mock automorphic forms from the point of view of representation
theory, that is, obtained from weak harmonic Maass forms give rise to
nontrivial $(\mathfrak{g},K)$-cohomology. We consider the possibility of
replacing the `holomorphic' condition with `cohomological' when generalizing to
general reductive groups. Such a candidate allows for growing Fourier
coefficients, in contrast to automorphic forms under the Miatello-Wallach
conjecture. In the second part, we provide an overview of the connection with
BPS black hole counts as a physical motivation for studying mock automorphic
forms.
|
An improved weighting algorithm applied to hadron showers has been developed
for a fine grained LAr calorimeter. The new method uses tabulated weights which
depend on the density of energy deposited in individual cells and in a
surrounding cone whose symmetry axis connects the interaction vertex with the
highest energy cluster in the shower induced by a hadron. The weighting of the
visible energy and the correction for losses due to noise cuts are applied in
separate steps. In contrast to standard weighting procedures the new algorithm
allows to reconstruct the total energy as well as the spatial energy deposition
on the level of individual calorimeter cells. The linearity and the energy
resolution of the pion signal in the momentum interval 2 GeV/c <= p <= 20 GeV/c
studied in this analysis are considerably improved in comparison to the
standard weighting algorithm as practiced presently by the H1 collaboration.
Moreover the energy spectra reconstructed with the new method follow in a broad
interval a Gaussian distribution and have less pronounced tails.
|
We propose a new approximation scheme to obtain analytic expressions for the
bound state energies and eigenfunctions of Yukawa like potentials. The
predicted energies are in excellent agreement with the accurate numerical
values reported in the literature.
|
This paper describes low-cost techniques used to collect video data in two
different tutorial classrooms - one in which the recording equipment is
permanently installed and one in which it is temporary. The author explains
what to do before, during, and after class in these two situations, providing
general strategies and technical advice for researchers interested in
videotaping tutorials or similar classrooms.
|
We study operator dynamics in many-body quantum systems, focusing on generic
features of systems which are ergodic, spatially extended, and lack conserved
densities, as exemplified by spin chains with Floquet time evolution. To
characterise dynamics we examine, in solvable models and numerically, the
behaviour of operator autocorrelation functions, as a function of time and the
size of the operator support. The standard expectation is that operator
autocorrelation functions in such systems are maximum at time zero and decay,
over a few Floquet periods, to a fluctuating value that reduces to zero under
an average over an ensemble of statistically similar systems. Our central
result is that ensemble-averaged correlation functions also display a second
generic feature, which consists of a peak at a later time. In individual
many-body systems, this peak can also be revealed by averaging autocorrelation
functions over complete sets of operators supported within a finite spatial
region, thereby generating a partial spectral form factor. The duration of the
peak grows indefinitely with the size of the operator support, and its
amplitude shrinks, but both are essentially independent of system size provided
this is sufficiently large to contain the operator. In finite systems, the
averaged correlation functions also show a further feature at still later
times, which is a counterpart to the so-called ramp and plateau of the spectral
form factor; its amplitude in the autocorrelation function decreases to zero
with increasing system size. Both the later-time peak and the ramp-and-plateau
feature are specific to models with time-translation symmetry, such as Floquet
systems or models with a time-independent Hamiltonian, and are absent in models
with an evolution operator that is a random function of time, such as the
extensively-studied random unitary circuits.
|
A scalar-response functional model describes the association between a scalar
response and a set of functional covariates. An important problem in the
functional data literature is to test the nullity or linearity of the effect of
the functional covariate in the context of scalar-on-function regression. This
article provides an overview of the existing methods for testing both the null
hypotheses that there is no relationship and that there is a linear
relationship between the functional covariate and scalar response, and a
comprehensive numerical comparison of their performance. The methods are
compared for a variety of realistic scenarios: when the functional covariate is
observed at dense or sparse grids and measurements include noise or not.
Finally, the methods are illustrated on the Tecator data set.
|
For a continuous field of the Cuntz algebra over a finite CW complex, we
introduce a topological invariant, which is an element in Dadarlat-Pennig's
generalized cohomology group, and prove that the invariant is trivial if and
only if the field comes from a vector bundle via Pimsner's construction.
|
In this correspondence, a novel simultaneous transmitting and reflecting
(STAR) reconfigurable intelligent surfaces (RISs) design is proposed in a
non-orthogonal multiple access (NOMA) enhanced coordinated multi-point
transmission (CoMP) network. Based on the insights of signal-enhancement-based
(SEB) and signal-cancellation-based (SCB) designs, we propose a novel
simultaneous-signal-enhancement-and-cancellation-based (SSECB) design, where
the inter-cell interferences and desired signals can be simultaneously
eliminated and boosted. Our simulation results demonstrate that: i) the
inter-cell interference can be perfectly eliminated, and the desired signals
can be enhanced simultaneously with the aid of a large number of RIS elements;
ii) the proposed SSECB design is capable of outperforming the conventional SEB
and SCB designs.
|
The boundary element method (BEM) enables the efficient electromagnetic
modelling of lossy conductors with a surface-based discretization. Existing BEM
techniques for conductor modelling require either expensive dual basis
functions or the use of both single- and double-layer potential operators to
obtain a well-conditioned system matrix. The associated computational cost is
particularly significant when conductors are embedded in stratified media, and
the expensive multilayer Green's function (MGF) must be invoked. In this work,
a novel single-source BEM formulation is proposed, which leads to a
well-conditioned system matrix without the need for dual basis functions. The
proposed single-layer impedance matrix (SLIM) formulation does not require the
double-layer potential to model the background medium, which reduces the cost
associated with the MGF. The accuracy and efficiency of the proposed method is
demonstrated through realistic examples drawn from different applications.
|
In crowd labeling, a large amount of unlabeled data instances are outsourced
to a crowd of workers. Workers will be paid for each label they provide, but
the labeling requester usually has only a limited amount of the budget. Since
data instances have different levels of labeling difficulty and workers have
different reliability, it is desirable to have an optimal policy to allocate
the budget among all instance-worker pairs such that the overall labeling
accuracy is maximized. We consider categorical labeling tasks and formulate the
budget allocation problem as a Bayesian Markov decision process (MDP), which
simultaneously conducts learning and decision making. Using the dynamic
programming (DP) recurrence, one can obtain the optimal allocation policy.
However, DP quickly becomes computationally intractable when the size of the
problem increases. To solve this challenge, we propose a computationally
efficient approximate policy, called optimistic knowledge gradient policy. Our
MDP is a quite general framework, which applies to both pull crowdsourcing
marketplaces with homogeneous workers and push marketplaces with heterogeneous
workers. It can also incorporate the contextual information of instances when
they are available. The experiments on both simulated and real data show that
the proposed policy achieves a higher labeling accuracy than other existing
policies at the same budget level.
|
The dynamics of N/Z and mass equilibration are investigated in the reactions
112,124Sn + 239Pu by employing the isospin-dependent quantum molecular dynamics
model. It is found that N/Z and mass equilibration take place at different
collision stages. The N/Z relaxation is observed in the approaching phase (from
first contact to deepest contact) with a very short time, whereas interestingly
we find for the first time that mass equilibration only takes place in the
separation phase (from the deepest contact to re-separation), which are
explained by investigating the dynamical asymmetry between the approaching and
separation phases. The mass equilibration also could be clarified with a
dynamical potential energy surface. Our results provide a new insight into the
equilibration dynamics of the quantum systems.
|
In this paper, we propose two iterative methods for finding a common solution
of a finite family of equilibrium problems for pseudomonotone bifunctions. The
first is a parallel hybrid extragradient-cutting algorithm which is extended
from the previously known one for variational inequalities to equilibrium
problems. The second is a new cyclic hybrid extragradient-cutting algorithm. In
the cyclic algorithm, using the known techniques, we can perform and develop
practical numerical experiments.
|
We study the $\mathfrak{gl}_{1|1}$ supersymmetric XXX spin chains. We give an
explicit description of the algebra of Hamiltonians acting on any cyclic tensor
products of polynomial evaluation $\mathfrak{gl}_{1|1}$ Yangian modules. It
follows that there exists a bijection between common eigenvectors (up to
proportionality) of the algebra of Hamiltonians and monic divisors of an
explicit polynomial written in terms of the Drinfeld polynomials. In particular
our result implies that each common eigenspace of the algebra of Hamiltonians
has dimension one. We also give dimensions of the generalized eigenspaces. We
show that when the tensor product is irreducible, then all eigenvectors can be
constructed using Bethe ansatz. We express the transfer matrices associated to
symmetrizers and anti-symmetrizers of vector representations in terms of the
first transfer matrix and the center of the Yangian.
|
Common techniques in Gravitational Wave data analysis assume, to some extent,
the stationarity and Gaussianity of the detector noise. These assumptions are
not always satisfied because of the presence of short-duration transients,
namely glitches, and other slower variations in the statistical properties of
the noise, which might be related to malfunctioning subsystems. We present here
a new technique to test the stationarity hypothesis with minimal assumptions on
the data, exploiting the band-limited root mean square and the two-sample
Kolmogorov-Smirnov test. The outcome is a time-frequency map showing where the
hypothesis is to be rejected. This technique was used as part of the event
validation procedure for assessing the quality of the LIGO and Virgo data
during O3. We also report on the applications of the test to both simulated and
real data, highlighting its sensitivity to various kinds of non-stationarities.
|
We present stellar mass measurements for a sample of 36 Luminous Compact Blue
Galaxies (LCBGs) at redshifts z = 0.4-1.2 in the Flanking Fields around the
Hubble Deep Field North. The technique is based on fitting a two-component
galaxy population model to multi-broadband photometry. Best-fit models are
found to be largely independent on the assumed values for the IMF and the
metallicity of the stellar populations, but are sensitive to the amount of
extinction and the extinction law adopted. On average, the best-fit model
corresponds to a LMC extinction law with E(B-V)=0.5. Stellar mass estimates,
however, are remarkably independent on the final model choice. Using a Salpeter
IMF, the derived median stellar mass for this sample is 5 x 10^9 Mo, i.e., ~2
times smaller than previous virial mass estimates. Despite uncertainties of a
factor 2-3, our results strengthen prior claims that L* CBGs at intermediate
redshifts are, on average, about 10 times less massive than a typical L* galaxy
today.
|
Subsets and Splits