text
stringlengths
6
128k
This paper is a thorough study of a digital broadcasting system adapted to the small mountainous island of Mauritius. A digital LAN was designed with MPEG-2 signals. The compressed signals were transmitted using DVB-T and QAM modulators. QAM-16 and QAM-64 modulators were designed and tested with a simulator under critical conditions of AWGN and phase noises. Results obtained from simulation have shown that Digital video broadcast with a single frequency network (SFN) is possible in Mauritius with QAM-64 and QAM-16 modulators applying COFDM mode of transmission. However, this study has also shown that QAM-16 modulator had a better performance at low AWGN values (less than 12 dB) and can be adopted for Mauritius Island, provided that the number of transmitted channels is not high enough.
We have shown that the $B-L$ generation due to the decay of the thermally produced superheavy fields can explain the Baryon assymmetry in the universe if the superheavy fields are heavier than $10^{13-14}$ GeV. Note that although the superheavy fields have non-vanishing charges under the standard model gauge interactions, the thermally prduced baryon asymmetry is sizable. The $B-L$ violating effective operators induced by integrating the superheavy fields have dimension 7, while the operator in the famous leptogenesis has dimension 5. Therefore, the constraints from the nucleon stability can be easily satisfied.
Environmental contours are widely used as basis for design of structures exposed to environmental loads. The basic idea of the method is to decouple the environmental description from the structural response. This is done by establishing an envelope of environmental conditions, such that any structure tolerating loads on this envelope will have a failure probability smaller than a prescribed value. Specifically, given an $n$-dimensional random variable $\mathbf{X}$ and a target probability of failure $p_{e}$, an environmental contour is the boundary of a set $\mathcal{B} \subset \mathbb{R}^{n}$ with the following property: For any failure set $\mathcal{F} \subset \mathbb{R}^{n}$, if $\mathcal{F}$ does not intersect the interior of $\mathcal{B}$, then the probability of failure, $P(\mathbf{X} \in \mathcal{F})$, is bounded above by $p_{e}$. As is common for many real-world applications, we work under the assumption that failure sets are convex. In this paper, we show that such environmental contours may be regarded as boundaries of Voronoi cells. This geometric interpretation leads to new theoretical insights and suggests a simple novel construction algorithm that guarantees the desired probabilistic properties. The method is illustrated with examples in two and three dimensions, but the results extend to environmental contours in arbitrary dimensions. Inspired by the Voronoi-Delaunay duality in the numerical discrete scenario, we are also able to derive an analytical representation where the environmental contour is considered as a differentiable manifold, and a criterion for its existence is established.
Recent observations of the cosmic microwave background (CMB) indicate that a successful theory of cosmological inflation needs to have flat potential of the inflaton scalar field. Realizing the inflaton to be a pseudo-Nambu Goldstone boson (pNGB) could ensure the flatness and the sub-Planckian scales related to the dynamics of the paradigm. In this work, we have taken the most general form of such a scenario: Goldstone inflation, and studied the model in the noncanonical domain. Natural inflation is a limiting case of this model, which is also studied here in the noncanonical regime. Our result is compared with the recent release by Planck collaboration and it is shown that for some combination of the model parameters, a Goldstone inflationary model in the noncanonical realisation obeys the current observational bounds. Then, we studied the era of reheating after the end of inflation. For different choice of model parameters, constraints on the reheating temperature ($T_{\rm re}$) and number of e-folds during reheating($N_{\rm re}$) for the allowed inflationary observables (e.g. scalar spectral index($n_s$) and tensor to scalar ratio($r$)) are predicted for this model.
In this work, we consider a tight binding lattice with two non-Hermitian impurities. The system is described by a non-Hermitian generalization of the Aubry Andre model. We show for the first time that there exists topologically nontrivial edge states with real spectra in the PT symmetric region.
A search for pair production of second-generation scalar leptoquarks in the final state with two muons and two jets is performed using proton-proton collision data at sqrt(s) = 7 TeV collected by the CMS detector at the LHC. The data sample used corresponds to an integrated luminosity of 34 inverse picobarns. The number of observed events is in good agreement with the predictions from the standard model processes. An upper limit is set on the second-generation leptoquark cross section times beta^2 as a function of the leptoquark mass, and leptoquarks with masses below 394 GeV are excluded at a 95% confidence level for beta = 1, where beta is the leptoquark branching fraction into a muon and a quark. These limits are the most stringent to date.
We discuss stability for a class of learning algorithms with respect to noisy labels. The algorithms we consider are for regression, and they involve the minimization of regularized risk functionals, such as L(f) := 1/N sum_i (f(x_i)-y_i)^2+ lambda ||f||_H^2. We shall call the algorithm `stable' if, when y_i is a noisy version of f*(x_i) for some function f* in H, the output of the algorithm converges to f* as the regularization term and noise simultaneously vanish. We consider two flavors of this problem, one where a data set of N points remains fixed, and the other where N -> infinity. For the case where N -> infinity, we give conditions for convergence to f_E (the function which is the expectation of y(x) for each x), as lambda -> 0. For the fixed N case, we describe the limiting 'non-noisy', 'non-regularized' function f*, and give conditions for convergence. In the process, we develop a set of tools for dealing with functionals such as L(f), which are applicable to many other problems in learning theory.
We present a first result towards the use of entailment in- side relational dual tableau-based decision procedures. To this end, we introduce a fragment of RL(1) which admits a restricted form of composition, (R ; S) or (R ; 1), where the left subterm R of (R ; S) is only allowed to be either the constant 1, or a Boolean term neither containing the complement operator nor the constant 1, while in the case of (R ; 1), R can only be a Boolean term involving relational variables and the operators of intersection and of union. We prove the decidability of the fragment by defining a dual tableau- based decision procedure with a suitable blocking mechanism and where the rules to decompose compositional formulae are modified so to deal with the constant 1 while preserving termination. The fragment properly includes the logics presented in previous work and, therefore, it allows one to express, among others, the multi-modal logic K with union and intersection of accessibility relations, and the description logic ALC with union and intersection of roles.
Beginning with Anderson (1972), spontaneous symmetry breaking (SSB) in infinite quantum systems is often put forward as an example of (asymptotic) emergence in physics, since in theory no finite system should display it. Even the correspondence between theory and reality is at stake here, since numerous real materials show SSB in their ground states (or equilibrium states at low temperature), although they are finite. Thus against what is sometimes called `Earman's Principle', a genuine physical effect (viz. SSB) seems theoretically recovered only in some idealization (namely the thermodynamic limit), disappearing as soon as the the idealization is removed. We review the well-known arguments that (at first sight) no finite system can exhibit SSB, using the formalism of algebraic quantum theory in order to control the thermodynamic limit and unify the description of finite- and infinite-volume systems. Using the striking mathematical analogy between the thermodynamic limit and the classical limit, we show that a similar situation obtains in quantum mechanics (which typically forbids SSB) versus classical mechanics (which allows it). This discrepancy between formalism and reality is quite similar to the measurement problem, and hence we address it in the same way, adapting an argument of the author and Reuvers (2013) that was originally intended to explain the collapse of the wave-function within conventional quantum mechanics. Namely, exponential sensitivity to (asymmetric) perturbations of the (symmetric) dynamics as the system size increases causes symmetry breaking already in finite but very large quantum systems. This provides continuity between finite- and infinite-volume descriptions of quantum systems featuring SSB and hence restores Earman's Principle (at least in this particularly threatening case).
We propose a new scenario for compound chondrule formation named as "fragment-collision model," in the framework of the shock-wave heating model. A molten cm-sized dust particle (parent) is disrupted in the high-velocity gas flow. The extracted fragments (ejectors) are scattered behind the parent and the mutual collisions between them will occur. We modeled the disruption event by analytic considerations in order to estimate the probability of the mutual collisions assuming that all ejectors have the same radius. We found that the estimated collision probability, which is the probability of collisions experienced by an ejector in one disruption event, can account for the observed fraction of compound chondrules. In addition, the model predictions are qualitatively consistent with other observational data (oxygen isotopic composition, textural types, and size ratios of constituents). Based on these results, we concluded that this new model can be one of the strongest candidates for the compound chondrule formation. It should be noted that all collisions do not necessarily lead to the compound chondrule formation. The formation efficiency and the future works which should be investigated in the forthcoming paper are also discussed.
We sketch the basic ideas of the lattice regularization in Quantum Field Theory, the corresponding Monte Carlo simulations, and applications to Quantum Chromodynamics (QCD). This approach enables the numerical measurement of observables at the non-perturbative level. We comment on selected results, with a focus on hadron masses and the link to Chiral Perturbation Theory. At last we address two outstanding issues: topological freezing and the sign problem.
Advances of deep learning for Artificial Neural Networks(ANNs) have led to significant improvements in the performance of digital signal processing systems implemented on digital chips. Although recent progress in low-power chips is remarkable, neuromorphic chips that run Spiking Neural Networks (SNNs) based applications offer an even lower power consumption, as a consequence of the ensuing sparse spike-based coding scheme. In this work, we develop a SNN-based Voice Activity Detection (VAD) system that belongs to the building blocks of any audio and speech processing system. We propose to use the bin encoding, a novel method to convert log mel filterbank bins of single-time frames into spike patterns. We integrate the proposed scheme in a bilayer spiking architecture which was evaluated on the QUT-NOISE-TIMIT corpus. Our approach shows that SNNs enable an ultra low-power implementation of a VAD classifier that consumes only 3.8$\mu$W, while achieving state-of-the-art performance.
Code comment generation which aims to automatically generate natural language descriptions for source code, is a crucial task in the field of automatic software development. Traditional comment generation methods use manually-crafted templates or information retrieval (IR) techniques to generate summaries for source code. In recent years, neural network-based methods which leveraged acclaimed encoder-decoder deep learning framework to learn comment generation patterns from a large-scale parallel code corpus, have achieved impressive results. However, these emerging methods only take code-related information as input. Software reuse is common in the process of software development, meaning that comments of similar code snippets are helpful for comment generation. Inspired by the IR-based and template-based approaches, in this paper, we propose a neural comment generation approach where we use the existing comments of similar code snippets as exemplars to guide comment generation. Specifically, given a piece of code, we first use an IR technique to retrieve a similar code snippet and treat its comment as an exemplar. Then we design a novel seq2seq neural network that takes the given code, its AST, its similar code, and its exemplar as input, and leverages the information from the exemplar to assist in the target comment generation based on the semantic similarity between the source code and the similar code. We evaluate our approach on a large-scale Java corpus, which contains about 2M samples, and experimental results demonstrate that our model outperforms the state-of-the-art methods by a substantial margin.
Recent strides in nonlinear model predictive control (NMPC) underscore a dependence on numerical advancements to efficiently and accurately solve large-scale problems. Given the substantial number of variables characterizing typical whole-body optimal control (OC) problems - often numbering in the thousands - exploiting the sparse structure of the numerical problem becomes crucial to meet computational demands, typically in the range of a few milliseconds. Addressing the linear-quadratic regulator (LQR) problem is a fundamental building block for computing Newton or Sequential Quadratic Programming (SQP) steps in direct optimal control methods. This paper concentrates on equality-constrained problems featuring implicit system dynamics and dual regularization, a characteristic of advanced interiorpoint or augmented Lagrangian solvers. Here, we introduce a parallel algorithm for solving an LQR problem with dual regularization. Leveraging a rewriting of the LQR recursion through block elimination, we first enhanced the efficiency of the serial algorithm and then subsequently generalized it to handle parametric problems. This extension enables us to split decision variables and solve multiple subproblems concurrently. Our algorithm is implemented in our nonlinear numerical optimal control library ALIGATOR. It showcases improved performance over previous serial formulations and we validate its efficacy by deploying it in the model predictive control of a real quadruped robot.
High-order Gaussian beams with multiple propagation modes have been studied for free-space optical communications. Fast classification of beams using a diffractive deep neural network, D2NN, has been proposed. D2NN optimization is important because it has numerous hyperparameters, such as interlayer distances and mode combinations. In this study, we classify Hermite-Gaussian beams, which are high-order Gaussian beams, using a D2NN, and automatically tune one of its hyperparameters known as the interlayer distance. We used the tree-structured Parzen estimator, a hyperparameter auto-tuning algorithm, to search for the best model. Results indicated that classification accuracy obtained by auto-tuning hyperparameters was higher than that obtained by manually setting interlayer distances at equal intervals. In addition, we confirmed that accuracy by auto-tuning improves as the number of classification modes increases.
The moduli space $\mathcal{G}^r_{g,d} \to \mathcal{M}_g$ parameterizing algebraic curves with a linear series of degree $d$ and rank $r$ has expected relative dimension $\rho = g - (r+1)(g-d+r)$. Classical Brill-Noether theory concerns the case $\rho \geq 0$; we consider the non-surjective case $\rho < 0$. We prove the existence of components of this moduli space with the expected relative dimension when $0 > \rho \geq -g+3$, or $0 > \rho \geq -C_r g + \mathcal{O}(g^{5/6})$, where $C_r$ is a constant depending on the rank of the linear series such that $C_r \to 3$ as $r \to \infty$. These results are proved via a two-marked-point generalization suitable for inductive arguments, and the regeneration theorem for limit linear series.
We give a construction of generalized cluster varieties and generalized cluster scattering diagrams for reciprocal generalized cluster algebras, the latter of which were defined by Chekhov and Shapiro. These constructions are analogous to the structures given for ordinary cluster algebras in the work of Gross, Hacking, Keel, and Kontsevich. As a consequence of these constructions, we are also able to construct theta functions for generalized cluster algebras, again in the reciprocal case, and demonstrate a number of their structural properties.
We study the optical and electrical properties of silver films with a graded thickness obtained through metallic evaporation in vacuum on a tilted substrate to evaluate their use as semitransparent electrical contacts. We measure their ellipsometric coefficients, optical transmissions and electrical conductivity for different widths, and we employ an efficient recursive method to calculate their macroscopic dielectric function, their optical properties and their microscopic electric fields. The topology of very thin films corresponds to disconnected islands, while very wide films are simply connected. For intermediate widths the film becomes semicontinuous, multiply connected, and its microscopic electric field develops hotspots at optical resonances which appear near the percolation threshold of the conducting phase, yielding large ohmic losses that increase the absorptance above that of a corresponding homogeneous film. Optimizing the thickness of the film to maximize its transmittance above the percolation threshold of the conductive phase we obtained a film with transmittance T = 0.41 and a sheet resistance $R_{\square}^{\mathrm{max}}\approx2.7\Omega$. We also analyze the observed emission frequency shift of porous silicon electroluminescent devices when Ag films are used as solid electrical contacts in replacement of electrolytic ones.
We present a general formulation for ray-tracing calculations in curved space-time. The formulation takes full account of relativistic effects in the photon transport and the relative motions of the emitters and the light-of-sight absorbing material. We apply the formulation to calculate the emission from accretion disks and tori around rotating black holes.
Ultracold molecules are associated from an atomic Bose-Einstein condensate by ramping a magnetic field across a Feshbach resonance. The reverse ramp dissociates the molecules. The kinetic energy released in the dissociation process is used to measure the widths of 4 Feshbach resonances in 87Rb. This method to determine the width works remarkably well for narrow resonances even in the presence of significant magnetic-field noise. In addition, a quasi-mono-energetic atomic wave is created by jumping the magnetic field across the Feshbach resonance.
Motivated by the motion of biopolymers and membranes in solution, this article presents a formulation of the equations of motion for curves and surfaces in a viscous fluid. We focus on geometrical aspects and simple variational methods for calculating internal stresses and forces, and we derive the full nonlinear equations of motion. In the case of membranes, we pay particular attention to the formulation of the equations of hydrodynamics on a curved, deforming surface. The formalism is illustrated by two simple case studies: (1) the twirling instability of straight elastic rod rotating in a viscous fluid, and (2) the pearling and buckling instabilities of a tubular liposome or polymersome.
Our earlier papers explore the nature of large wave vector spin waves in ultrathin ferromagnets, and also the properties and damping of spin waves of zero wave vector, at the center of the two dimensional Brillouin zone, with application to FMR studies. The present paper explores the behavior of spin waves in such films at intermediate wave vectors, which connect the two regimes. For the case of Fe films on Au(100), we study the wave vector dependence of the linewidth of the lowest frequency mode, to find that it contains a term which varies as the fourth power of the wave vector. It is argued that this behavior is expected quite generally. We also explore the nature of the eigenvectors of the two lowest lying modes of the film, as a function of wave vector. Interestingly, as wave vector increases, the lowest mode localizes onto the interface between the film and the substrate, while the second mode evolves into a surface spin wave, localized on the outer layer. We infer similar behavior for a Co film on Cu(100), though this evolution occurs at rather larger wave vectors where, as we have shown previously, the modes are heavily damped with the consequence that identification of distinct eigenmodes is problematical.
Causal models communicate our assumptions about causes and effects in real-world phe- nomena. Often the interest lies in the identification of the effect of an action which means deriving an expression from the observed probability distribution for the interventional distribution resulting from the action. In many cases an identifiability algorithm may return a complicated expression that contains variables that are in fact unnecessary. In practice this can lead to additional computational burden and increased bias or inefficiency of estimates when dealing with measurement error or missing data. We present graphical criteria to detect variables which are redundant in identifying causal effects. We also provide an improved version of a well-known identifiability algorithm that implements these criteria.
Shubnikov-de Haas (SdH) oscillations are observed in Bi2Se3 flakes with high carrier concentration and low bulk mobility. These oscillations probe the protected surface states and enable us to extract their carrier concentration, effective mass and Dingle temperature. The Fermi momentum obtained is in agreement with angle resolved photoemission spectroscopy (ARPES) measurements performed on crystals from the same batch. We study the behavior of the Berry phase as a function of magnetic field. The standard theoretical considerations fail to explain the observed behavior.
We consider the problem of optimal recovery of true ranking of $n$ items from a randomly chosen subset of their pairwise preferences. It is well known that without any further assumption, one requires a sample size of $\Omega(n^2)$ for the purpose. We analyze the problem with an additional structure of relational graph $G([n],E)$ over the $n$ items added with an assumption of \emph{locality}: Neighboring items are similar in their rankings. Noting the preferential nature of the data, we choose to embed not the graph, but, its \emph{strong product} to capture the pairwise node relationships. Furthermore, unlike existing literature that uses Laplacian embedding for graph based learning problems, we use a richer class of graph embeddings---\emph{orthonormal representations}---that includes (normalized) Laplacian as its special case. Our proposed algorithm, {\it Pref-Rank}, predicts the underlying ranking using an SVM based approach over the chosen embedding of the product graph, and is the first to provide \emph{statistical consistency} on two ranking losses: \emph{Kendall's tau} and \emph{Spearman's footrule}, with a required sample complexity of $O(n^2 \chi(\bar{G}))^{\frac{2}{3}}$ pairs, $\chi(\bar{G})$ being the \emph{chromatic number} of the complement graph $\bar{G}$. Clearly, our sample complexity is smaller for dense graphs, with $\chi(\bar G)$ characterizing the degree of node connectivity, which is also intuitive due to the locality assumption e.g. $O(n^\frac{4}{3})$ for union of $k$-cliques, or $O(n^\frac{5}{3})$ for random and power law graphs etc.---a quantity much smaller than the fundamental limit of $\Omega(n^2)$ for large $n$. This, for the first time, relates ranking complexity to structural properties of the graph. We also report experimental evaluations on different synthetic and real datasets, where our algorithm is shown to outperform the state-of-the-art methods.
Despite many years of efforts, attempts to reach the quantum regime of topological surface states (TSS) on an electrically tunable topological insulator (TI) platform have so far failed on binary TI compounds such as Bi2Se3 due to high density of interfacial defects. Here, utilizing an optimal buffer layer on a gatable substrate, we demonstrate the first electrically tunable quantum Hall effects (QHE) on TSS of Bi2Se3. On the n-side, well-defined QHE shows up, but it diminishes near the charge neutrality point (CNP) and completely disappears on the p-side. Furthermore, around the CNP the system transitions from a metallic to a highly resistive state as the magnetic field is increased, whose temperature dependence indicates presence of an insulating ground state at high magnetic fields.
We show that for any uncountable cardinal $\lambda$, the category of sets of cardinality at least $\lambda$ and monomorphisms between them cannot appear as the category of point of a topos, in particular is not the category of models of a $L_{(\infty,\omega)}$-theory. More generally we show that for any regular cardinal $\kappa < \lambda$ it is neither the category of $\kappa$-points of a $\kappa$-topos, in particular, not the category of models of a $L_{(\infty,\kappa)}$-theory. The proof relies on the construction of a categorified version of the Scott topology, which constitute a left adjoint to the functor sending any topos to its category of points and the computation of this left adjoint evaluated on the category of sets of cardinality at least $\lambda$ and monomorphisms between them. The same techniques also applies to a few other categories. At least to the category of vector spaces of with bounded below dimension and the category of algebraic closed fields of fixed characteristic with bounded below transcendence degree.
Let (M, g) be an asymptotically flat static vacuum initial data set with non-empty compact boundary. We prove that (M, g) is isometric to a spacelike slice of a Schwarzschild spacetime under the mere assumption that the boundary of (M, g) has zero mean curvature, hence generalizing a classic result of Bunting and Masood-ul-Alam. In the case that the boundary has constant positive mean curvature and satisfies a stability condition, we derive an upper bound of the ADM mass of (M, g) in terms of the area and mean curvature of the boundary. Our discussion is motivated by Bartnik's quasi-local mass definition.
State-of-the-art methods for handwriting recognition are based on Long Short Term Memory (LSTM) recurrent neural networks (RNN), which now provides very impressive character recognition performance. The character recognition is generally coupled with a lexicon driven decoding process which integrates dictionaries. Unfortunately these dictionaries are limited to hundred of thousands words for the best systems, which prevent from having a good language coverage, and therefore limit the global recognition performance. In this article, we propose an alternative to the lexicon driven decoding process based on a lexicon verification process, coupled with an original cascade architecture. The cascade is made of a large number of complementary networks extracted from a single training (called cohort), making the learning process very light. The proposed method achieves new state-of-the art word recognition performance on the Rimes and IAM databases. Dealing with gigantic lexicon of 3 millions words, the methods also demonstrates interesting performance with a fast decision stage.
This Letter presents a novel structured light system model that effectively considers local lens distortion by pixel-wise rational functions. We leverage the stereo method for initial calibration and then estimate the rational model for each pixel. Our proposed model can achieve high measurement accuracy within and outside the calibration volume, demonstrating its robustness and accuracy.
The apps installed on a smartphone can reveal much information about a user, such as their medical conditions, sexual orientation, or religious beliefs. Additionally, the presence or absence of particular apps on a smartphone can inform an adversary who is intent on attacking the device. In this paper, we show that a passive eavesdropper can feasibly identify smartphone apps by fingerprinting the network traffic that they send. Although SSL/TLS hides the payload of packets, side-channel data such as packet size and direction is still leaked from encrypted connections. We use machine learning techniques to identify smartphone apps from this side-channel data. In addition to merely fingerprinting and identifying smartphone apps, we investigate how app fingerprints change over time, across devices and across different versions of apps. Additionally, we introduce strategies that enable our app classification system to identify and mitigate the effect of ambiguous traffic, i.e., traffic in common among apps such as advertisement traffic. We fully implemented a framework to fingerprint apps and ran a thorough set of experiments to assess its performance. We fingerprinted 110 of the most popular apps in the Google Play Store and were able to identify them six months later with up to 96% accuracy. Additionally, we show that app fingerprints persist to varying extents across devices and app versions.
We discuss $4$-dimensional achiral Lefschetz fibrations bounding $3$-dimensional open books and study their Lefschetz fibration (LF) embedding in a bounded $6$-dimensional manifold, in the sense of Ghanwat--Pancholi. As an application we give another proof of the fact that every closed orientable $4$-manifold embeds in $S^2 \times S^2 \times S^2$ . We also show that every achiral Lefschetz fibration with hyperelliptic monodromy admits LF embedding in $D^6 = D^2 \times D^4$ and discuss an obstruction to such LF embeddings.
This paper summarizes the discussions, presentations, and activity of the Future Light Sources Workshop 2012 (FLS 2012) working group dedicated to Electron Sources. The focus of the working group was to discuss concepts and technologies that might enable much higher peak and average brightness from electron beam sources. Furthermore the working group was asked to consider methods to greatly improve the robustness of operation and lower the costs of providing electrons.
In this paper, we report a capillary-based M-Z interferometer that could be used for precise detection of variations in refractive indices of gaseous samples. This sensing mechanism is quite straightforward. Cladding and core modes of a capillary are simultaneously excited by coupling coherent laser beams to the capillary cladding and core, respectively. Interferogram would be generated as the light transmitted from the core interferes with the light transmitted from the cladding. Variations in refractive index of the air filling the core lead to variations in phase difference between the core and cladding modes, thus shifting the interference fringes. Using a photodiode together with a narrow slit, we could analyze the fringe shifts. The resolution of the sensor was found to be 1*10-8 RIU, that is comparable to the highest resolution obtained by other interferometric sensors reported in previous literatures. Finally, we also analyze the temperature cross sensitivity of the sensor. The advantages of our sensor include very low cost, high sensitivity, straightforward sensing mechanism, and ease of fabrication.
AT2019wey is a Galactic low mass X-ray binary with a candidate black hole accretor first discovered as an optical transient by ATLAS in December 2019. It was then associated with an X-ray source discovered by SRG in March 2020. After observing a brightening in X-rays in August 2020, VLA observations of the source revealed an optically thin spectrum that subsequently shifted to optically thick, as the source continued to brighten in radio. This motivated observations of the source with the VLBA. We found a resolved source that we interpret to be a steady compact jet, a feature associated with black hole X-ray binary systems in the hard X-ray spectral state. The jet power is comparable to the accretion-disk X-ray luminosity. Here, we summarize the results from these observations.
Natural language offers a highly intuitive interface for image editing. In this paper, we introduce the first solution for performing local (region-based) edits in generic natural images, based on a natural language description along with an ROI mask. We achieve our goal by leveraging and combining a pretrained language-image model (CLIP), to steer the edit towards a user-provided text prompt, with a denoising diffusion probabilistic model (DDPM) to generate natural-looking results. To seamlessly fuse the edited region with the unchanged parts of the image, we spatially blend noised versions of the input image with the local text-guided diffusion latent at a progression of noise levels. In addition, we show that adding augmentations to the diffusion process mitigates adversarial results. We compare against several baselines and related methods, both qualitatively and quantitatively, and show that our method outperforms these solutions in terms of overall realism, ability to preserve the background and matching the text. Finally, we show several text-driven editing applications, including adding a new object to an image, removing/replacing/altering existing objects, background replacement, and image extrapolation. Code is available at: https://omriavrahami.com/blended-diffusion-page/
Biunit pairs are introduced as pairs of elements in a semiheap that generalize the notion of unit. Families of functions generalizing involutions and conjugations, called switches and warps, are investigated. The main theorem establishes that there is a one-to-one correspondence between monoids equipped with a particular switch and semiheaps with a biunit pair. This generalizes a well-established result in semiheap theory that connects involuted semigroups and semiheaps with biunit elements. Furthermore, diheaps are introduced as semiheaps whose elements belong to biunit pairs and they are shown to be non-isomorphic but warp-equivalent to heaps.
Plasmonic antennas are attractive optical structures for many applications in nano and quantum technologies. By providing enhanced interaction between a nanoemitter and light, they efficiently accelerate and direct spontaneous emission. One challenge, however, is the precise nanoscale positioning of the emitter in the structure. Here we present a laser etching protocol that deterministically positions a single colloidal CdSe/CdS core/shell quantum dot emitter inside a subwavelength plasmonic patch antenna with three-dimensional nanoscale control. By exploiting the properties of metal-insulator-metal structures at the nanoscale, the fabricated single emitter antenna exhibits an extremely high Purcell factor (>72) and brightness enhancement by a factor of 70. Due to the unprecedented quenching of Auger processes and the strong acceleration of multiexciton emission, more than 4 photons per pulse can be emitted by a single quantum dot. Our technology permits the fabrication of bright room-temperature single-emitter sources emitting either multiple or single photons.
We present the first lattice QCD calculation of coupled $\pi\omega$ and $\pi\phi$ scattering, incorporating coupled $S$ and $D$-wave $\pi\omega$ in $J^P=1^+$. Finite-volume spectra in three volumes are determined via a variational analysis of matrices of two-point correlation functions, computed using large bases of operators resembling single-meson, two-meson and three-meson structures, with the light-quark mass corresponding to a pion mass of $m_\pi \approx 391$ MeV. Utilizing the relationship between the discrete spectrum of finite-volume energies and infinite-volume scattering amplitudes, we find a narrow axial-vector resonance ($J^{PC}=1^{+-}$), the analogue of the $b_1$ meson, with mass $m_{R}\approx1380$ MeV and width $\Gamma_{R}\approx 91$ MeV. The resonance is found to couple dominantly to $S$-wave $\pi\omega$, with a much-suppressed coupling to $D$-wave $\pi\omega$, and a negligible coupling to $\pi\phi$ consistent with the `OZI rule'. No resonant behavior is observed in $\pi\phi$, indicating the absence of a putative low-mass $Z_s$ analogue of the $Z_c$ claimed in $\pi J/\psi$. In order to minimally present the contents of a unitary three-channel scattering matrix, we introduce an $n$-channel generalization of the traditional two-channel Stapp parameterization.
Generalized category discovery (GCD) is a recently proposed open-world task. Given a set of images consisting of labeled and unlabeled instances, the goal of GCD is to automatically cluster the unlabeled samples using information transferred from the labeled dataset. The unlabeled dataset comprises both known and novel classes. The main challenge is that unlabeled novel class samples and unlabeled known class samples are mixed together in the unlabeled dataset. To address the GCD without knowing the class number of unlabeled dataset, we propose a co-training-based framework that encourages clustering consistency. Specifically, we first introduce weak and strong augmentation transformations to generate two sufficiently different views for the same sample. Then, based on the co-training assumption, we propose a consistency representation learning strategy, which encourages consistency between feature-prototype similarity and clustering assignment. Finally, we use the discriminative embeddings learned from the semi-supervised representation learning process to construct an original sparse network and use a community detection method to obtain the clustering results and the number of categories simultaneously. Extensive experiments show that our method achieves state-of-the-art performance on three generic benchmarks and three fine-grained visual recognition datasets. Especially in the ImageNet-100 data set, our method significantly exceeds the best baseline by 15.5\% and 7.0\% on the \texttt{Novel} and \texttt{All} classes, respectively.
We report on two quantitative, morphological estimators of the filamentary structure of the Cosmic Web, the so-called global and local skeletons. The first, based on a global study of the matter density gradient flow, allows us to study the connectivity between a density peak and its surroundings, with direct relevance to the anisotropic accretion via cold flows on galactic halos. From the second, based on a local constraint equation involving the derivatives of the field, we can derive predictions for powerful statistics, such as the differential length and the relative saddle to extrema counts of the Cosmic web as a function of density threshold (with application to percolation of structures and connectivity), as well as a theoretical framework to study their cosmic evolution through the onset of gravity-induced non-linearities.
We present a simple method to quantitatively capture the heterogeneity in the degree distribution of a network graph using a single parameter $\sigma$. Using an exponential transformation of the shape parameter of the Weibull distribution, this control parameter allows the degree distribution to be easily interpolated between highly symmetric and highly heterogeneous distributions on the unit interval. This parameterization of heterogeneity also recovers several other canonical distributions as intermediate special cases, including the Gaussian, Rayleigh, and exponential distributions. We then outline a general graph generation algorithm to produce graphs with a desired amount of heterogeneity. The utility of this formulation of a heterogeneity parameter is demonstrated with examples relating to epidemiological modeling and spectral analysis.
We prove that the 2d Euler equation is globally well-posed in a space of vector fields having spatial asymptotic expansion at infinity of any a priori given order. The asymptotic coefficients of the solutions are holomorphic functions of $t$, do not involve (spacial) logarithmic terms, and develop even when the initial data has fast decay at infinity. We discuss the evolution in time of the asymptotic terms and their approximation properties.
Recent neural networks based surface reconstruction can be roughly divided into two categories, one warping templates explicitly and the other representing 3D surfaces implicitly. To enjoy the advantages of both, we propose a novel 3D representation, Neural Vector Fields (NVF), which adopts the explicit learning process to manipulate meshes and implicit unsigned distance function (UDF) representation to break the barriers in resolution and topology. This is achieved by directly predicting the displacements from surface queries and modeling shapes as Vector Fields, rather than relying on network differentiation to obtain direction fields as most existing UDF-based methods do. In this way, our approach is capable of encoding both the distance and the direction fields so that the calculation of direction fields is differentiation-free, circumventing the non-trivial surface extraction step. Furthermore, building upon NVFs, we propose to incorporate two types of shape codebooks, \ie, NVFs (Lite or Ultra), to promote cross-category reconstruction through encoding cross-object priors. Moreover, we propose a new regularization based on analyzing the zero-curl property of NVFs, and implement this through the fully differentiable framework of our NVF (ultra). We evaluate both NVFs on four surface reconstruction scenarios, including watertight vs non-watertight shapes, category-agnostic reconstruction vs category-unseen reconstruction, category-specific, and cross-domain reconstruction.
We investigate the $k$-error linear complexity of pseudorandom binary sequences of period $p^{\mathfrak{r}}$ derived from the Euler quotients modulo $p^{\mathfrak{r}-1}$, a power of an odd prime $p$ for $\mathfrak{r}\geq 2$. When $\mathfrak{r}=2$, this is just the case of polynomial quotients (including Fermat quotients) modulo $p$, which has been studied in an earlier work of Chen, Niu and Wu. In this work, we establish a recursive relation on the $k$-error linear complexity of the sequences for the case of $\mathfrak{r}\geq 3$. We also state the exact values of the $k$-error linear complexity for the case of $\mathfrak{r}=3$. From the results, we can find that the $k$-error linear complexity of the sequences (of period $p^{\mathfrak{r}}$) does not decrease dramatically for $k<p^{\mathfrak{r}-2}(p-1)^2/2$.
Autonomous mobile robots (e.g., warehouse logistics robots) often need to traverse complex, obstacle-rich, and changing environments to reach multiple fixed goals (e.g., warehouse shelves). Traditional motion planners need to calculate the entire multi-goal path from scratch in response to changes in the environment, which result in a large consumption of computing resources. This process is not only time-consuming but also may not meet real-time requirements in application scenarios that require rapid response to environmental changes. In this paper, we provide a novel Multi-Goal Motion Memory technique that allows robots to use previous planning experiences to accelerate future multi-goal planning in changing environments. Specifically, our technique predicts collision-free and dynamically-feasible trajectories and distances between goal pairs to guide the sampling process to build a roadmap, to inform a Traveling Salesman Problem (TSP) solver to compute a tour, and to efficiently produce motion plans. Experiments conducted with a vehicle and a snake-like robot in obstacle-rich environments show that the proposed Motion Memory technique can substantially accelerate planning speed by up to 90\%. Furthermore, the solution quality is comparable to state-of-the-art algorithms and even better in some environments.
We develop a theoretical approach for nuclear spectral functions at high missing momenta and removal energies based on the multi-nucleon short-range correlation~(SRC) model. The approach is based on the effective Feynman diagrammatic method which allows to account for the relativistic effects important in the SRC domain. In addition to two-nucleon SRC with center of mass motion we derive also the contribution of three-nucleon SRCs to the nuclear spectral functions. The latter is modeled based on the assumption that 3N SRCs are a product of two sequential short range NN interactions. This approach allowed us to express the 3N SRC part of the nuclear spectral function as a convolution of two NN SRCs. Thus the knowledge of 2N SRCs allows us to model both two- and three-nucleon SRC contributions to the spectral function. The derivations of the spectral functions are based on the two theoretical frameworks in evaluating covariant Feynman diagrams: In the first, referred as virtual nucleon approximation, we reduce Feynman diagrams to the time ordered noncovariant diagrams by evaluating nucleon spectators in the SRC at their positive energy poles, neglecting explicitly the contribution from vacuum diagrams. In the second approach, referred as light-front approximation, we formulate the boost invariant nuclear spectral function in the light-front reference frame in which case the vacuum diagrams are generally suppressed and the bound nucleon is described by its light-cone variables such as momentum fraction, transverse momentum and invariant mass.
Ubiquitous systems with End-Edge-Cloud architecture are increasingly being used in healthcare applications. Federated Learning (FL) is highly useful for such applications, due to silo effect and privacy preserving. Existing FL approaches generally do not account for disparities in the quality of local data labels. However, the clients in ubiquitous systems tend to suffer from label noise due to varying skill-levels, biases or malicious tampering of the annotators. In this paper, we propose Federated Opportunistic Computing for Ubiquitous Systems (FOCUS) to address this challenge. It maintains a small set of benchmark samples on the FL server and quantifies the credibility of the client local data without directly observing them by computing the mutual cross-entropy between performance of the FL model on the local datasets and that of the client local FL model on the benchmark dataset. Then, a credit weighted orchestration is performed to adjust the weight assigned to clients in the FL model based on their credibility values. FOCUS has been experimentally evaluated on both synthetic data and real-world data. The results show that it effectively identifies clients with noisy labels and reduces their impact on the model performance, thereby significantly outperforming existing FL approaches.
We study by kinetic Monte Carlo simulations the dynamic behavior of a Ziff-Gulari-Barshad model with CO desorption for the reaction CO + O $\to$ CO$_2$ on a catalytic surface. Finite-size scaling analysis of the fluctuations and the fourth-order order-parameter cumulant show that below a critical CO desorption rate, the model exhibits a nonequilibrium first-order phase transition between low and high CO coverage phases. We calculate several points on the coexistence curve. We also measure the metastable lifetimes associated with the transition from the low CO coverage phase to the high CO coverage phase, and {\it vice versa}. Our results indicate that the transition process follows a mechanism very similar to the decay of metastable phases associated with {\it equilibrium} first-order phase transitions and can be described by the classic Kolmogorov-Johnson-Mehl-Avrami theory of phase transformation by nucleation and growth. In the present case, the desorption parameter plays the role of temperature, and the distance to the coexistence curve plays the role of an external field or supersaturation. We identify two distinct regimes, depending on whether the system is far from or close to the coexistence curve, in which the statistical properties and the system-size dependence of the lifetimes are different, corresponding to multidroplet or single-droplet decay, respectively. The crossover between the two regimes approaches the coexistence curve logarithmically with system size, analogous to the behavior of the crossover between multidroplet and single-droplet metastable decay near an equilibrium first-order phase transition.
Aircraft models may be considered as flat if one neglects some terms associated to aerodynamics. Computational experiments in Maple show that in some cases a suitably designed feed-back allows to follow such trajectories, when applied to the non-flat model. However some maneuvers may be hard or even impossible to achieve with this flat approximation. In this paper, we propose an iterated process to compute a more achievable trajectory, starting from the flat reference trajectory. More precisely, the unknown neglected terms in the flat model are iteratively re-evaluated using the values obtained at the previous step. This process may be interpreted as a new trajectory parametrization, using an infinite number of derivatives, a property that may be called \emph{generalized flatness}. We illustrate the pertinence of this approach in flight conditions of increasing difficulties, from single engine flight, to aileron roll.
Exact expressions for ensemble averaged Madelung energies of finite volumes are derived. The extrapolation to the thermodynamic limit converges unconditionally and can be used as a parameter-free real-space summation method of Madelung constants. In the large volume limit, the surface term of the ensemble averaged Madelung energy has a universal form, independent of the crystal structure. The scaling of the Madelung energy with system size provides a simple explanation for the structural phase transition observed in cesium halide clusters.
Underlying many Bayesian inference techniques that seek to approximate the posterior as a Gaussian distribution is a fundamental linear algebra problem that must be solved for both the mean and key entries of the covariance. Even when the true posterior is not Gaussian (e.g., in the case of nonlinear measurement functions) we can use variational schemes that repeatedly solve this linear algebra problem at each iteration. In most cases, the question is not whether a solution to this problem exists, but rather how we can exploit problem-specific structure to find it efficiently. Our contribution is to clearly state the Fundamental Linear Algebra Problem of Gaussian Inference (FLAPOGI) and to provide a novel presentation (using Kronecker algebra) of the not-so-well-known result of Takahashi et al. (1973) that makes it possible to solve for key entries of the covariance matrix. We first provide a global solution and then a local version that can be implemented using local message passing amongst a collection of agents calculating in parallel. Contrary to belief propagation, our local scheme is guaranteed to converge in both the mean and desired covariance quantities to the global solution even when the underlying factor graph is loopy; in the case of synchronous updates, we provide a bound on the number of iterations required for convergence. Compared to belief propagation, this guaranteed convergence comes at the cost of additional storage, calculations, and communication links in the case of loops; however, we show how these can be automatically constructed on the fly using only local information.
Binary pulsars provide some of the tightest current constraints on modified theories of gravity and these constraints will only get tighter as radio astronomers continue timing these systems. These binary pulsars are particularly good at constraining scalar-tensor theories in which gravity is mediated by a scalar field in addition to the metric tensor. Scalar-tensor theories can predict large deviations from General Relativity due to the fact that they allow for violation of the strong-equivalence principle through a phenomenon known as scalarization. This effect appears directly in the timing model for binary pulsars, and as such, it can be tightly constrained through precise timing. In this paper, we investigate these constraints for two scalar-tensor theories and a large set of realistic equations of state. We calculate the constraints that can be placed by saturating the current $1\sigma$ bounds on single post-Keplerian parameters, as well as employing Bayesian methods through Markov-Chain-Monte-Carlo simulations to explore the constraints that can be achieved when one considers all measured parameters simultaneously. Our results demonstrate that both methods are able to place similar constraints and that they are both indeed dominated by the measurements of the orbital period decay. The Bayesian approach, however, allows one to simultaneously explore the posterior distributions of not only the theory parameters but of the masses as well.
We discuss scalar quantum field theories in a Lorentz-invariant three-dimensional noncommutative space-time. We first analyze the one-loop diagrams of the two-point functions, and show that the non-planar diagrams are finite and have infrared singularities from the UV/IR mixing. The scalar quantum field theories have the problem that the violation of the momentum conservation from the non-planar diagrams does not vanish even in the commutative limit. A way to obtain an exact translational symmetry by introducing an infinite number of tensor fields is proposed. The translational symmetry transforms local fields into non-local ones in general. We also discuss an analogue of thermodynamics of free scalar field theory in the noncommutative space-time.
Graphs naturally appear in several real-world contexts including social networks, the web network, and telecommunication networks. While the analysis and the understanding of graph structures have been a central area of study in algorithm design, the rapid increase of data sets over the last decades has posed new challenges for designing efficient algorithms that process large-scale graphs. These challenges arise from two usual assumptions in classical algorithm design, namely that graphs are static and that they fit into a single machine. However, in many application domains, graphs are subject to frequent changes over time, and their massive size makes them infeasible to be stored in the memory of a single machine. Driven by the need to devise new tools for overcoming such challenges, this thesis focuses on two areas of modern algorithm design that directly deal with processing massive graphs, namely dynamic graph algorithms and graph sparsification. We develop new algorithmic techniques from both dynamic and sparsification perspective for a multitude of graph-based optimization problems which lie at the core of Spectral Graph Theory, Graph Partitioning, and Metric Embeddings. Our algorithms are faster than any previous one and design smaller sparsifiers with better (approximation) quality. More importantly, this work introduces novel reduction techniques that show unexpected connections between seemingly different areas such as dynamic graph algorithms and graph sparsification.
The Sivers function, an asymmetric transverse-momentum distribution of the quarks in a transversely polarized nucleon, is calculated in the MIT bag model. The bag quark wave functions contain both $S$-wave and $P$-wave components, and their interference leads to nonvanishing Sivers function in the presence of the final state interactions. We approximate these interactions through one-gluon exchange. An estimate of another transverse momentum dependent distribution $h_1^\perp$ is also performed in the same model.
We try to apply the cosmic crystallography method of Lehoucq, Lachieze-Rey, and Luminet to a universe model with closed spatial section of negative curvature. But the sharp peaks predicted for Einstein-de Sitter closed models do not appear in our hyperbolic example. So we turn to a variant of that method, by subtracting from the distribution of distances between images in the closed model, the similar distribution in Friedmann's open model. The result is a plot with much oscillation in small scales, modulated by a long wavelength quasi-sinusoidal pattern.
This paper presents a robust multi-class multi-object tracking (MCMOT) formulated by a Bayesian filtering framework. Multi-object tracking for unlimited object classes is conducted by combining detection responses and changing point detection (CPD) algorithm. The CPD model is used to observe abrupt or abnormal changes due to a drift and an occlusion based spatiotemporal characteristics of track states. The ensemble of convolutional neural network (CNN) based object detector and Lucas-Kanede Tracker (KLT) based motion detector is employed to compute the likelihoods of foreground regions as the detection responses of different object classes. Extensive experiments are performed using lately introduced challenging benchmark videos; ImageNet VID and MOT benchmark dataset. The comparison to state-of-the-art video tracking techniques shows very encouraging results.
The correspondences between logarithmic operators in the CFTs on the boundary of AdS_3 and on the world-sheet and dipole fields in the bulk are studied using the free field formulation of the SL(2,C)/SU(2) WZNW model. We find that logarithmic operators on the boundary are related to operators on the world-sheet which are in indecomposable representations of SL(2). The Knizhnik-Zamolodchikov equation is used to determine the conditions for those representations to appear in the operator product expansions of the model.
We explain how perturbative string theory can be viewed as an exactly renormalizable Weyl invariant quantum mechanics in the worldsheet representation clarifying why string scattering amplitudes are both finite and unambiguously normalized and explaining the origin of UV-IR relations in spacetime. As applications we examine the worldsheet representation of nonperturbative type IB states and of string solitons. We conclude with an analysis of the thermodynamics of a free closed string gas establishing the absence of the Hagedorn phase transition. We show that the 10D heterotic strings share a stable finite temperature ground state with gauge group SO(16)xSO(16). The free energy at the self-dual Kosterlitz-Thouless phase transition is minimized with finite entropy and positive specific heat. The open and closed string gas transitions to a confining long string phase at a temperature at or below the string scale in the presence of an external electric field.
Important gains have recently been obtained in object detection by using training objectives that focus on {\em hard negative} examples, i.e., negative examples that are currently rated as positive or ambiguous by the detector. These examples can strongly influence parameters when the network is trained to correct them. Unfortunately, they are often sparse in the training data, and are expensive to obtain. In this work, we show how large numbers of hard negatives can be obtained {\em automatically} by analyzing the output of a trained detector on video sequences. In particular, detections that are {\em isolated in time}, i.e., that have no associated preceding or following detections, are likely to be hard negatives. We describe simple procedures for mining large numbers of such hard negatives (and also hard {\em positives}) from unlabeled video data. Our experiments show that retraining detectors on these automatically obtained examples often significantly improves performance. We present experiments on multiple architectures and multiple data sets, including face detection, pedestrian detection and other object categories.
Artificial Intelligence (AI) systems such as autonomous vehicles, facial recognition, and speech recognition systems are increasingly integrated into our daily lives. However, despite their utility, these AI systems are vulnerable to a wide range of attacks such as adversarial, backdoor, data poisoning, membership inference, model inversion, and model stealing attacks. In particular, numerous attacks are designed to target a particular model or system, yet their effects can spread to additional targets, referred to as transferable attacks. Although considerable efforts have been directed toward developing transferable attacks, a holistic understanding of the advancements in transferable attacks remains elusive. In this paper, we comprehensively explore learning-based attacks from the perspective of transferability, particularly within the context of cyber-physical security. We delve into different domains -- the image, text, graph, audio, and video domains -- to highlight the ubiquitous and pervasive nature of transferable attacks. This paper categorizes and reviews the architecture of existing attacks from various viewpoints: data, process, model, and system. We further examine the implications of transferable attacks in practical scenarios such as autonomous driving, speech recognition, and large language models (LLMs). Additionally, we outline the potential research directions to encourage efforts in exploring the landscape of transferable attacks. This survey offers a holistic understanding of the prevailing transferable attacks and their impacts across different domains.
Spiking neural networks (SNNs) can be run on neuromorphic devices with ultra-high speed and ultra-low energy consumption because of their binary and event-driven nature. Therefore, SNNs are expected to have various applications, including as generative models being running on edge devices to create high-quality images. In this study, we build a variational autoencoder (VAE) with SNN to enable image generation. VAE is known for its stability among generative models; recently, its quality advanced. In vanilla VAE, the latent space is represented as a normal distribution, and floating-point calculations are required in sampling. However, this is not possible in SNNs because all features must be binary time series data. Therefore, we constructed the latent space with an autoregressive SNN model, and randomly selected samples from its output to sample the latent variables. This allows the latent variables to follow the Bernoulli process and allows variational learning. Thus, we build the Fully Spiking Variational Autoencoder where all modules are constructed with SNN. To the best of our knowledge, we are the first to build a VAE only with SNN layers. We experimented with several datasets, and confirmed that it can generate images with the same or better quality compared to conventional ANNs. The code is available at https://github.com/kamata1729/FullySpikingVAE
We investigate the production of cosmic ray (CR) protons at cosmological shocks by performing, for the first time, numerical simulations of large scale structure formation that include directly the acceleration, transport and energy losses of the high energy particles. CRs are injected at shocks according to the thermal leakage model and, thereafter, accelerated to a power-law distribution as indicated by the test particle limit of the diffusive shock acceleration theory. The evolution of the CR protons accounts for losses due to adiabatic expansion/compression, Coulomb collisions and inelastic p-p scattering. Our results suggest that CR protons produced at shocks formed in association with the process of large scale structure formation could amount to a substantial fraction of the total pressure in the intra-cluster medium. Their presence should be easily revealed by GLAST through detection of gamma-ray flux from the decay of neutral pions produced in inelastic p-p collisions of such CR protons with nuclei of the intra-cluster gas. This measurement will allow a direct determination of the CR pressure contribution in the intra-cluster medium. We also find that the spatial distribution of CR is typically more irregular than that of the thermal gas because it is more influenced by the underlying distribution of shocks. This feature is reflected in the appearance of our gamma-ray synthetic images. Finally, the average CR pressure distribution appears statistically slightly more extended than the thermal pressure.
We study invariant Einstein metrics on the indicated homogeneous manifolds $M$, the corresponding algebraic Einstein equations $E$, the associated with $M$ and $E$ Newton polytopes $P(M)$, and the integer volumes $\nu = \nu(P(M))$ of it (the Newton numbers). We show that $\nu = 80, 152,...,152$ respectively. It is claimed that the numbers $\epsilon = \epsilon(M)$ of complex solutions of $E$ equals $ \nu - 18, \nu - 18, \nu,..., \nu $. The results are consistent with classification of non K\"ahler invariant Einstein metrics on $G_2/T^2$ obtained recently by Y.Sakane, A. Arvanitoyeorgos, and I. Chrysikos. We present also a short description of all invariant complex Einstein metrics on $ SU_4/T^3 $. We prove existence of Riemannian non K\"ahler invariant Einstein metrics on $G_2/T^2$-like K\"ahler homogeneous spaces $ E_6/T^2\cdot(A_2)^2, E_7/T^2\cdot A_5, E_8/T^2\cdot E_6, F_4/T^2\cdot A_2$, where $ T^2\cdot A_5 \subset A_2\cdot A_5\subset E_7 $ and some other results.
We discuss the current picture of the standard model's scalar sector at strong coupling. We compare the pattern observed in the scalar sector in perturbation theory up to two-loop with the nonperturbative solution obtained by a next-to-leading order 1/N expansion. In particular, we analyze two resonant Higgs scattering processes, ff -> H -> f'f' and ff -> H -> ZZ, WW. We describe the ingredients of the nonperturbative calculation, such as the tachyonic regularization, the higher order 1/N intermediate renormalization, and the numerical methods for evaluating the graphs. We discuss briefly the perspectives and usefulness of extending these nonperturbative methods to other theories.
We give piecewise affine maps on the unit cube whose symbolic representation is the Dyck shift. This leads to a different way of verifying the chaotic nature of this system, including the computation of entropy.
In the recent years, public use of artistic effects for editing and beautifying images has encouraged researchers to look for new approaches to this task. Most of the existing methods apply artistic effects to the whole image. Exploitation of neural network vision technologies like object detection and semantic segmentation could be a new viewpoint in this area. In this paper, we utilize an instance segmentation neural network to obtain a class mask for separately filtering the background and foreground of an image. We implement a top prior-mask selection to let us select an object class for filtering purpose. Different artistic effects are used in the filtering process to meet the requirements of a vast variety of users. Also, our method is flexible enough to allow the addition of new filters. We use pre-trained Mask R-CNN instance segmentation on the COCO dataset as the segmentation network. Experimental results on the use of different filters are performed. System's output results show that this novel approach can create satisfying artistic images with fast operation and simple interface.
We show the existence of a natural Dirichlet-to-Neumann map on Riemannian manifolds with boundary and bounded geometry, such that the bottom of the Dirichlet spectrum is positive. This map regarded as a densely defined operator in the $L^2$-space of the boundary admits Friedrichs extension. We focus on the spectrum of this operator on covering spaces and total spaces of Riemannian principal bundles over compact manifolds.
We present preliminary results of a study of Delta S = 2 matrix elements originating from physics beyond the Standard Model. Using 2+1 flavour Domain Wall Fermions we obtain the non-perturbative renormalisation (mixing) matrix in the RI scheme. We also discuss plans for the chiral extrapolation of the renormalised matrix elements in a partially quenched set-up.
The two-particle correlation function employed in Hanbury-Brown Twiss interferometry and femtoscopy is traditionally parameterized by a Gaussian form. Other forms, however, have also been used, including the somewhat more general L\'evy form. Here we consider a variety of effects present in realistic femtoscopic studies which may modify the shape of the correlation function and thereby influence the physical interpretation of a given parameterization.
The accumulation of residual stress during welding and additive manufacturing is an important effect that can significantly anticipate the workpiece failure. In this work we exploit the physical and analytical transparency of a 1.5D model to show that the deposition of thermally expanded material onto an elastic substrate leads to the accumulation of strain incompatibility. This field, which is the source of residual stress in the system, introduces memory of the construction history even in the absence of plastic deformations. The model is then applied to describe the onset and the progression of residual stresses during deposition, their evolution upon cooling, and the fundamental role played by the velocity of the moving heat source.
Supervised learning, while deployed in real-life scenarios, often encounters instances of unknown classes. Conventional algorithms for training a supervised learning model do not provide an option to detect such instances, so they miss-classify such instances with 100% probability. Open Set Recognition (OSR) and Non-Exhaustive Learning (NEL) are potential solutions to overcome this problem. Most existing methods of OSR first classify members of existing classes and then identify instances of new classes. However, many of the existing methods of OSR only makes a binary decision, i.e., they only identify the existence of the unknown class. Hence, such methods cannot distinguish test instances belonging to incremental unseen classes. On the other hand, the majority of NEL methods often make a parametric assumption over the data distribution, which either fail to return good results, due to the reason that real-life complex datasets may not follow a well-known data distribution. In this paper, we propose a new online non-exhaustive learning model, namely, Non-Exhaustive Gaussian Mixture Generative Adversarial Networks (NE-GM-GAN) to address these issues. Our proposed model synthesizes Gaussian mixture based latent representation over a deep generative model, such as GAN, for incremental detection of instances of emerging classes in the test data. Extensive experimental results on several benchmark datasets show that NE-GM-GAN significantly outperforms the state-of-the-art methods in detecting instances of novel classes in streaming data.
At online retail platforms, it is crucial to actively detect the risks of transactions to improve customer experience and minimize financial loss. In this work, we propose xFraud, an explainable fraud transaction prediction framework which is mainly composed of a detector and an explainer. The xFraud detector can effectively and efficiently predict the legitimacy of incoming transactions. Specifically, it utilizes a heterogeneous graph neural network to learn expressive representations from the informative heterogeneously typed entities in the transaction logs. The explainer in xFraud can generate meaningful and human-understandable explanations from graphs to facilitate further processes in the business unit. In our experiments with xFraud on real transaction networks with up to 1.1 billion nodes and 3.7 billion edges, xFraud is able to outperform various baseline models in many evaluation metrics while remaining scalable in distributed settings. In addition, we show that xFraud explainer can generate reasonable explanations to significantly assist the business analysis via both quantitative and qualitative evaluations.
Heralded near-deterministic multi-qubit controlled phase gates with integrated error detection have recently been proposed by Borregaard et al. [Phys. Rev. Lett. 114, 110502 (2015)]. This protocol is based on a single four-level atom (a heralding quartit) and $N$ three-level atoms (operational qutrits) coupled to a single-resonator mode acting as a cavity bus. Here we generalize this method for two distant resonators without the cavity bus between the heralding and operational atoms. Specifically, we analyze the two-qubit controlled-Z gate and its multi-qubit-controlled generalization (i.e., a Toffoli-like gate) acting on the two-lowest levels of $N$ qutrits inside one resonator, with their successful actions being heralded by an auxiliary microwave-driven quartit inside the other resonator. Moreover, we propose a circuit-quantum-electrodynamics realization of the protocol with flux and phase qudits in linearly-coupled transmission-line resonators with dissipation. These methods offer a quadratic fidelity improvement compared to cavity-assisted deterministic gates.
Cartesian differential categories come equipped with a differential combinator that formalizes the derivative from multi-variable differential calculus, and also provide the categorical semantics of the differential $\lambda$-calculus. An important source of examples of Cartesian differential categories are the coKleisli categories of the comonads of differential categories, where the latter concept provides the categorical semantics of differential linear logic. In this paper, we generalize this construction by introducing Cartesian differential comonads, which are precisely the comonads whose coKleisli categories are Cartesian differential categories, and thus allows for a wider variety of examples of Cartesian differential categories. As such, we construct new examples of Cartesian differential categories from Cartesian differential comonads based on power series, divided power algebras, and Zinbiel algebras.
Assume $AD+V=L(\mathbb{R})$. Let $\kappa=\utilde{\delta}^2_1$, the supremum of all $\utilde{\Delta}^2_1$ prewellorderings. We prove that extenders on the sequence of $\H$ that have critical point $\kappa$ are generated by countably complete measures. This provides a partial reversal of Woodin's result that the $<\Theta$-strongness of $\kappa$ in $\H$ is witnessed by $\kappa$-complete ultrafilters on $\k$. The aforementioned characterization of extenders works in a more general setting for all cutpoint measurable cardinals of $\H$ in all models of determinacy where the fine structural analysis of $\H$ has been carried out. For example, it holds in the minimal model of the Largest Suslin Axiom. It also gives a simple proof of a theorem of Steel that the successor members of the Solovay sequence are cutpoints in $\H$ (in models where $\H$ analysis is carried out).
We show that the quasi-skutterudite superconductor Sr_3Ir_4Sn_{13} undergoes a structural transition from a simple cubic parent structure, the I-phase, to a superlattice variant, the I'-phase, which has a lattice parameter twice that of the high temperature phase. We argue that the superlattice distortion is associated with a charge density wave transition of the conduction electron system and demonstrate that the superlattice transition temperature T* can be suppressed to zero by combining chemical and physical pressure. This enables the first comprehensive investigation of a superlattice quantum phase transition and its interplay with superconductivity in a cubic charge density wave system.
In this work, the zero-temperature limit of the thermodynamic spin-density functional theory is investigated. The coarse-grained approach to the equilibrium density operator is used to describe the equilibrium state. The characteristic functions of a macrostate are introduced and their zero-temperature limits are investigated. A detailed discussion of the spin-grand-canonical ensemble in the entropy and energy representations is performed. The maps between the state function variables at 0K limit are rigorously studied for both representations. In the spin-canonical ensemble, the energy surface and the discontinuity pattern are investigated. Finally, based on the maps between the state function variables at 0K limit, the Hohenberg-Kohn theorem for the systems with non-integer electron and spin numbers at zero-temperature limit is formulated.
Rough membership function defines the measurement of relationship between conditional and decision attribute from an Information system. In this paper we propose a new method to construct rough graph through rough membership function $\omega_{G}^F(f)$. Rough graph identifies the pattern between the objects with imprecise and uncertain information. We explore the operations and properties of rough graph in various stages of its structure.
We propose a neural network model for MDG and optical SNR estimation in SDM transmission. We show that the proposed neural-network-based solution estimates MDG and SNR with high accuracy and low complexity from features extracted after DSP.
We have created a knowledge graph based on major data sources used in ecotoxicological risk assessment. We have applied this knowledge graph to an important task in risk assessment, namely chemical effect prediction. We have evaluated nine knowledge graph embedding models from a selection of geometric, decomposition, and convolutional models on this prediction task. We show that using knowledge graph embeddings can increase the accuracy of effect prediction with neural networks. Furthermore, we have implemented a fine-tuning architecture that adapts the knowledge graph embeddings to the effect prediction task and leads to better performance. Finally, we evaluate certain characteristics of the knowledge graph embedding models to shed light on the individual model performance.
Recent advances in distributed learning raise environmental concerns due to the large energy needed to train and move data to/from data centers. Novel paradigms, such as federated learning (FL), are suitable for decentralized model training across devices or silos that simultaneously act as both data producers and learners. Unlike centralized learning (CL) techniques, relying on big-data fusion and analytics located in energy hungry data centers, in FL scenarios devices collaboratively train their models without sharing their private data. This article breaks down and analyzes the main factors that influence the environmental footprint of FL policies compared with classical CL/Big-Data algorithms running in data centers. The proposed analytical framework takes into account both learning and communication energy costs, as well as the carbon equivalent emissions; in addition, it models both vanilla and decentralized FL policies driven by consensus. The framework is evaluated in an industrial setting assuming a real-world robotized workplace. Results show that FL allows remarkable end-to-end energy savings (30%-40%) for wireless systems characterized by low bit/Joule efficiency (50 kbit/Joule or lower). Consensus-driven FL does not require the parameter server and further reduces emissions in mesh networks (200 kbit/Joule). On the other hand, all FL policies are slower to converge when local data are unevenly distributed (often 2x slower than CL). Energy footprint and learning loss can be traded off to optimize efficiency.
We calculate the magnetic moments of light nuclei ($A < 20$) using the auxiliary field diffusion Monte Carlo method and local two- and three-nucleon forces with electromagnetic currents from chiral effective field theory. For all nuclei under consideration, we also calculate the ground-state energies and charge radii. We generally find a good agreement with experimental values for all of these observables. For the electromagnetic currents, we explore the impact of employing two different power countings, and study theoretical uncertainties stemming from the truncation of the chiral expansion order-by-order for select nuclei within these two approaches. We find that it is crucial to employ consistent power countings for interactions and currents to achieve a systematic order-by-order convergence.
For static fluid spheres, the condition of hydrostatic equilibrium is given by the generalized Tolman--Oppenheimer--Volkoff (TOV) equation, a Riccati equation in the radial pressure. For a perfect fluid source, it is known that finding a new solution from an existing solution requires solving a Bernoulli equation, if the density profile is kept the same. In this paper, we consider maps between static (an)isotropic fluid spheres with the same (arbitrary) density profile and present solution-generating techniques to find new solutions from existing ones. The maps, in general, require solving an associated Riccati equation, which, unlike the Bernoulli equation, cannot be solved by quadrature. In any case, it can be shown that the output solution is not, in general, regular for a given regular input solution. However, if pressure anisotropy is kept the same, the new solution is both regular and can be found by solving a Bernoulli equation. We give a few examples where the generalized TOV equation, under algebraic constraints, can be converted into a Bernoulli equation and thus, solved exactly. We discuss the physical significance of these Bernoulli equations. Since the density profile remains the same in our approach, the spatial line element is identical for all solutions, which facilitates direct comparison between various equilibrium configurations using fluid variables as functions of the radial coordinate. Finally, combining with the previous study on generation algorithms, we show how this study leads us to a new three-parameter family of exact solutions that satisfy all desirable physical conditions.
We investigate the cosmological behavior of mimetic F(R) gravity. This scenario is the F(R) extension of usual mimetic gravity classes, which are based on re-parametrizations of the metric using new, but not propagating, degrees of freedom, that can lead to a wider family of solutions. Performing a detailed dynamical analysis for exponential, power-law, and arbitrary F(R) forms, we extracted the corresponding critical points. Interestingly enough, we found that although the new features of mimetic F(R) gravity can affect the universe evolution at early and intermediate times, at late times they will not have any effect, and the universe will result at stable states that coincide with those of usual F(R) gravity. However, this feature holds for the late-time background evolution only. On the contrary, the behavior of the perturbations is expected to be different since the new term contributes to the perturbations even if it does not contribute at the background level.
We consider the allocation of limited resources to heterogeneous customers who arrive in an online fashion. We would like to allocate the resources "fairly", so that no group of customers is marginalized in terms of their overall service rate. We study whether this is possible to do so in an online fashion, and if so, what a good online allocation policy is. We model this problem using online bipartite matching under stationary arrivals, a fundamental model in the literature typically studied under the objective of maximizing the total number of customers served. We instead study the objective of maximizing the minimum service rate across all groups, and propose two notions of fairness: long-run and short-run. For these fairness objectives, we analyze how competitive online algorithms can be, in comparison to offline algorithms which know the sequence of demands in advance. For long-run fairness, we propose two online heuristics (Sampling and Pooling) which establish asymptotic optimality in different regimes (no specialized supplies, no rare demand types, or imbalanced supply/demand). By contrast, outside all of these regimes, we show that the competitive ratio of online algorithms is between 0.632 and 0.732. For short-run fairness, we show for complete bipartite graphs that the competitive ratio of online algorithms is between 0.863 and 0.942; we also derive a probabilistic rejection algorithm which is asymptotically optimal in the total demand. Depending on the overall scarcity of resources, either our Sampling or Pooling heuristics could be desirable. The most difficult situation for online allocation occurs when the total supply is just enough to serve the total demand. We simulate our algorithms on a public ride-hailing dataset, which both demonstrates the efficacy of our heuristics and validates our managerial insights.
Metal mixing plays critical roles in the enrichment of metals in galaxies. The abundance of elements such as Mg, Fe, and Ba in metal-poor stars help us understand the metal mixing in galaxies. However, the efficiency of metal mixing in galaxies is not yet understood. Here we report a series of $N$-body/smoothed particle hydrodynamics simulations of dwarf galaxies with different efficiencies of metal mixing using turbulence-induced mixing model. We show that metal mixing apparently occurs in dwarf galaxies from Mg and Ba abundance. We find that the scaling factor for metal diffusion larger than 0.01 is necessary to reproduce the observation of Ba abundance in dwarf galaxies. This value is consistent with the value expected from turbulence theory and experiment. We also find that timescale of metal mixing is less than 40 Myr. This timescale is shorter than that of typical dynamical times of dwarf galaxies. We demonstrate that the determination of a degree of scatters of Ba abundance by the observation will help us to constrain the efficiency of metal mixing more precisely.
A free-floating planet is a planetary-mass object that orbits around a non-stellar massive object (e.g. a brown dwarf) or around the Galactic Center. The presence of exomoons orbiting free-floating planets has been theoretically predicted by several models. Under specific conditions, these moons are able to retain an atmosphere capable of ensuring the long-term thermal stability of liquid water on their surface. We model this environment with a one-dimensional radiative-convective code coupled to a gas-phase chemical network including cosmic rays and ion-neutral reactions. We find that, under specific conditions and assuming stable orbital parameters over time, liquid water can be formed on the surface of the exomoon. The final amount of water for an Earth-mass exomonoon is smaller than the amount of water in Earth oceans, but enough to host the potential development of primordial life. The chemical equilibrium time-scale is controlled by cosmic rays, the main ionization driver in our model of the exomoon atmosphere.
Inducing chirality in optically and electronically active materials is interesting for applications in sensing and quantum information transmission. Two-dimensional (2D) transition metal chalcogenides (TMDs) possess excellent electronic and optical properties but are achiral. Here we demonstrate chirality induction in atomically thin layers of 2D MoS2 by functionalization with chiral thiol molecules. Analysis of X-ray absorption near-edge structure and Raman optical activity with circularly polarized excitation suggest chemical and electronic interactions that leads chirality transfer from the molecules to the MoS2. We confirm chirality induction in 2D MoS2 with circular dichroism measurements that show absorption bands at wavelengths of 380-520 nm and 520-600 nm with giant molar ellipticity of 10^8 deg cm2/dmol 2-3 orders of magnitude higher than 3D chiral materials. Phototransistors fabricated from atomically thin chiral MoS2 for detection of circularly polarized light exhibit responsivity of >10^2 A/W and maximum anisotropy g-factor of 1.98 close to the theoretical maximum of 2.0, which indicates that the chiral states of photons are fully distinguishable by the photodetectors. Our results demonstrate that it is possible achieve chirality induction in monolayer MoS2 by molecular functionalization and realise ultra-sensitive detectors for circularly polarized photons.
We investigate theoretically a topological vortex phase transition induced by a superradiant phase transition in an atomic Bose-Einstein condensate driven by a Laguerre-Gaussian optical mode. We show that superradiant radiation can either carry zero angular momentum, or be in a rotating Laguerre-Gaussian mode with angular momentum. The conditions leading to these two regimes are determined in terms of the width for the pump laser and the condensate size for the limiting cases where the recoil energy is both much smaller and larger than the atomic interaction energy.
Like generic multi-task learning, continual learning has the nature of multi-objective optimization, and therefore faces a trade-off between the performance of different tasks. That is, to optimize for the current task distribution, it may need to compromise performance on some previous tasks. This means that there exist multiple models that are Pareto-optimal at different times, each addressing a distinct task performance trade-off. Researchers have discussed how to train particular models to address specific trade-off preferences. However, existing algorithms require training overheads proportional to the number of preferences -- a large burden when there are multiple, possibly infinitely many, preferences. As a response, we propose Imprecise Bayesian Continual Learning (IBCL). Upon a new task, IBCL (1) updates a knowledge base in the form of a convex hull of model parameter distributions and (2) obtains particular models to address task trade-off preferences with zero-shot. That is, IBCL does not require any additional training overhead to generate preference-addressing models from its knowledge base. We show that models obtained by IBCL have guarantees in identifying the Pareto optimal parameters. Moreover, experiments on standard image classification and NLP tasks support this guarantee. Statistically, IBCL improves average per-task accuracy by at most 23% and peak per-task accuracy by at most 15% with respect to the baseline methods, with steadily near-zero or positive backward transfer. Most importantly, IBCL significantly reduces the training overhead from training 1 model per preference to at most 3 models for all preferences.
We present narrow-band near-infrared images of a sample of 11 Galactic planetary nebulae (PNe) obtained in the molecular hydrogen (H$_{2}$) 2.122 $\mu$m and Br$\gamma$ 2.166 $\mu$m emission lines and the $K_{\rm c}$ 2.218 $\mu$m continuum. These images were collected with the Wide-field InfraRed Camera (WIRCam) on the 3.6m Canada-France-Hawaii Telescope (CFHT); their unprecedented depth and wide field of view allow us to find extended nebular structures in H$_{2}$ emission in several PNe, some of these being the first detection. The nebular morphologies in H$_{2}$ emission are studied in analogy with the optical images, and indication on stellar wind interactions is discussed. In particular, the complete structure of the highly asymmetric halo in NGC6772 is witnessed in H$_{2}$, which strongly suggests interaction with the interstellar medium. Our sample confirms the general correlation between H$_{2}$ emission and the bipolarity of PNe. The knotty/filamentary fine structures of the H$_{2}$ gas are resolved in the inner regions of several ring-like PNe, also confirming the previous argument that H2 emission mostly comes from knots/clumps embedded within fully ionized material at the equatorial regions. Moreover, the deep H$_{2}$ image of the butterfly-shaped Sh1-89, after removal of field stars, clearly reveals a tilted ring structure at the waist. These high-quality CFHT images justify follow-up detailed morpho-kinematic studies that are desired to deduce the true physical structures of a few PNe in the sample.
The moduli spaces of trigonal curves of odd genus $g>4$ are proven to be rational.
We present a general scheme for the calculation of the Renyi entropy of a subsystem in quantum many-body models that can be efficiently simulated via quantum Monte Carlo. When the simulation is performed at very low temperature, the above approach delivers the entanglement Renyi entropy of the subsystem, and it allows to explore the crossover to the thermal Renyi entropy as the temperature is increased. We implement this scheme explicitly within the Stochastic Series expansion as well as within path-integral Monte Carlo, and apply it to quantum spin and quantum rotor models. In the case of quantum spins, we show that relevant models in two dimensions with reduced symmetry (XX model or hardcore bosons, transverse-field Ising model at the quantum critical point) exhibit an area law for the scaling of the entanglement entropy.
Supervised trackers trained on labeled data dominate the single object tracking field for superior tracking accuracy. The labeling cost and the huge computational complexity hinder their applications on edge devices. Unsupervised learning methods have also been investigated to reduce the labeling cost but their complexity remains high. Aiming at lightweight high-performance tracking, feasibility without offline pre-training, and algorithmic transparency, we propose a new single object tracking method, called the green object tracker (GOT), in this work. GOT conducts an ensemble of three prediction branches for robust box tracking: 1) a global object-based correlator to predict the object location roughly, 2) a local patch-based correlator to build temporal correlations of small spatial units, and 3) a superpixel-based segmentator to exploit the spatial information of the target frame. GOT offers competitive tracking accuracy with state-of-the-art unsupervised trackers, which demand heavy offline pre-training, at a lower computation cost. GOT has a tiny model size (<3k parameters) and low inference complexity (around 58M FLOPs per frame). Since its inference complexity is between 0.1%-10% of DL trackers, it can be easily deployed on mobile and edge devices.
Bounded time series consisting of rates or proportions are often encountered in applications. This manuscript proposes a practical approach to analyze bounded time series, through a beta regression model. The method allows the direct interpretation of the regression parameters on the original response scale, while properly accounting for the heteroskedasticity typical of bounded variables. The serial dependence is modeled by a Gaussian copula, with a correlation matrix corresponding to a stationary autoregressive and moving average process. It is shown that inference, prediction, and control can be carried out straightforwardly, with minor modifications to standard analysis of autoregressive and moving average models. The methodology is motivated by an application to the influenza-like-illness incidence estimated by the Google${}^\circledR$ Flu Trends project.
A large-scale gradient in the metal abundance has been detected with ASCA from an X-ray bright cluster of galaxies AWM7. The metal abundance shows a peak of 0.5 solar at the center and smoothly declines to <~ 0.2 solar at a radius of 500 kpc. The gas temperature is found to be constant at 3.8 keV. The radial distribution of iron can be fit with a beta-model with beta ~ 0.8 assuming the same core radius (115 kpc) as that of the intracluster medium. The metal distribution in AWM7 suggests that the gas injected from galaxies is not efficiently mixed in the cluster space and traces the distribution of galaxies.
The launching of up to four astrometric missions in the next decade will enable us to make measurements of stars in tidal streamers from Galactic satellites of sufficient accuracy to place strong constraints on the mass distribution in the Milky Way. In this paper we simulate observations of debris populations in order to assess the required properties of any data set chosen to implement this experiment. We apply our results to find the desired target properties of stars (e.g., accuracy of velocity and proper motion measurements) associated with the dwarf spheroidal satellites of the Milky Way.
We design an arbitrary high-order accurate nodal discontinuous Galerkin spectral element approximation for the nonlinear two dimensional shallow water equations with non-constant, possibly discontinuous, bathymetry on unstructured, possibly curved, quadrilateral meshes. The scheme is derived from an equivalent flux differencing formulation of the split form of the equations. We prove that this discretisation exactly preserves the local mass and momentum. Furthermore, combined with a special numerical interface flux function, the method exactly preserves the mathematical entropy, which is the total energy for the shallow water equations. By adding a specific form of interface dissipation to the baseline entropy conserving scheme we create a provably entropy stable scheme. That is, the numerical scheme discretely satisfies the second law of thermodynamics. Finally, with a particular discretisation of the bathymetry source term we prove that the numerical approximation is well-balanced. We provide numerical examples that verify the theoretical findings and furthermore provide an application of the scheme for a partial break of a curved dam test problem.